B.C. Government Slams OpenAI Over Silence Following Mass Shooting

By Tax assistant

Published on:

B.C. Government Slams OpenAI Over Silence Following Mass Shooting

The British Columbia government is expressing “profound” concern after learning that OpenAI met with provincial officials just one day after a deadly mass shooting in Tumbler Ridge, yet failed to disclose that they had previously flagged the shooter as a high-risk user.

Thank you for reading this post, don't forget to subscribe!

The Missing Disclosure

In June 2025—eight months before the attack—OpenAI internal staff debated whether to report 18-year-old Jesse Van Rootselaar to the police after her account was banned for promoting violence. They ultimately opted not to involve authorities, citing a lack of “imminent threat.”

The friction points identified by the province include:

  • The Meeting: OpenAI representatives sat down with B.C. officials on February 11, 2026, for a scheduled discussion about AI technology.
  • The Silence: Despite the province being in a state of shock over the shooting just 24 hours prior, OpenAI did not mention their prior interactions with the shooter or the ban they had placed on her account.
  • The Discovery: The B.C. government only learned of OpenAI’s prior knowledge through recent investigative reporting, rather than through direct communication from the company.

Comparison of Perspectives

StakeholderKey Argument
B.C. Premier David EbyIt is “unacceptable” and “disturbing” that a tech company held potentially life-saving intelligence and chose not to share it during a face-to-face meeting.
OpenAI SpokespersonThe company followed established safety protocols. At the time of the 2025 ban, the content did not meet the “specific and imminent” threshold required for law enforcement referral.
RCMPInvestigators are currently working to secure digital evidence that was only disclosed by the company after the tragedy occurred.

The Policy Debate

This incident has reignited a fierce debate over AI corporate responsibility:

  1. Privacy vs. Safety: OpenAI argued that “over-reporting” every policy violation could lead to privacy overreach and a flood of low-quality leads for police.
  2. Moral Obligation: B.C. officials argue that when a tragedy occurs, companies have a moral duty to be transparent about any known red flags, especially during active government consultations.

“The families of the victims deserve to know why this information remained in a boardroom in California instead of in the hands of local law enforcement.” — Summary of Provincial Sentiment

Leave a Comment