ai safety

Anthropic endorses California’s AI safety bill, SB 53

On Monday, Anthropic announced an official endorsement of SB 53, a California bill from state Senator Scott Wiener that would impose first-in-the-nation transparency requirements on the world’s largest AI model developers. Anthropic’s endorsement marks a rare and major win for SB 53, at a time when major tech groups like CTA and Chamber for Progress […]

Anthropic endorses California’s AI safety bill, SB 53 Read More »

OpenAI to route sensitive conversations to GPT-5, introduce parental controls

OpenAI said Tuesday it plans to route sensitive conversations to reasoning models like GPT-5 and roll out parental controls within the next month – part of an ongoing response to recent safety incidents involving ChatGPT failing to detect mental distress. The new guardrails come in the aftermath of the suicide of teenager Adam Raine, who

OpenAI to route sensitive conversations to GPT-5, introduce parental controls Read More »

Trillion with a ‘T’? That’s a lot of dollars, Nvidia.

Nvidia reported another massive quarter this week with $46.7 billion in revenue, a 56% year-over-year increase driven almost entirely by AI demand. But despite CEO Jensen Huang’s bold prediction of $3 to 4 trillion in global AI infrastructure spending in the next five years, the stock slid as investors questioned how long this kind of

Trillion with a ‘T’? That’s a lot of dollars, Nvidia. Read More »

OpenAI and Anthropic researchers decry ‘reckless’ safety culture at Elon Musk’s xAI

AI safety researchers from OpenAI, Anthropic, and other organizations are speaking out publicly against the “reckless” and “completely irresponsible” safety culture at xAI, the billion-dollar AI startup owned by Elon Musk. The criticisms follow weeks of scandals at xAI that have overshadowed the company’s technological advances. Last week, the company’s AI chatbot, Grok, spouted antisemitic

OpenAI and Anthropic researchers decry ‘reckless’ safety culture at Elon Musk’s xAI Read More »

Research leaders urge tech industry to monitor AI’s ‘thoughts’

AI researchers from OpenAI, Google DeepMind, Anthropic, as well as a broad coalition of companies and nonprofit groups, are calling for deeper investigation into techniques for monitoring the so-called thoughts of AI reasoning models in a position paper published Tuesday. A key feature of AI reasoning models, such as OpenAI’s o3 and DeepSeek’s R1, are

Research leaders urge tech industry to monitor AI’s ‘thoughts’ Read More »

California lawmaker behind SB 1047 reignites push for mandated AI safety reports

California State Senator Scott Wiener on Wednesday introduced new amendments to his latest bill, SB 53, that would require the world’s largest AI companies to publish safety and security protocols and issue reports when safety incidents occur. If signed into law, California would be the first state to impose meaningful transparency requirements onto leading AI

California lawmaker behind SB 1047 reignites push for mandated AI safety reports Read More »

Anthropic says most AI models, not just Claude, will resort to blackmail

Several weeks after Anthropic released research claiming that its Claude Opus 4 AI model resorted to blackmailing engineers who tried to turn the model off in controlled test scenarios, the company is out with new research suggesting the problem is more widespread among leading AI models. On Friday, Anthropic published new safety research testing 16

Anthropic says most AI models, not just Claude, will resort to blackmail Read More »

ChatGPT will avoid being shut down in some life-threatening scenarios, former OpenAI researcher claims

Former OpenAI research leader Steven Adler published a new independent study on Wednesday claiming that, in certain scenarios, his former employer’s AI models will go to great lengths to try to avoid being shut down. In a blog post, Adler describes a series of experiments he ran on OpenAI’s latest GPT-4o model, the default model

ChatGPT will avoid being shut down in some life-threatening scenarios, former OpenAI researcher claims Read More »

Yoshua Bengio launches LawZero, a nonprofit AI safety lab

Turing Award winner Yoshua Bengio is launching a nonprofit AI safety lab called LawZero to build safer AI systems, he told the Financial Times on Monday. LawZero raised $30 million in philanthropic contributions from Skype founding engineer Jaan Tallinn, former Google chief Eric Schmidt, Open Philanthropy, and the Future of Life Institute, among others. The

Yoshua Bengio launches LawZero, a nonprofit AI safety lab Read More »