ai models

OpenAI’s o3 suggests AI models are scaling in new ways — but so are the costs

Last month, AI founders and investors told TechCrunch that we’re now in the “second era of scaling laws,” noting how established methods of improving AI models were showing diminishing returns. One promising new method they suggested could keep gains was “test-time scaling,” which seems to be what’s behind the performance of OpenAI’s o3 model — […]

OpenAI’s o3 suggests AI models are scaling in new ways — but so are the costs Read More »

HuggingFace CEO has concerns about Chinese open source AI models

China’s open source AI models have been making the news lately for their strong performance on various AI tasks such as coding and ‘reasoning.’  However, they have also attracted criticism – including from OpenAI employees – for censoring topics sensitive to the Chinese government, such as the Tiananmen Square massacre. HuggingFace’s CEO Clement Delangue says

HuggingFace CEO has concerns about Chinese open source AI models Read More »

Inflection CEO says it’s done trying to make next generation AI models

Just last year, Inflection AI was as hot as a startup could be, releasing best-in-class AI models it claimed could outperform technology from OpenAI, Meta, and Google. That’s a stark contrast compared to today, as Inflection’s new CEO tells TechCrunch that his startup is simply no longer trying to compete on that front. Between then

Inflection CEO says it’s done trying to make next generation AI models Read More »

Current AI scaling laws are showing diminishing returns, forcing AI labs to change course

AI labs traveling the road to super-intelligent systems are realizing they might have to take a detour. “AI scaling laws,” the methods and expectations that labs have used to increase the capabilities of their models for the last five years, are now showing signs of diminishing returns, according to several AI investors, founders, and CEOs

Current AI scaling laws are showing diminishing returns, forcing AI labs to change course Read More »

Zuckerberg says Meta will need 10x more computing power to train Llama 4 than Llama 3

Meta, which develops one of the biggest foundational open-source large language models, Llama, believes it will need significantly more computing power to train models in the future. Mark Zuckerberg said on Meta’s second-quarter earnings call on Tuesday that to train Llama 4 the company will need 10x more compute than what was needed to train

Zuckerberg says Meta will need 10x more computing power to train Llama 4 than Llama 3 Read More »