March 2010 April 2010 May 2010 June 2010 July 2010
August 2010
September 2010 October 2010
November 2010
December 2010 January 2011 February 2011 March 2011 April 2011 May 2011 June 2011 July 2011 August 2011 September 2011 October 2011 November 2011 December 2011 January 2012 February 2012 March 2012 April 2012 May 2012 June 2012 July 2012 August 2012 September 2012 October 2012 November 2012 December 2012 January 2013 February 2013 March 2013 April 2013 May 2013 June 2013 July 2013 August 2013 September 2013 October 2013 November 2013 December 2013 January 2014 February 2014 March 2014 April 2014 May 2014 June 2014 July 2014 August 2014 September 2014 October 2014 November 2014 December 2014 January 2015 February 2015 March 2015 April 2015 May 2015 June 2015 July 2015 August 2015 September 2015 October 2015 November 2015 December 2015 January 2016 February 2016 March 2016 April 2016 May 2016 June 2016 July 2016 August 2016 September 2016 October 2016 November 2016 December 2016 January 2017 February 2017 March 2017 April 2017 May 2017 June 2017 July 2017 August 2017 September 2017 October 2017 November 2017 December 2017 January 2018 February 2018 March 2018 April 2018 May 2018 June 2018 July 2018 August 2018 September 2018 October 2018 November 2018 December 2018 January 2019 February 2019 March 2019 April 2019 May 2019 June 2019 July 2019 August 2019 September 2019 October 2019 November 2019 December 2019 January 2020 February 2020 March 2020 April 2020 May 2020 June 2020 July 2020 August 2020 September 2020 October 2020 November 2020 December 2020 January 2021 February 2021 March 2021 April 2021 May 2021 June 2021 July 2021 August 2021 September 2021 October 2021 November 2021 December 2021 January 2022 February 2022 March 2022 April 2022 May 2022 June 2022 July 2022 August 2022 September 2022 October 2022 November 2022 December 2022 January 2023 February 2023 March 2023 April 2023 May 2023 June 2023 July 2023 August 2023 September 2023 October 2023 November 2023 December 2023 January 2024 February 2024 March 2024 April 2024 May 2024 June 2024 July 2024 August 2024 September 2024 October 2024 November 2024
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
28
29
30
News Every Day |

AI improvements are slowing down. Companies have a plan to break through the wall.

 The tech world has been debating if AI models are plateauing.
  • The rate of AI model improvement appears to be slowing, but some tech leaders say there is no wall.
  • It's prompted a debate over how companies can overcome AI bottlenecks.
  • Business Insider spoke to 12 people at the forefront of the AI boom to find out the path forward.

Silicon Valley leaders all-in on the artificial intelligence boom have a message for critics: their technology has not hit a wall.

A fierce debate over whether improvements in AI models have hit their limit has taken hold in recent weeks, forcing several CEOs to respond. OpenAI boss Sam Altman was among the first to speak out, posting on X this month that "there is no wall."

Dario Amodei, CEO of rival firm Anthropic, and Jensen Huang, the CEO of Nvidia, have also disputed reports that AI progress has slowed. Others, including Marc Andreessen, say AI models aren't getting noticeably better and are all converging to perform at roughly similar levels.

This is a trillion-dollar question for the tech industry. If tried-and-tested AI model training methods are providing diminishing returns, it could undermine the core reason for an unprecedented investment cycle that's funding new startups, products, and data centers — and even rekindling idled nuclear power plants.

Business Insider spoke to 12 people at the forefront of the AI industry, including startup founders, investors, and current and former insiders at Google DeepMind and OpenAI, about the challenges and opportunities ahead in the quest for superintelligent AI.

Together, they said that tapping into new types of data, building reasoning into systems, and creating smaller but more specialized models are some of the ways to keep the wheels of AI progress turning.

The pre-training dilemma

Researchers point to two key blocks that companies may encounter in an early phase of AI development, known as pre-training. The first is access to computing power. More specifically, this means getting hold of specialist chips called GPUs. It's a market dominated by Santa Clara-based chip giant Nvidia, which has battled with supply constraints in the face of nonstop demand.

"If you have $50 million to spend on GPUs but you're on the bottom of Nvidia's list — we don't have enough kimchi to throw at this, and it will take time," said Henri Tilloy, partner at French VC firm Singular.

Jensen Huang's Nvidia has become the world's most valuable company off the back of the AI boom.

There is another supply problem, too: training data. AI companies have run into limits on the quantity of public data they can secure to feed into their large language models, or LLMs, in pre-training.

This phase involves training an LLM on a vast corpus of data, typically scraped from the internet, and then processed by GPUs. That information is then broken down into "tokens," which form the fundamental units of data processed by a model.

While throwing more data and GPUs at a model has reliably produced smarter models year after year, companies have been exhausting the supply of publicly available data on the internet. Research firm Epoch AI predicts usable textual data could be squeezed dry by 2028.

"The internet is only so large," Matthew Zeiler, founder and CEO of Clarifai, told BI.

Multimodal and private data

Eric Landau, cofounder and CEO of data startup Encord, said that this is where other data sources will offer a path forward in the scramble to overcome the bottleneck in public data.

One example is multimodal data, which involves feeding AI systems visual and audio sources of information, such as photos or podcast recordings. "That's one part of the picture," Landau said. "Just adding more modalities of data." AI labs have already started using multimodal data as a tool, but Landau says it remains "very underutilized."

Sharon Zhou, cofounder and CEO of LLM platform Lamini, sees another vastly untapped area: private data. Companies have been securing licensing agreements with publishers to gain access to their vast troves of information. OpenAI, for instance, has struck partnerships with organizations such as Vox Media and Stack Overflow, a Q&A platform for developers, to bring copyrighted data into their models.

"We are not even close to using all of the private data in the world to supplement the data we need for pre-training," Zhou said. "From work with our enterprise and even startup customers, there's a lot more signal in that data that is very useful for these models to capture."

A data quality problem

A great deal of research effort is now focused on enhancing the quality of data that an LLM is trained on rather than just the quantity. Researchers could previously afford to be "pretty lazy about the data" in pre-training, Zhou said, by just chucking as much as possible at a model to see what stuck. "This isn't totally true anymore," she said.

One solution that companies are exploring is synthetic data, an artificial form of data generated by AI.

According to Daniele Panfilo, CEO of startup Aindo AI, synthetic data can be a "powerful tool to improve data quality," as it can "help researchers construct datasets that meet their exact information needs." This is particularly useful in a phase of AI development known as post-training, where techniques such as fine-tuning can be used to give a pre-trained model a smaller dataset that has been carefully crafted with specific domain expertise, such as law or medicine.

One former employee at Google DeepMind, the search giant's AI lab, told BI that "Gemini has shifted its strategy" from going bigger to more efficient. "I think they've realized that it is actually very expensive to serve such large models, and it is better to specialize them for various tasks through better post-training," the former employee said.

Google launched Gemini, formerly known as Bard, in 2023.

In theory, synthetic data offers a useful way to hone a model's knowledge and make it smaller and more efficient. In practice, there's no full consensus on how effective synthetic data can be in making models smarter.

"What we discovered this year with our synthetic data, called Cosmopedia, is that it can help for some things, but it's not the silver bullet that's going to solve our data problem," Thomas Wolf, cofounder and chief science officer at open-source platform Hugging Face, told BI.

Jonathan Frankle, the chief AI scientist at Databricks, said there's no "free lunch " when it comes to synthetic data and emphasized the need for human oversight. "If you don't have any human insight, and you don't have any process of filtering and choosing which synthetic data is most relevant, then all the model is doing is reproducing its own behavior because that's what the model is intended to do," he said.

Concerns around synthetic data came to a head after a paper published in July in the journal Nature said there was a risk of "model collapse" with "indiscriminate use" of synthetic data. The message was to tread carefully.

Building a reasoning machine

For some, simply focusing on the training portion won't cut it.

Former OpenAI chief scientist and Safe Superintelligence cofounder Ilya Sutskever told Reuters this month that results from scaling models in pre-training had plateaued and that "everyone is looking for the next thing."

That "next thing" looks to be reasoning. Industry attention has increasingly turned to an area of AI known as inference, which focuses on the ability of a trained model to respond to queries and information it might not have seen before with reasoning capabilities.

At Microsoft's Ignite event this month, the company's CEO Satya Nadella said that instead of seeing so-called AI scaling laws hit a wall, he was seeing the emergence of a new paradigm for "test-time compute," which is when a model has the ability to take longer to respond to more complex prompts from users. Nadella pointed to a new "think harder" feature for Copilot — Microsoft's AI agent — which boosts test time to "solve even harder problems."

Aymeric Zhuo, cofounder and CEO of AI startup Agemo, said that AI reasoning "has been an active area of research," particularly as "the industry faces a data wall." He told BI that improving reasoning requires increasing test-time or inference-time compute.

Typically, the longer a model takes to process a dataset, the more accurate the outcomes it generates. Right now, models are being queried in milliseconds. "It doesn't quite make sense," Sivesh Sukumar, an investor at investment firm Balderton, told BI. "If you think about how the human brain works, even the smartest people take time to come up with solutions to problems."

In September, OpenAI released a new model, o1, which tries to "think" about an issue before responding. One OpenAI employee, who asked not to be named, told BI that "reasoning from first principles" is not the forte of LLMs as they work based on "a statistical probability of which words come next," but if we "want them to think and solve novel problem areas, they have to reason."

Noam Brown, a researcher at OpenAI, thinks the impact of a model with greater reasoning capabilities can be extraordinary. "It turned out that having a bot think for just 20 seconds in a hand of poker got the same boosting performance as scaling up the model by 100,000x and training it for 100,000 times longer," he said during a talk at TED AI last month.

Google and OpenAI did not respond to a request for comment from Business Insider.

The AI boom meets its tipping point

These efforts give researchers reasons to remain hopeful, even if current signs point to a slower rate of performance leaps. As a separate former DeepMind employee who worked on Gemini told BI, people are constantly "trying to find all sorts of different kinds of improvements."

That said, the industry may need to adjust to a slower pace of improvement.

"I just think we went through this crazy period of the models getting better really fast, like, a year or two ago. It's never been like that before," the former DeepMind employee told BI. "I don't think the rate of improvement has been as fast this year, but I don't think that's like some slowdown."

Lamini's Zhou echoed this point. Scaling laws — an observation that AI models improve with size, more data, and greater computing power —work on a logarithmic scale rather than a linear one, she said. In other words, think of AI advances as a curve rather than a straight upward line on a graph. That makes development far more expensive "than we'd expect for the next substantive step in this technology," Zhou said.

She added: "That's why I think our expectations are just not going to be met at the timeline we want, but also why we'll be more surprised by capabilities when they do appear."

Amazon Web Services CEO Adam Selipsky speaks with Anthropic CEO Dario Amodei during a 2023 conference.

Companies will also need to consider how much more expensive it will be to create the next versions of their highly prized models. According to Anthropic's Amodei, a training run in the future could one day cost $100 billion. These costs include GPUs, energy needs, and data processing.

Whether investors and customers are willing to wait around longer for the superintelligence they've been promised remains to be seen. Issues with Microsoft's Copilot, for instance, are leading some customers to wonder if the much-hyped tool is worth the money.

For now, AI leaders maintain that there are plenty of levers to pull — from new data sources to a focus on inference — to ensure models continue improving. Investors and customers just might have to be prepared for them to come at a slower pace compared to the breakneck pace set by OpenAI when it launched ChatGPT two years ago.

Bigger problems lie ahead if they don't.

Read the original article on Business Insider
Москва

«МедАльянсГрупп» – надежный партнер в медицинской сфере

Michail Antonio reveals he was barred from entering the UK after passport blunder in nightmare international break

Sky Sports commentator stunned by ‘one of the strangest reactions to a goal I’ve ever seen’ by Watford fans

Exclusive: Sumit Kaul on joining the new season of Tenali Rama as Girgit; says ‘It will be a challenge for me to live up to the expectations of audience’

F1 Las Vegas Grand Prix – Start time, starting grid, how to watch, & more

Ria.city






Read also

Mum-of-four Coleen Rooney praised for being ‘so relatable’ after sharing how she deals with sons when they’re naughty

Two in three Nigerians affected by NTDs -FG

Shrinking Recap: Mic Drop

News, articles, comments, with a minute-by-minute update, now on Today24.pro

News Every Day

Sky Sports commentator stunned by ‘one of the strangest reactions to a goal I’ve ever seen’ by Watford fans

Today24.pro — latest news 24/7. You can add your news instantly now — here


News Every Day

Sky Sports commentator stunned by ‘one of the strangest reactions to a goal I’ve ever seen’ by Watford fans



Sports today


Новости тенниса
Елена Веснина

Елена Веснина рассказала, как впервые увидела Мирру Андрееву в девятилетнем возрасте



Спорт в России и мире
Москва

Певец из Москвы Лев Соловьев выпустил гимн медиафутбола и футбольной медиалиги



All sports news today





Sports in Russia today

Москва

Победа с минимальным отрывом: АРХИWOOD выиграл во II туре Лиги Чемпионов Бизнеса


Новости России

Game News

Five new Steam games you probably missed (November 25, 2024)


Russian.city


Москва

«МедАльянсГрупп» – надежный партнер в медицинской сфере


Губернаторы России
УФСИН

Фильмы осужденных посмотрели в псковских учреждениях УФСИН


Собянин и Голикова открыли четвертый центр женского здоровья

Юрист Черепанова объяснила штрафы за нарушение правил содержания животных

В квартирах ЖК «Квартал Триумфальный» стартовал монтаж радиаторов отопления

Филиал № 4 ОСФР по Москве и Московской области информирует: В Москве и Московской области 650 тысяч пенсионеров старше 80 лет получают пенсию в повышенном размере


Менеджер Песни. Менеджер Релиза Песни.

Релиз трека. Релиз новой песни. Релиз сингла. Релиз Музыкального альбома.

В Гнесинке прошел концерт ведущих казачьих коллективов

Кабинет Артиста. Яндекс кабинет артиста.


Миранчук с «Атлантой» выбыл из плей-офф МЛС, Синнер выиграл Кубок Дэвиса. Главное к утру

Динара Сафина назвала турниры Большого шлема, которые Даниил Медведев может выиграть в следующем сезоне

Зарина Дияс узнала хорошую новость от WTA

Елена Веснина: Синнер с итальянского переводится «грешник», но он играет как полубог



В Подмосковье офицер Росгвардии оказал помощь в эвакуации  пострадавших в результате ДТП

Сеть клиник «Будь Здоров» запустила медицинского GPT-ассистента

В Подмосковье росгвардейцы пришли на помощь пострадавшим в результате ДТП

Филиал № 4 ОСФР по Москве и Московской области информирует: Отделение СФР по Москве и Московской области оплатило свыше 243 тысяч дополнительных выходных дней по уходу за детьми с инвалидностью


С глаз долой, из сердца - вон: что делают россияне с подарками бывших

Собянин сообщил об открытии 13 обновленных поликлиник

Филиал № 4 ОСФР по Москве и Московской области информирует: В Москве и Московской области 650 тысяч пенсионеров старше 80 лет получают пенсию в повышенном размере

Нижегородское "Торпедо" разгромило на выезде московский "Спартак" в КХЛ


Московская биржа запустит торги мини- и микрофьючерсами на Brent и Henry Hub

Уже завтра пройдет церемония награждения AdIndex Awards

Семёнов Николай Олегович – Спортсмен, -блогер, тренер, предприниматель, актёр.

Суть народа. Журналисты страны обсудили в Москве традиционные ценности



Путин в России и мире






Персональные новости Russian.city
Концерт

Концерт ко Дню матери прошел во Дворце культуры «Звезда» в Наро-Фоминске



News Every Day

F1 Las Vegas Grand Prix – Start time, starting grid, how to watch, & more




Friends of Today24

Музыкальные новости

Персональные новости