Friday, November 22, 2024
Google search engine

OpenAI and rivals search new path to smarter AI as present strategies hit limitations


(Reuters) – Artificial intelligence corporations like OpenAI are in search of to beat sudden delays and challenges within the pursuit of ever-bigger giant language fashions by growing coaching strategies that use extra human-like methods for algorithms to “think”.

A dozen AI scientists, researchers and buyers advised Reuters they consider that these strategies, that are behind OpenAI’s just lately launched o1 mannequin, might reshape the AI arms race, and have implications for the varieties of assets that AI corporations have an insatiable demand for, from vitality to varieties of chips.

OpenAI declined to remark for this story. After the discharge of the viral ChatGPT chatbot two years in the past, expertise corporations, whose valuations have benefited enormously from the AI increase, have publicly maintained that “scaling up” present fashions by including extra knowledge and computing energy will persistently result in improved AI fashions.

But now, among the most distinguished AI scientists are talking out on the restrictions of this “bigger is better” philosophy.

Ilya Sutskever, co-founder of AI labs Safe Superintelligence (SSI) and OpenAI, advised Reuters just lately that outcomes from scaling up pre-training – the part of coaching an AI mannequin that makes use of an enormous quantity of unlabeled knowledge to grasp language patterns and buildings – have plateaued.

Sutskever is extensively credited as an early advocate of attaining huge leaps in generative AI development by the usage of extra knowledge and computing energy in pre-training, which ultimately created ChatGPT. Sutskever left OpenAI earlier this yr to discovered SSI.

“The 2010s were the age of scaling, now we’re back in the age of wonder and discovery once again. Everyone is looking for the next thing,” Sutskever mentioned. “Scaling the right thing matters more now than ever.”

Sutskever declined to share extra particulars on how his staff is addressing the problem, apart from saying SSI is engaged on another strategy to scaling up pre-training.

Behind the scenes, researchers at main AI labs have been operating into delays and disappointing outcomes within the race to launch a big language mannequin that outperforms OpenAI’s GPT-4 mannequin, which is sort of two years previous, in accordance with three sources aware of non-public issues.

The so-called ‘training runs’ for big fashions can price tens of tens of millions of {dollars} by concurrently operating a whole bunch of chips. They usually tend to have hardware-induced failure given how difficult the system is; researchers could not know the eventual efficiency of the fashions till the tip of the run, which may take months.



Source link

- Advertisment -
Google search engine

Must Read