Friday, November 22, 2024
Google search engine

OpenAI and rivals search new path to smarter AI as present strategies hit limitations


(Reuters) – Artificial intelligence firms like OpenAI are searching for to beat surprising delays and challenges within the pursuit of ever-bigger massive language fashions by growing coaching methods that use extra human-like methods for algorithms to “think”.

A dozen AI scientists, researchers and buyers instructed Reuters they imagine that these methods, that are behind OpenAI’s not too long ago launched o1 mannequin, might reshape the AI arms race, and have implications for the sorts of assets that AI firms have an insatiable demand for, from vitality to sorts of chips.

OpenAI declined to remark for this story. After the discharge of the viral ChatGPT chatbot two years in the past, expertise firms, whose valuations have benefited vastly from the AI growth, have publicly maintained that “scaling up” present fashions by means of including extra information and computing energy will persistently result in improved AI fashions.

But now, a number of the most distinguished AI scientists are talking out on the constraints of this “bigger is better” philosophy.

Ilya Sutskever, co-founder of AI labs Safe Superintelligence (SSI) and OpenAI, instructed Reuters not too long ago that outcomes from scaling up pre-training – the part of coaching an AI mannequin that makes use of an enormous quantity of unlabeled information to know language patterns and buildings – have plateaued.

Sutskever is broadly credited as an early advocate of attaining huge leaps in generative AI development by means of the usage of extra information and computing energy in pre-training, which ultimately created ChatGPT. Sutskever left OpenAI earlier this 12 months to discovered SSI.

“The 2010s were the age of scaling, now we’re back in the age of wonder and discovery once again. Everyone is looking for the next thing,” Sutskever mentioned. “Scaling the right thing matters more now than ever.”

Sutskever declined to share extra particulars on how his crew is addressing the problem, aside from saying SSI is engaged on another method to scaling up pre-training.

Behind the scenes, researchers at main AI labs have been working into delays and disappointing outcomes within the race to launch a big language mannequin that outperforms OpenAI’s GPT-4 mannequin, which is sort of two years previous, in keeping with three sources aware of personal issues.

The so-called ‘training runs’ for big fashions can value tens of hundreds of thousands of {dollars} by concurrently working lots of of chips. They usually tend to have hardware-induced failure given how sophisticated the system is; researchers might not know the eventual efficiency of the fashions till the tip of the run, which might take months.



Source link

- Advertisment -
Google search engine

Must Read