News Report Technology

A member of the OpenAI team has indicated that the large language models of today may soon become notably smaller.

In a thought-provoking address lasting 45 minutes, AI expert Hyung Won Chung, who previously authored the significant Google paper 'Scaling Instruction-Finetuned Language Models,' explored the transformative realm of large language models in 2023. His insight underlines how these models can be fine-tuned to better adhere to instructions. Chung describes the landscape of large language models as exceptionally fluid. In this evolving domain, the guiding concepts are continuously shifting, unlike traditional fields where foundational principles often remain fixed. With advancements, what is deemed impossible today might become achievable tomorrow. He stresses the importance of framing claims regarding LLM capabilities with a clause like 'for now,' indicating that a model might still be capable of performing a task, even if it hasn't yet demonstrated that ability. The expansive models we utilize today are anticipated to transition into smaller versions within just a few years.

Hyung Won Chung, OpenAI

An essential takeaway from Chung's presentation is the critical need for thorough documentation and reproducibility in research.

Chung's talk delves into the complexities surrounding data and model parallelism. This section offers valuable insights for anyone keen to explore the technical depths of AI, emphasizing the importance of understanding these parallelism techniques for optimizing performance.

Hyung Won Chung, OpenAI

Chung argues that the Maximum Likelihood method, currently employed for pre-training large language models, acts as a significant limitation to achieving truly expansive capabilities, like scaling to 10,000 times what GPT-4 can handle. As machine learning evolves, relying on manually devised loss functions may become increasingly constraining. AI research He envisions a future where AI development's next phase involves algorithmically learned functions separate from traditional methods. Although still in its early stages, this approach could unlock scalability that overcomes current limitations. He also highlights ongoing projects like Reinforcement Learning from Human Feedback (RLHF) with Rule Modeling as progress in this direction, yet acknowledges that challenges still need addressing.

It’s important to understand that the information presented here should not be construed as financial, legal, investment, or similar types of advice. Always invest what you can afford to lose and consider seeking independent financial counsel if you're uncertain. We recommend reviewing the terms and conditions along with any support resources provided by the issuer or advertiser. MetaversePost strives for accuracy and impartiality in reporting, keeping in mind that market trends can change rapidly. large-scale model training .

Damir leads the team and functions as a product manager and editor at Metaverse Post, where he covers a range of topics including AI, machine learning, AGI, large language models, the Metaverse, and Web3. His engaging articles reach a vast audience, drawing over a million readers every month. With a decade of expertise in SEO and digital marketing, Damir is recognized as an authority, having been featured in renowned publications like Mashable, Wired, and Cointelegraph. Traveling as a digital nomad among the UAE, Turkey, Russia, and the CIS, he attributes his physics degree to equipping him with the critical thinking skills essential for navigating the dynamic world of the internet.

Blum celebrates the one-year milestone with accolades for 'Best GameFi App' and 'Best Trading App' at the Blockchain Forum 2025.

Disclaimer

In line with the Trust Project guidelines Omniston tackles the challenge of DeFi fragmentation by enhancing liquidity on the TON network.

Copyright, Permissions, and Linking Policy

A researcher from OpenAI suggests that today's large language models will soon be considered small by future standards, according to Metaverse Post.

Know More

Hyung Won Chung, a distinguished AI researcher with a notable history at Google Brain and currently part of the OpenAI team, delivered an insightful 45-minute presentation where he discussed the landscape of large language models as we move into 2023.

The landscape of large language models is shifting, as highlighted by a researcher from OpenAI.

Know More
Read More
Read more
News Report Technology
The Solv Protocol has partnered with Fragmetric and Zeus Network to launch FragBTC, a yield-generating Bitcoin product native to Solana.
News Report Technology
Polygon has initiated the 'Agglayer Breakout Program' to spur innovation and provide airdrop benefits to POL token holders.
Press Releases Business Markets Technology
Exploring the trajectory from Ripple to The Big Green DAO, we observe how cryptocurrency projects are contributing to charitable initiatives.
News Report Technology
Let's examine various efforts leveraging the potential of digital currencies for philanthropic endeavors.