Today's AI/ML headlines are brought to you by ThreatPerspective

Biz & IT Ars Technica

ChatGPT’s success could have come sooner, says former Google AI researcher

A co-author of Attention Is All You Need reflects on ChatGPT's surprise and Google's conservatism.
In 2017, eight machine-learning researchers at Google released a groundbreaking research paper called Attention Is All You Need, which introduced the Transformer AI architecture that underpins almost all of today's high-profile generative AI models.

The Transformer has made a key component of the modern AI boom possible by translating (or transforming, if you will) input chunks of data called "tokens" into another desired form of output using a neural network. Variations of the Transformer architecture power language models like GPT-4o (and ChatGPT), audio synthesis models that run Google's NotebookLM and OpenAI's Advanced Voice Mode, video synthesis models like Sora, and image synthesis models like Midjourney.

At TED AI 2024 in October, one of those eight researchers, Jakob Uszkoreit, spoke with Ars Technica about the development of transformers, Google's early work on large language models, and his new venture in biological computing.

Read full article

Comments


Published: 2024-11-14T17:00:52











© Digital Event Horizon . All rights reserved.

Privacy | Terms of Use | Contact Us