Tech »  Topic »  Beyond GPT architecture: Why Google’s Diffusion approach could reshape LLM deployment

Beyond GPT architecture: Why Google’s Diffusion approach could reshape LLM deployment


Last month, along with a comprehensive suite of new AI tools and innovations, Google DeepMind unveiled Gemini Diffusion. This experimental research model uses a diffusion-based approach to generate text. Traditionally, large language models (LLMs) like GPT and Gemini itself have relied on autoregression, a step-by-step approach where each word is generated based on the previous one. Diffusion language models (DLMs), also known as diffusion-based large language models (dLLMs), leverage a method more commonly seen in image generation, starting with random noise and gradually refining it into a coherent output. This approach dramatically increases generation speed and can improve coherency and consistency. 

Gemini Diffusion is currently available as an experimental demo; sign up for the waitlist here to get access

(Editor’s note: We’ll be unpacking paradigm shifts like diffusion-based language models—and what it takes to run them in production—at VB Transform, June 24–25 in San Francisco ...


Copyright of this story solely belongs to venturebeat . To see the full text click HERE