Tech »  Topic »  Forget data labeling: Tencent’s R-Zero shows how LLMs can train themselves

Forget data labeling: Tencent’s R-Zero shows how LLMs can train themselves


Image credit: VentureBeat with ChatGPT

A new training framework developed by researchers at Tencent AI Lab and Washington University in St. Louis enables large language models (LLMs) to improve themselves without requiring any human-labeled data. The technique, called R-Zero, uses reinforcement learning to generate its own training data from scratch, addressing one of the main bottlenecks in creating self-evolving AI systems. R-Zero works by having two independent models co-evolve by interacting with and challenging each other.

Experiments show that R-Zero substantially improves reasoning capabilities across different LLMs, which could lower the complexity and costs of training advanced AI. For enterprises, this approach could accelerate the development of specialized models for complex reasoning tasks without the massive expense of curating labeled datasets.

The challenge of self-evolving LLMs

The idea behind self-evolving LLMs is to create AI systems that can autonomously generate, refine, and learn from their own experiences. This offers a ...


Copyright of this story solely belongs to venturebeat . To see the full text click HERE