How DeepSeek's new way to train advanced AI models could disrupt everything - again
zdnet.comThe Chinese AI lab may have just found an approach to training frontier LLMs that's both practical and scalable, even for more cash-strapped developers.

Follow ZDNET: Add us as a preferred source on Google.
ZDNET's key takeaways
- DeepSeek debuted Manifold-Constrained Hyper-Connections, or mHCs.
- They offer a way to scale LLMs without incurring huge costs.
- The company postponed the release of its R2 model in mid-2025.
Just before the start of the new year, the AI world was introduced to a potential game-changing new method for training advanced models.
A team of researchers from Chinese AI firm DeepSeek released a paper on Wednesday outlining what it called Manifold-Constrained Hyper-Connections, or mHC for short, which may provide a pathway for engineers to build and scale large language models without the huge computational costs that are typically required.
Also: Is DeepSeek's new model ...
Copyright of this story solely belongs to zdnet.com . To see the full text click HERE

