Tech »  Topic »  How DeepSeek's new way to train advanced AI models could disrupt everything - again

How DeepSeek's new way to train advanced AI models could disrupt everything - again


The Chinese AI lab may have just found an approach to training frontier LLMs that's both practical and scalable, even for more cash-strapped developers.

Flavio Coelho/ Moment via Getty

Follow ZDNET: Add us as a preferred source on Google.

ZDNET's key takeaways

  • DeepSeek debuted Manifold-Constrained Hyper-Connections, or mHCs.
  • They offer a way to scale LLMs without incurring huge costs.
  • The company postponed the release of its R2 model in mid-2025.

Just before the start of the new year, the AI world was introduced to a potential game-changing new method for training advanced models.

A team of researchers from Chinese AI firm DeepSeek released a paper on Wednesday outlining what it called Manifold-Constrained Hyper-Connections, or mHC for short, which may provide a pathway for engineers to build and scale large language models without the huge computational costs that are typically required.

Also: Is DeepSeek's new model ...


Copyright of this story solely belongs to zdnet.com . To see the full text click HERE