Moneycontrol PRO
HomeArtificial IntelligenceDeepSeek touts new training method as China pushes AI efficiency

DeepSeek touts new training method as China pushes AI efficiency

Such publications from DeepSeek have foreshadowed the release of major models in the past

January 02, 2026 / 06:59 IST
The document, co-authored by founder Liang Wenfeng, introduces a framework it called Manifold-Constrained Hyper-Connections
Snapshot AI
  • DeepSeek unveils efficient AI framework to boost scalability and cut energy use
  • New method addresses training instability and scalability in large AI models
  • Anticipation grows for DeepSeek's next flagship R2 model, expected in February

DeepSeek published a paper outlining a more efficient approach to developing AI, illustrating the Chinese artificial intelligence industry’s effort to compete with the likes of OpenAI despite a lack of free access to Nvidia Corp. chips.

The document, co-authored by founder Liang Wenfeng, introduces a framework it called Manifold-Constrained Hyper-Connections. It’s designed to improve scalability while reducing the computational and energy demands of training advanced AI systems, according to the authors.

Such publications from DeepSeek have foreshadowed the release of major models in the past. The Hangzhou-based startup stunned the industry with the R1 reasoning model a year ago, developed at a fraction of the cost of its Silicon Valley rivals. DeepSeek has since released several smaller platforms but anticipation is mounting for its next flagship system, widely dubbed the R2, expected around the Spring Festival in February.

Chinese startups continue to operate under significant constraints, with the US preventing access to the most advanced semiconductors essential to developing and running AI. Those restrictions have forced researchers to pursue unconventional methods and architectures.

DeepSeek, known for its unorthodox innovations, published its latest paper this week through the open repository arXiv and open-source platform Hugging Face. The paper lists 19 authors, with Liang’s name appearing last.

The founder, who’s consistently steered DeepSeek’s research agenda, has pushed his team to rethink how large-scale AI systems are conceived and built.

The latest research addresses challenges such as training instability and limited scalability, noting that the new method incorporates “rigorous infrastructure optimization to ensure efficiency.” Tests were conducted on models ranging from 3 billion to 27 billion parameters, building on ByteDance Ltd.’s 2024 research into hyper-connection architectures.

The technique holds promise “for the evolution of foundational models,” the authors said.

Bloomberg
first published: Jan 2, 2026 06:59 am

Discover the latest Business News, Sensex, and Nifty updates. Obtain Personal Finance insights, tax queries, and expert opinions on Moneycontrol or download the Moneycontrol App to stay updated!

Subscribe to Tech Newsletters

  • On Saturdays

    Find the best of Al News in one place, specially curated for you every weekend.

  • Daily-Weekdays

    Stay on top of the latest tech trends and biggest startup news.

Advisory Alert: It has come to our attention that certain individuals are representing themselves as affiliates of Moneycontrol and soliciting funds on the false promise of assured returns on their investments. We wish to reiterate that Moneycontrol does not solicit funds from investors and neither does it promise any assured returns. In case you are approached by anyone making such claims, please write to us at grievanceofficer@nw18.com or call on 02268882347