CES文章详情顶部

Infinigence Unveils Next-Gen AI Infrastructure Suite, Aims to Lead China's AI Deployment Revolution

When discussing the resumption of H20 sales and its impact on the development of domestic chips, Xia Lixue stated that it still depends on long-term value. H20 itself is not the most cutting-edge technology; whether in training or inference, it is already an optimized version. Therefore, at least for the domestic industry, there will definitely be short-term fluctuations this time, but the long-term impact will not be that significant.

WuWenXinQiong Co-founder and CEO Xia Lixue

Xia Lixue, Co-founder and CEO of Infinigence

TMTPOST -- Infinigence, an AI infrastructure startup backed by Tsinghua University, introduced a sweeping portfolio of performance-optimized computing platforms targeting the full spectrum of AI deployment at this year’s World Artificial Intelligence Conference (WAIC 2025) .

The company officially launched three flagship products under its integrated solution suite: Infinicloud, a global-scale AI cloud platform for clusters of up to 100,000 GPUs; InfiniCore, a high-performance intelligent computing platform designed for multi-thousand-GPU clusters; and InfiniEdge, a lean, edge computing solution optimized for terminal deployments with as few as one GPU.

Together, the platforms represent what CEO Xia Lixue calls a “software-hardware co-designed infrastructure system for the AI 2.0 era.” Built for compatibility across heterogeneous computing environments, the Infinigence stack offers full lifecycle support—from model scheduling and performance optimization to large-scale application deployment.

“We’re addressing a core bottleneck in China’s AI industry: fragmentation in compute infrastructure,” Xia said. “With InfiniCloud, InfiniCore, and InfiniEdge, we’re enabling AI developers to move seamlessly between different chips, architectures, and workloads—unlocking intelligent performance at scale.”

In a fast-evolving AI landscape dominated by open-source large language models such as DeepSeek, GLM-4.5, and MiniMax M1, Chinese infra startups are racing to build the backbone that powers model deployment and inference.

Early on July 29, Infinigence announced that InfiniCloud now supports Zhipu AI’s latest GLM-4.5 and GLM-4.5-air models, which currently rank third globally in performance. The move signals Infinigence’s ambition to anchor the growing synergy between Chinese model developers and domestic chipmakers.

Xia likened the trio of newly launched platforms to “three bundled boxes” that can be matched to AI workloads of any scale. “From a single smartphone to clusters of 100,000 GPUs—our system is designed to ensure resource efficiency and intelligent elasticity,” he said.

Infinigence’s platforms are already powering Shanghai ModelSpeed Space, the world’s largest AI incubator. The facility sees daily token call volumes exceed 10 billion, supports over 100 AI use cases, and reaches tens of millions of monthly active users across its applications.

A key challenge for China’s AI infrastructure sector is hardware heterogeneity. With dozens of domestic chip vendors and proprietary architectures, developers often struggle to port models across systems.

Xia emphasized that Infinigence has developed a “universal compute language” that bridges chips with disparate instruction sets. “We treat computing resources like supermarket goods—plug-and-play, interoperable, and composable,” he said.

The company’s infrastructure has already achieved full-stack adaptation for more than a dozen domestic chips, delivering 50%–200% performance gains through algorithm and compiler optimization. It also supports unified scheduling and mixed-precision computing, enabling cost-performance ratios that beat many international offerings.

“What’s missing in China’s ecosystem is a feedback loop,” Xia said. “In the U.S., NVIDIA and OpenAI form a tight cycle: model developers know what chips are coming, and chipmakers know what models are being built. We’re building that loop domestically.”

Infinigence is also targeting AI democratization with a first-of-its-kind cross-regional federated reinforcement learning system. The system links idle GPU resources from different regional AIDC centers into a unified compute cluster—allowing SMEs to build and fine-tune domain-specific inference models using consumer-grade cards.

To support this, Infinigence launched the “AIDC Joint Operations Innovation Ecosystem Initiative” in partnership with China’s three major telecom providers and 20+ AIDC institutions.

Xia noted that while training still depends heavily on NVIDIA hardware, inference workloads are rapidly migrating to domestic accelerators. “Users often start with international chips on our platform, but we help them transition to Chinese cards—many of which now deliver strong commercial value,” he said.

Infinigence has also rolled out a series of on-device and edge inference engines under its Infini-Ask line. These include:

  • Infini-Megrez2.0, co-developed with the Shanghai Institute of Creative Intelligence, the world’s first on-device intrinsic model.

  • Infini-Mizar2.0, built with Lenovo, which enables heterogeneous computing across AI PCs, boosting local model capacity from 7B to 30B parameters.

  • A low-cost FPGA-based large model inference engine, jointly developed with Suzhou Yige Technology.

Founded in May 2023, Infinigence has raised more than RMB 1 billion in just two years, including a record-setting RMB 500 million Series A round in 2024—the largest to date in China’s AI infrastructure sector.

Its product portfolio now spans everything from model hosting and cloud management to edge optimization and model migration—serving clients across intelligent computing centers, model providers, and industrial sectors.

The company’s broader mission, Xia said, is to balance scale, performance, and resource availability. “Our vision is to deliver ‘boundless intelligence and flawless computing’—wherever there's compute, we want Infinigence to be the intelligence that flows through it.”

IEEE Fellow and Tsinghua professor Wang Yu, also a co-founder of Infinigence, argued that the future of China’s AI economy depends on interdisciplinary collaboration. “We need people who understand chips, models, commercialization, and investment,” Wang said. “Only then can we solve the ‘last mile’ problem—connecting AI research with real-world deployment.”

As China looks to decouple from foreign hardware dependence while competing globally in next-gen AI, Infinigence is positioning itself as a vital enabler—fusing chip-level control with cloud-scale ambition.

“Every AI system runs on two forces: models and compute,” Xia said. “They cannot evolve in silos—they must move forward in sync.”

本文系作者 zhangxinyue 授权钛媒体发表,并经钛媒体编辑,转载请注明出处、作者和本文链接
本内容来源于钛媒体钛度号,文章内容仅供参考、交流、学习,不构成投资建议。
想和千万钛媒体用户分享你的新奇观点和发现,点击这里投稿 。创业或融资寻求报道,点击这里

敬原创,有钛度,得赞赏

赞赏支持
发表评论
0 / 300

根据《网络安全法》实名制要求,请绑定手机号后发表评论

登录后输入评论内容

快报

更多

13:50

现货黄金失守4310美元/盎司,日内跌0.67%

13:34

铂期货主力合约日内走低13%,报516.7元/克

13:30

沪深两市成交额超1.5万亿元,较上日此时缩量497亿元

13:17

可控核聚变概念局部异动,王子新材直线涨停

13:13

商业航天概念再度拉升,中国卫通4天2板续创历史新高

13:11

全国政协举行新年茶话会 习近平发表重要讲话

13:05

数字货币概念震荡走强,拉卡拉涨超10%

13:04

富时中国A50指数期货下跌超1%

13:00

现货白银日内跌6%,现报71.58美元/盎司

12:59

商务部:各地可在消费品以旧换新政策框架内,自主实施补贴政策

12:51

新加坡股市收低0.2%,2025年累计上涨23%

12:45

预告:国新办1月6日就推进绿色消费有关情况举行新闻发布会

12:31

商务部等五部门:鼓励家电销售企业联合生产、回收企业开展以旧换新促销活动,对消费者购新并交旧给予适当优惠

12:20

港股25年收官:恒生指数全年涨27.77%,恒生科技指数全年涨23.45%

12:16

五部门:引导家电、手机等品牌企业健全合理的价格体系,防止出现“内卷式”竞争

12:09

工行、农行:2026年1月1日起为数字人民币实名钱包余额计付利息

12:08

五部门:各地要科学制定补贴资金使用及兑付计划,减轻政策参与经营主体垫资压力

12:04

五部门:2026年家电产品每件补贴不超过1500元,数码和智能产品每件补贴不超过500元

12:02

国务院任免国家工作人员:任命林泽昌为财政部副部长

11:57

商务部:今年1-11月服务进出口总额同比增长7.1%

扫描下载App