Rising Competition in China’s AI Infra Market Set to Cut Costs for AI Models

"I am increasingly confident that if China wants to develop its own ecosystem and its own AI, it must have full control over the entire industry chain. After accumulating experience in optimizing AI hardware and software, we will push China to fully utilize its computing power in the age of large AI models and compete with the United States," said Professor Wang Yu, founder of Infinigence, department head of Tsinghua University Department of Electronic Engineering, in Mandarin and as translated by AsianFin.

TMTPOST -- In the fast-evolving AI landscape, the push for lower operational costs and more efficient systems is becoming crucial. At the heart of this effort lies AI infrastructure (AI Infra), a sector that is rapidly gaining importance in China’s burgeoning artificial intelligence market.

One of the most significant events in this sector occurred just before the World Artificial Intelligence Conference (WAIC) in July this year, when executives from two competing AI Infra companies took to the stage at a conference in Shanghai. Despite being competitors, both companies were focused on securing collaboration agreements with an AI chip company that could help them advance their goals.

The CEO of one of the companies, upon hearing that a competitor would also be attending, immediately reached out to the event organizers to secure an additional speaking slot. The CEO flew from Beijing to Shanghai to present the company’s vision, marking their first public address since its establishment.

This bold move paid off, as the company not only secured the collaboration with the target client but also successfully won another partnership agreement at a subsequent AI computation release event just three days later. This episode is a prime example of the fierce competition in China’s AI Infra industry, a sector that has grown to become integral to the AI ecosystem.

The Rise of AI Infra in China

AI Infra acts as a bridge between computing power and AI applications, providing critical infrastructure such as software systems, data storage and processing, and network facilities. Its goal is to address challenges such as the limitations on high-end AI computing power from U.S. firms, decoupling Chinese models from overseas GPU technology like Nvidia’s, and solving issues related to computing, storage, and communication networks.

As the demand for AI-powered applications continuously increases but China faces a severe shortage of available computing power, the AI Infra sector has emerged as an indispensable part of the country’s AI model development process.

One of the main catalysts for this growth is the strategic efforts of companies in China to reduce costs while increasing the efficiency of their AI model training and inference systems.

With the United States imposing restrictions on the export of high-performance AI chips like Nvidia’s, and the overall high cost of AI hardware, the AI Infra sector is rapidly expanding to address these issues. According to a report by CICC Securities, AI Infra is currently in its early stage but is expected to grow at a rate of more than 30% annually over the next three to five years. Meanwhile, KKR & Co. estimates that global investments in data centers could reach $250 billion annually, as demand for AI infrastructure continues to rise. 

The global AI market is poised for explosive growth. According to Sequoia Capital and Bain & Company, the AI market could surge to nearly $1 trillion by 2027, with the AI hardware and services market growing at an annual rate of 40% to 55%. 

Within this market, over $600 billion in investments are expected to flow into AI infrastructure. As the cost of training AI models skyrockets—sometimes reaching several billion dollars per model—there is a massive push to make AI infrastructure more cost-effective and accessible.

The Battle for Computing Power and Efficiency

Over the past 60 years, AI has revolutionized various sectors, including design, healthcare, and even the pricing of GPUs. With the new AI boom, computing power has become a critical strategic asset, a factor that determines national competitiveness.

The three pillars of AI development—chips, infrastructure, and data—are inextricably linked. Chips, such as CPUs, GPUs, and memory semiconductors, play a decisive role in determining the power of AI systems. Infrastructure, including 5G networks, data centers, cloud computing clusters, and supercomputers, drives the development of computing power. Data, in turn, is the ultimate measure of the value of computing power. 

The scaling laws of AI models indicate that as the amount of computing resources and data increases, the capabilities of the models grow exponentially. Over the last decade, the growth of computing resources and data has been nothing short of remarkable.

For instance, the GPT-1 model developed by OpenAI had 170 million parameters, while GPT-4 now has over a trillion parameters—a stunning increase. The computational cost of training these models is equally staggering. For example, training GPT-3.5 on Microsoft’s Azure AI supercomputing infrastructure required roughly 3,640 petaflop-days (the equivalent of one quadrillion calculations per second over 3,640 days).

As AI models become more complex, the demand for computing power surges. The costs associated with this are also climbing. A single chip produced by TSMC's 3nm process can cost over $20,000, while a set of eight Nvidia A100 GPUs can run as high as $25 million. In early 2023, ChatGPT used nearly 30,000 high-end Nvidia GPUs to handle millions of daily requests. The cost of building the underlying infrastructure for services like GPT-powered Bing AI was estimated at over $4 billion—more than the GDP of South Sudan.

As AI models require more and more resources, the cost of training large models has become a significant barrier. For example, ByteDance’s MegaScale infrastructure, designed for training models with 175 billion parameters, requires several days of training on a large cluster, with costs exceeding millions of dollars.

Despite these high costs, the AI community faces the challenge of improving the utilization efficiency of computing power, as the training efficiency of many AI models is often less than 60%.

The Role of AI Infra in Addressing These Challenges

The role of AI Infra is becoming more critical as AI models grow in size and complexity. As Shen Dou, the executive vice president of Baidu, has noted, the explosion of AI applications and the increasing demand for model training and inference require companies to focus on improving both computing power efficiency and reducing costs.

In China, the high cost of obtaining AI computing resources and the challenges associated with China’s national strategy of channeling computing resources from the east to the west have made it even more urgent to optimize the use of computing resources. This is where AI Infra comes in: by decoupling models from hardware, AI Infra helps ensure that computing resources are used more efficiently and that AI models can scale more rapidly without incurring prohibitive costs. 

Companies like SiliconFlow and Infinigence have introduced innovative AI Infra solutions. Infinigence, for example, has achieved a 90% increase in computational efficiency for one of its internet clients, significantly lowering the cost of running AI models. AI Infra solutions will allow companies to achieve scalable profits within the next three to five years, Xia Lixue, the co-founder and CEO of Infinigence, told AsianFin. 

The Commercialization of AI Infra 

Rapid progress is being made in the commercialization of AI Infra. Startups in this field are offering a range of solutions to address the growing demand for efficient AI model training and inference. These solutions include cloud-based AI platforms, software-as-a-service (SaaS) offerings, and end-to-end solutions that combine chips and software. Companies like SiliconFlow and Infinigence are leading the way in providing customized computing power solutions to meet the needs of developers and businesses. As AI models continue to grow in scale and complexity, the demand for efficient and cost-effective infrastructure solutions will only increase.

In contrast to established tech giants like Nvidia, which dominates the global AI infrastructure space with its A100 and H100 GPUs, China’s AI Infra companies are still in the early stages of development. However, they are catching up, and their ability to tailor solutions to local needs positions them as key players in the domestic AI landscape.

Nvidia CEO Jensen Huang recently admitted that since the birth of general-purpose computing 60 years ago, there has been a shift toward accelerated computing. Through parallel computing, GPU-based computing power has significantly surpassed that of CPUs.

The development of neural networks and deep learning has also accelerated the speed at which computers acquire knowledge, bringing about a leap in computer intelligence. He believes that traditional computing methods rely on pre-set algorithmic models, lacking the ability to learn and understand. However, by incorporating deep learning, systems can adjust and optimize data to enhance the utilization of computing power. 

Huang emphasized that computing technology advances a million times every decade, and within just two years, Nvidia and the entire industry will undergo profound changes. He described the future of AI as "incredible," believing that AI has narrowed the technological gap between humans, and in the next ten years, computing power will increase another million times.

"I am increasingly confident that if China wants to develop its own ecosystem and its own AI, it must have full control over the entire industry chain. After accumulating experience in optimizing AI hardware and software, we will push China to fully utilize its computing power in the age of large AI models and compete with the United States," said Professor Wang Yu, founder of Infinigence, department head of Tsinghua University Department of Electronic Engineering, in Mandarin and as translated by AsianFin.

As the competition heats up, the ability to reduce costs and enhance computational efficiency will be pivotal for AI Infra companies striving to support China’s rapidly expanding AI model ecosystem.

With a projected global investment in AI infrastructure of over $600 billion by 2027, the stakes are high. Companies that can deliver innovative, cost-effective solutions will have a significant role to play in the AI revolution.

转载请注明出处、作者和本文链接
声明:文章内容仅供参考、交流、学习、不构成投资建议。
想和千万钛媒体用户分享你的新奇观点和发现,点击这里投稿 。创业或融资寻求报道,点击这里

敬原创,有钛度,得赞赏

赞赏支持
发表评论
0 / 300

根据《网络安全法》实名制要求,请绑定手机号后发表评论

登录后输入评论内容

快报

更多

20:08

此前被传“失联”的王莉现身茅台说明会:2026年不设经营目标

20:07

华熙生物:控股股东拟近期履行捐赠协议,无偿捐赠1%公司股份

20:07

科大智能:拟1.51亿元收购源光电气70%股权

20:06

北京微纳星空科技股份有限公司科创板IPO获受理

20:05

5月11日新闻联播速览20条

20:02

Strategy于5月4日至5月10日期间以平均购买价80340美元购入535枚比特币

19:54

*ST正平:回复上交所监管工作函,称不存在故意损毁财务资料等情形

19:46

斯塔默:不会辞职

19:45

马克龙:从未考虑武力重开霍尔木兹海峡

19:43

央行:展望未来,人工智能行业将为经济高质量发展注入更多动能

19:32

摩根大通策略师提醒,美国股市高位之下潜藏通胀风险

19:32

*ST观典:收到股票终止上市决定,预计最后交易日期为6月8日

19:31

欧盟委员会对OpenAI开放新型ChatGPT模型访问权限表示欢迎

19:31

央行:对一些不合理的、容易削减货币政策传导的市场行为加强规范

19:31

央行:拓展丰富中央银行宏观审慎与金融稳定功能,创新金融工具,维护金融市场平稳运行

19:30

央行:推进人民币国际化,提升资本项目开放水平

19:29

央行:保持汇率弹性,发挥汇率调节宏观经济和国际收支自动稳定器功能

19:28

央行:引导金融机构提高利率定价能力,促进社会综合融资成本低位运行

19:28

央行:继续实施好适度宽松的货币政策,保持流动性充裕和社会融资条件相对宽松

19:25

贵州茅台回应是否会再次提价

扫描下载App