The Complete Record of the AI Cloud Battle Among ByteDance, Alibaba, and Baidu: An Investment in the Future

02/24 2026 516

The Spring Festival Gala is just the beginning; the AI cloud war has only just started.

When the 2026 CCTV Spring Festival Gala live broadcast signal lit up for the first second, it not only marked the end of the old year but also signaled the start of China's AI cloud era.

This was no ordinary technical support for a gala event. According to official disclosures, Volcano Engine was the exclusive AI cloud partner for the 2026 Spring Festival Gala, providing full technical support throughout the event, covering core scenarios such as ultra-high-definition live broadcasting, real-time interactive scheduling, and massive concurrent user processing. This meant that the smooth presentation of live broadcasts and the interaction requests of hundreds of millions of viewers during the gala relied entirely on its AI cloud technology.

For Volcano Engine, this was a showcase of its capabilities to the entire industry. "Only cloud services that can withstand the triple challenges of traffic, stability, and security at the scale of the Spring Festival Gala are qualified to compete in the core battleground of the AI era," an industry insider told Xinmou.

Thousands of miles away, at Alibaba's Xixi Park in Hangzhou, Alibaba Cloud Intelligence was equally busy. Although it did not secure the exclusive AI cloud partnership for the Spring Festival Gala, it still reserved ample computing power for billions of AI calls across Qianwen App, Taobao, Alipay, and Fliggy during the Spring Festival period. Just a month earlier, according to LatePost, Alibaba was considering increasing its planned RMB 380 billion investment in AI infrastructure and cloud computing over the next three years to RMB 480 billion.

As for Baidu Intelligent Cloud, the Spring Festival interactive features of Wenxin Yiyan, cloud service scheduling for intelligent driving, and the AI computing power demands of Baidu Search all rested on its clusters. Shen Dou, Executive Vice President of Baidu Group and President of Baidu Intelligent Cloud Business Group, once stated, "In the era of the intelligent economy, new infrastructure is essential for support, and that is the AI-first cloud."

Over the past two decades, China's internet giants have fought key battles in e-commerce, social media, short videos, and payments. But this time is different—previous wars were over traffic entry points; this time, it's about the underlying infrastructure of the AI era. If large models are the engines of the AI era, then AI cloud is the runway and airport that supports them. Microsoft leveraged its deep integration with OpenAI through Azure to breach AWS and Google's defenses in the global cloud market; Google redefined cloud service standards in the AI era with its full-stack capabilities of Gemini + TPU + Google Cloud.

In China, the three giants with the most core large model capabilities—ByteDance, Alibaba, and Baidu—are now engaged in a battle in the AI cloud arena that will determine the landscape of the next decade. This is not a local skirmish; losing here could mean losing entry tickets to the entire AI era.

01

The War Truly Begins After the Turning Point of Large Models

The real trigger for China's AI cloud war was not the launch of ChatGPT in 2023 but the full-scale eruption of the large model industry in 2025.

When the large model craze first emerged in 2023, cloud services were merely "computing power suppliers" for large model companies—startup and giant large model teams purchased GPU resources from cloud providers for training and inference, with a relationship more like that of a client and a vendor. The industry consensus at the time was that the core competition of large models lay in algorithms, with cloud services being just supporting infrastructure.

But everything changed in 2025. The advent of DeepSeek-R1 directly shifted the competition of large models from laboratory benchmarks to real-world scenarios for C-end users; with Doubao surpassing 100 million daily active users, Yuanbao's user base growing tenfold, and Qianwen App undergoing rapid iterations, large models transformed from a technical concept into products used daily by hundreds of millions of users.

This brought about disruptive demands for cloud services. An industry veteran calculated: A large model product with over 100 million daily active users processes more than 50 trillion tokens daily, requiring tens of thousands of GPU cards for inference alone each day; a full-scale pre-training of a large model with hundreds of billions of parameters requires thousands of cards running continuously for months, with computing power costs reaching hundreds of millions. More importantly, the upper limit of a large model's capabilities depends not only on algorithms but also on underlying computing power scheduling, network optimization, and inference acceleration—capabilities entirely controlled by cloud providers.

The industry logic completely reversed: Previously, large models needed the cloud; now, the cloud must bind with large models. Cloud providers without large model capabilities can only serve as low-margin "movers" of computing power; large models without proprietary cloud services will always be constrained by computing power and unable to deploy their capabilities to a vast number of enterprise clients.

Microsoft Azure's success set an example for domestic giants. By the end of 2025, Azure exceeded expectations, with AI clients contributing significant non-AI revenue, and the computing power shortage was expected to persist into 2026. The key was its deep integration with OpenAI—from dedicated clusters for model training to optimized inference architectures and Copilot services for enterprise clients, cloud and large models were fully integrated.

Google followed the same path, supporting Gemini's training and inference with its self-developed TPU chips and delivering a full-stack capability of "chips + models + cloud" to enterprise clients through Google Cloud. According to Gartner, Google Cloud's global market share rose to 13% in 2025, outpacing AWS in growth, with core increments also coming from AI-related services.

Against this backdrop, ByteDance, Alibaba, and Baidu—three domestic giants with full-stack capabilities in "chips-large models-cloud services"—made AI cloud their core battleground in 2026. The three chose entirely different paths, just as they did when large models first emerged three years ago:

Alibaba bet on a full-stack closed loop from the start, treating the "TongYunGe" strategy—integrating Tongyi Lab, Alibaba Cloud, and T-Head—as its core competitiveness, attempting to replicate Google's full-stack story;

Baidu was the earliest domestic player to layout (layout) AI-native clouds, using a combination of the PaddlePaddle deep learning framework, Wenxin large models, and self-developed Kunlun chips to deeply cultivate scenarios in government, enterprises, industry, and intelligent transportation, following a route of "technological precipitate + scenario landing";

ByteDance's Volcano Engine started latest but with the strongest momentum, leveraging extreme scenario validation within ByteDance—Doubao's daily processing of 63 trillion tokens, over 200% growth in six months, and the traffic tests of Douyin's 800 million daily active users—to replicate its internally refined capabilities externally.

The market landscape is being rewritten accordingly. Before 2025, the domestic cloud market was led by Alibaba Cloud, followed closely by Huawei Cloud and Tencent Cloud, with Baidu Intelligent Cloud and Volcano Engine ranking fourth and fifth. However, after 2025, AI cloud became the sole growth engine, with cloud providers lacking large model capabilities seeing slowed growth; the three with core large models rapidly reshaped the market landscape.

Among them, Volcano Engine's growth was the most astonishing. According to an Omdia report, as of October 2025, Volcano Engine held a 15% market share in the 2025 global enterprise-grade MaaS (Model as a Service) market, indicating its mature capabilities in serving B-end application scenarios.

It should be noted that MaaS is not the entirety of AI cloud and even accounts for only a small portion of the entire public cloud market's revenue. However, this does not prevent Alibaba from focusing its attention on AI cloud. In Liu Weiguang's view, Senior Vice President of Alibaba Cloud Intelligence Group and President of the Public Cloud Business Unit, even though MaaS has significant growth potential, the key is to win the entire new market of AI cloud, establish a full-stack AI cloud capability integrating software and hardware, and enable enterprises to leverage stronger AI models at lower costs—this is the decisive factor in competition.

02 The Frontline Battle: Full-Stack Capabilities, Customer Acquisition, and Computing Power Arms Race

The first direct clash among the three was the computing power arms race.

For AI cloud, computing power is the most critical ammunition. According to previous reports by LatePost, Alibaba is one of the companies with the largest GPU reserves in China, purchasing tens of thousands of high-end GPUs from overseas in 2025 alone, even buying large quantities of consumer-grade graphics cards like the RTX 4090 to build inference clusters and supplement inference throughput.

Alibaba's confidence comes from its self-developed chip layout (layout). T-Head's self-developed Zhenwu 810E AI chip has become one of the main chips in China's newly added AI computing power market, securing its first external major client order in 2025; in January 2026, sources close to the market said Alibaba decided to support T-Head's future independent listing, aiming to further strengthen its chip capabilities and provide underlying support for Alibaba Cloud's full-stack strategy.

ByteDance's Volcano Engine follows the route of "internal refinement, external output." ByteDance's internal computing power demands serve as the best training ground for Volcano Engine—Doubao's daily processing of 63 trillion tokens, equivalent to handling billions of user requests daily, imposes far higher requirements on computing power scheduling, inference optimization, and disaster recovery backup than most enterprise clients. Additionally, securing the AI cloud partnership for the 2026 Spring Festival Gala essentially proves to the entire industry that Volcano Engine's capabilities can withstand the most extreme traffic tests in China.

More noteworthy is ByteDance's global layout . According to relevant media reports, ByteDance will fully accelerate the globalization of its AI business in 2026, with Southeast Asia as the core focus, and Volcano Engine serving as the infrastructure for ByteDance's AI globalization. In other words, Volcano Engine must not only support the growth of Doubao's overseas version, Dola, but also serve a large number of Chinese enterprises going global, following the path Microsoft Azure took to rise in the global market.

Baidu Intelligent Cloud was the earliest domestic player to layout (layout) AI computing power. Baidu established the Institute of Deep Learning in 2013, becoming the first Chinese internet company to elevate deep learning to a core technological innovation; it launched China's first deep learning platform, PaddlePaddle, in 2016; announced the self-developed China's first cloud-based full-feature AI chip, "Kunlun," in 2018; released the world's first knowledge-enhanced hundred-billion-parameter large model in 2019; and in 2025, Baidu Intelligent Cloud lit up China's first fully self-developed 30,000-card Kunlun chip cluster, capable of simultaneously supporting the training of multiple hundred-billion-parameter large models, marking an important milestone in the construction of domestic AI computing power clusters.

Baidu's core advantage lies in deep scenario integration. Baidu Intelligent Cloud's AI capabilities are deeply embedded in all of Baidu's core businesses: the AI computing power demands of Search, user services of Wenxin Yiyan, cloud-edge-end collaboration for intelligent driving, and city-level implementation of ACE Intelligent Transportation are all supported by Baidu Intelligent Cloud. Especially in government, enterprise, industrial internet, and intelligent transportation, Baidu Intelligent Cloud has secured many city-level intelligent transportation projects in China, with the stable computing power demands from these projects forming its foundation.

Behind the computing power arms race is the competition for customers. Since 2025, the AI-related computing power demands of domestic internet companies, startups, and government-enterprise clients have exploded, becoming the core battleground for the three.

Alibaba Cloud's advantage lies in China's most comprehensive cloud service ecosystem and over a decade of enterprise client accumulation. Volcano Engine's breakthrough lies in internet clients and overseas clients. ByteDance itself is one of China's largest internet companies, with a far deeper understanding of internet industry demands than other cloud providers, while Baidu Intelligent Cloud firmly holds the foundation in government-enterprise and industrial sectors.

03 The Invisible Battlefield: Organizational Collaboration and Departmental Rivalry

The AI cloud war among the giants is not just about computing power and customers but also about internal organizational collaboration capabilities. The deep integration of cloud and large models essentially requires breaking down internal departmental silos within companies to achieve full-process collaboration from chips, infrastructure, models, to products—precisely the most challenging issue for large companies to solve.

Alibaba clearly recognizes this issue. Alibaba formally proposed the "TongYunGe" trinity strategy, fully integrating Tongyi Lab, Alibaba Cloud, and T-Head. The underlying logic is that previously, Tongyi Lab belonged to DAMO Academy, T-Head was an independent chip company, and Alibaba Cloud was an independent business group, with certain departmental silos among them, often working independently and struggling to form synergies.

After the "TongYunGe" strategy was implemented, it largely achieved unified planning of technological routes and unified allocation of resources. For example, the training of Tongyi's large models, from chip selection and cluster construction to training frameworks and inference optimization, was jointly designed by T-Head, Alibaba Cloud, and Tongyi Lab throughout the process, breaking down previous departmental barriers.

ByteDance's Volcano Engine follows a different collaboration route. ByteDance's large model team, the Seed department, has been deeply integrated with Volcano Engine from the start—all computing power services for Doubao are provided by Volcano Engine, with joint design and refinement in every detail from the underlying architecture of model training to inference optimization.

In February 2025, Wu Yonghui, former Vice President of Google DeepMind, joined ByteDance to lead the theoretical research of large models in the Seed department. His first move was to break down barriers between model departments and teams, achieving data sharing across all links and teams, with the most crucial aspect being integration with Volcano Engine's infrastructure team.

Baidu Intelligent Cloud was the earliest domestic player to achieve "cloud-model integration." Its technology R&D system underwent major reforms, with the Technology Platform Group (TPG) responsible for Wenxin's large model R&D split into a Basic Model R&D Department and an Application Model R&D Department, both reporting directly to Robin Li. This adjustment marked Baidu's shift from "basic R&D priority" to "application-driven," pushing backbone (backbone) talents skilled in technological implementation to the frontlines to accelerate the transition of large models from laboratories to industrial scenarios.

An internal source told Xinmou that Baidu's advantage lies in being an AI-native company from the start—PaddlePaddle, Wenxin, and Intelligent Cloud were all built around AI implementation, without much historical baggage. The difference is, "Alibaba Cloud started with the cloud and then moved to large models; we started with AI technology and then built the cloud, so the synergy between cloud and models is naturally smoother."

However, both ByteDance, Alibaba, and Baidu must face internal resource allocation issues. Especially since 2025, the incubation and cultivation of new businesses have required substantial computing power and technological resources, inevitably leading to resource competition with existing business foundations, which tests the strategic decisions and judgment of their leaders.

04 Rebuilding Infrastructure: Will AI Cloud Make the World More Open or More Closed?

In 2006, Amazon launched AWS, ushering in the cloud computing era. The industry's vision at the time was that cloud computing would turn computing power into a public infrastructure like water and electricity, allowing anyone or any company to access powerful computing capabilities at extremely low costs, breaking down the technological barriers of giants.

Two decades later, cloud computing has indeed transformed the entire tech industry, but it has not broken down barriers; instead, it has further entrenched the advantages of giants. Today, over 80% of the global cloud market is controlled by AWS, Microsoft, and Google; in the domestic market, the top five providers account for over 90% of the market share.

The arrival of AI cloud is pushing this trend to the extreme.

The founder of an AI startup has calculated: To stay in the game in the AI era, at least 100,000 high-end GPU cards are needed. Estimated conservatively at an annualized cost of $10,000 per card, this translates to an annual investment of about $1 billion—and this is just the basic investment in computing power, not including the costs of talent, technology, and scenario implementation. For startups, this threshold is almost impossible to overcome.

The more core barriers are data and scenarios. Among all the data in the world, only a very small portion is publicly accessible via cloud services. The rest is either offline cold data or held by tech giants. ByteDance has massive content and interaction data from Douyin and Doubao, Alibaba has transaction data from e-commerce and payments, and Baidu has scenario data from search and intelligent transportation. This data is the core fuel for training large models and optimizing AI cloud services, which startups simply cannot access.

The deep integration of cloud and large models is forming a new closed loop. For example, a client using Alibaba Cloud's services will be more inclined to use Tongyi Qianwen's large model; once using Tongyi Qianwen's large model, it becomes difficult to leave Alibaba Cloud's computing power support. The same logic applies to ByteDance and Baidu—this closed loop will make the barriers for tech giants higher and higher (higher and higher), while leaving startups with less and less room to survive.

The U.S. market has already validated this trend. Three years after ChatGPT ignited the AI wave, hundreds of AI startups have been established, funded, and explored new products globally. However, to date, the vast majority of market share remains in the hands of two giants, Microsoft and Google. Microsoft, through its integration of Azure and OpenAI, has captured a large number of enterprise clients; Google, with its full-stack AI capabilities, has defended its core business, making it difficult for startups to find sufficient room to survive within the giants' closed loops.

On the other hand, AI cloud is also breaking down some old barriers. Alibaba Cloud's Tongyi Qianwen and Baidu's Wenxin large models have both open-sourced their core versions, allowing any company or individual to develop their own applications based on these open models, to some extent lowering the barrier to AI adoption. Volcano Engine has also opened up many of ByteDance's internally refined AI capabilities to small and medium-sized enterprises through cloud services, enabling them to gain the same AI capabilities as ByteDance without having to build from scratch.

By the end of 2025, Volcano Engine released a preview version of the Doubao mobile assistant, deeply embedding the Doubao large model into the mobile system, bypassing various apps and allowing users to complete operations that would originally require repeated clicks through voice commands. Almost simultaneously, the Qianwen App announced its integration with Taobao, Alipay, Fliggy, and Gaode, with a "Task Assistant" feature capable of completing complex, multi-step tasks on behalf of users. Baidu's Wenxin Yiyan has also been fully integrated into all of Baidu's ecosystem products, including search, maps, and cloud storage.

The giants are all trying to create a unified entry point for the AI era, and AI cloud is the underlying support for this entry point. When users can complete all operations and access all information primarily through a single dialog box or entry point, it may seem on the surface that users have more freedom of choice. However, the control points that determine what they can see, what they can do, and the order in which various services appear are actually fewer.

We may now be building another Tower of Babel. In legend, humans attempted to build a tower to heaven using a single language, but the project ultimately collapsed due to the fragmentation of languages. Today, the giants are using AI to try to create a unified, omniscient entry point, building a Tower of Babel for the AI era.

The war for AI cloud has only just begun. By 2026, this war will enter a white-hot stage, with all three giants investing more resources and talent to compete for the underlying infrastructure of the AI era. This is not a quick battle but a protracted war; the final outcome will depend not only on the amount of computing power and the density of talent but also on who can truly open up AI capabilities to more people and who can genuinely create new scenarios and values for the AI era.

This article is an original piece from Xinmou.

— END —

Solemnly declare: the copyright of this article belongs to the original author. The reprinted article is only for the purpose of spreading more information. If the author's information is marked incorrectly, please contact us immediately to modify or delete it. Thank you.