Chinese AI developers explore renting Nvidia’s Rubin GPU in the cloud — cost, complexity, and regulatory hurdles could limit deployments

2 hours ago 1
Nvidia Vera Rubin, CES 2026
(Image credit: Tom's Hardware)

China-based AI hardware developers are making rapid progress with accelerators of their own design. However, China's most advanced AI developers are increasingly acknowledging that domestic hardware is unlikely to catch up with U.S. leaders in the near term, which greatly limits the development of competitive models. To that end, in a bid to stay competitive with American peers, Chinese AI developers are exploring ways to rent Nvidia's upcoming Rubin GPUs in the cloud, reports the Wall Street Journal.

When Nvidia introduced its Rubin datacenter platform for AI in January, it publicly named American customers and omitted Chinese ones. The company has taken this approach in recent quarters, reflecting U.S. export rules, its commitment to comply with them, and its intent not to signal to investors about the opening of the Chinese market. The message was received by Chinese companies, and they began exploring ways to obtain leading-edge processors from Nvidia remotely to avoid falling behind their U.S. rivals.

There is no surprise that using remote hardware to train frontier AI models is tricky, as the difference between renting Rubin in a remote cloud data center and deploying it locally is profound. U.S. hyperscalers can integrate Rubin accelerators at scale, tune their software stacks tightly around the new hardware, and reserve massive GPU clusters for long training runs. By contrast, Chinese developers that plan to rent Rubin capacity will have to cope with limited allocations, cross-border latency, limited freedom to customize systems, and, in some cases, queuing. If they rent enough systems — and there are cloud data centers in the U.S. that currently run hundreds of thousands of Blackwell GPUs — they may well train their models without much hassle. However, if they cannot find appropriate clouds on time, they will have fewer AI accelerators per project and, in some cases, be unable to run large training jobs, which will directly cap model size, experimentation cadence, and iteration speed.

Get Tom's Hardware's best news and in-depth reviews, straight to your inbox.

Google Preferred Source

Follow Tom's Hardware on Google News, or add us as a preferred source, to get our latest news, analysis, & reviews in your feeds.

Anton Shilov is a contributing writer at Tom’s Hardware. Over the past couple of decades, he has covered everything from CPUs and GPUs to supercomputers and from modern process technologies and latest fab tools to high-tech industry trends.

Read Entire Article