top of page

>

>

ByteDance and Alibaba are buying Huawei’s Ascend 950PR for easier AI deployment

Huawei Ascend 950PR

News

ByteDance and Alibaba are buying Huawei’s Ascend 950PR for easier AI deployment

Mar 27, 2026

14:15

Disruption snapshot


  • Huawei’s Ascend 950PR shifts focus from raw speed to easier deployment. It works better with CUDA-like tools, cuts migration effort, and is optimized for inference workloads.


  • Winners: Huawei, Chinese cloud platforms, inference-heavy AI apps. Losers: Nvidia ecosystem lock-in, firms reliant on proprietary CUDA advantages.


  • Watch real inference usage. Track whether ByteDance and Alibaba run production workloads on Ascend, not just test deployments.

Huawei’s new Ascend 950PR AI chip is starting to draw orders from major Chinese tech companies like ByteDance and Alibaba ahead of broader shipments in 2026.


That’s notable given Huawei’s earlier Ascend 910C didn’t see the same level of adoption from China’s largest private internet firms, even with strong policy pressure to use domestic chips.


They still mostly opted for Nvidia.


On paper, the 950PR doesn’t look like a major jump in raw computing power over the 910C. But that’s exactly why the demand matters. Chinese tech companies aren’t just chasing peak performance anymore. They’re trying to add usable AI capacity quickly, with less engineering friction and faster deployment into real products.


That’s where the 950PR seems better positioned. It’s reported to work more smoothly with the Nvidia-style software environment developers already know, deliver faster response times, and focus more directly on inference workloads like chatbots, search, and recommendation systems.


So the shift here isn’t mainly about Huawei closing the benchmark gap. It’s about whether the company is finally reducing the software and deployment pain that held back earlier Ascend chips. If that’s happening, adoption could follow much more quickly this time.

 

Why the 950PR is getting orders the 910C did not

 

Huawei struggled to persuade major private-sector Chinese tech firms to adopt the Ascend 910C in large quantities even with strong policy pressure to use domestic semiconductors. That matters because it rules out the easy explanation that the market simply needed more time or more patriotic pressure. Big customers already had a reason to diversify away from Nvidia. They still did not move at scale.

 

The difference now is that the product seems more deployable: easier migration from Nvidia-based workflows, better response speed, and a design optimized for inference rather than just headline chip comparisons. For companies like ByteDance and Alibaba, which are also trying to meet AI demand with new chips, that changes the procurement question from “Is this politically useful?” to “Can this run production workloads without burning months of engineering time?”

 

Price supports that case, but it is not the whole case. The 950PR is expected to sell for about 50,000 yuan per card, while a premium HBM-equipped version will be about 70,000 yuan. Those are meaningful numbers for scaled deployments, but large internet platforms do not commit to new accelerators just because they are cheaper. They commit when the hardware-software tradeoff becomes manageable enough to support real service delivery, especially as tech giants plan up to $700 billion for AI infrastructure in 2026.

 

The real bottleneck was software migration

 

Nvidia’s hold on China’s AI market has never been just about chip performance.

 

Developers have long relied on CUDA, and switching to domestic chips has required rewriting code and retraining on new platforms, making migration costly and slow. Even when Chinese chips could approach the performance of Nvidia’s downgraded offerings, internet companies still preferred Nvidia because there was no mature domestic alternative to CUDA. That lock-in matters even more if Nvidia can keep adapting its China strategy with approved H200 chips.

 

That is why the most important detail is not the modest performance bump. It is that Huawei appears to have moved away from forcing developers into a more proprietary, CANN-heavy path and now lets Chinese tech firms that have generally used Nvidia’s software system migrate models more easily. That is a direct attack on the adoption bottleneck that held back the 910C. A domestic chip does not need to beat Nvidia everywhere to win share. It needs to reduce the switching penalty enough that “good enough” hardware becomes operationally worth using.

 

That is also why the 950PR order story matters more than the product launch story. The 910C already showed that China could assemble a more competitive domestic AI chip. But that architectural progress did not automatically produce broad adoption among the biggest private buyers. The 950PR matters because it suggests Huawei may be improving the part customers actually feel first: deployment friction.

 

Why inference is the right battlefield

 

The timing fits the workload shift. The 950PR is designed to excel at inference, while demand for inference computing in China is rising as the sector moves from model development toward real-world deployment. That matters because inference rewards a different mix of attributes than frontier training does. Reliability, responsiveness, availability, and ease of integration can matter more than absolute peak compute leadership when the job is serving models at scale.

 

That makes “better response speed” more important than it sounds. In a training race, a modest compute shortfall can be disqualifying. In inference, a chip that is available domestically, priced competitively, and easier to fit into existing workflows can still win meaningful volume even if Nvidia remains ahead on ecosystem depth and top-end performance. The 950PR is not necessarily Huawei’s answer to Nvidia’s best chips. It may be Huawei’s answer to the much more practical question Chinese buyers are now asking: what can we deploy at scale this year?

 

The real test arrives in the second half of 2026

 

Huawei sent samples in January, expects mass production to begin in April 2026, and plans fully fledged shipments in the second half of 2026, targeting about 750,000 units this year. Those dates turn the story into a near-term operating test, not a distant roadmap promise.

 

The signal to watch is not the shipment target on its own. It is whether major Chinese cloud, app, and platform companies begin running meaningful inference volume on Ascend rather than just testing or politically endorsing it. If that happens, the implication is clear: the main barrier to domestic substitution was not simply raw chip weakness but software migration cost, and Huawei is starting to lower it. If it does not happen, Nvidia’s moat remains what it has been all along: not just faster silicon, but the ecosystem developers are least willing to leave.

Recommended Articles

loading-animation.gif

loading-animation.gif

loading-animation.gif

bottom of page