Back
Apr 27, 2026

In the fast-moving world of artificial intelligence, breakthroughs are often measured in parameters and benchmarks. But the real game-changer happens when cutting-edge models meet equally powerful hardware.
That moment has arrived.
DeepSeek’s latest V4-Pro — a massive 1.6 trillion parameter Mixture-of-Experts (MoE) model with only 49B active parameters — has been successfully validated for full-scale inference on Huawei’s Ascend 950PR heterogeneous chip. This is more than just another model release. It is the clearest sign yet that China has achieved a true “model-chip closed loop” — a self-sustaining ecosystem where world-class AI models and domestically developed AI chips work together in perfect synergy.
What Is the “Model-Chip Closed Loop”?
For years, the global AI industry has faced a painful bottleneck: the most powerful models were built on hardware that was increasingly difficult to access due to export restrictions.
The “model-chip closed loop” changes that equation.
It means:
Top-tier open-source models (like DeepSeek V4) are optimized from day one for domestic hardware.
Chinese AI chips (like Ascend 950PR) are designed with specific model architectures (especially MoE) in mind.
Software stack (CANN), memory (HiBL), and heterogeneous computing architecture form a complete, high-performance pipeline.
The result? Frontier-level AI that is faster, cheaper, and fully sovereign.
Technical Breakthrough: V4 + Ascend 950PR
Here’s why this combination is historic:
Model Side: DeepSeek V4-Pro delivers 1.6T total parameters while activating only 49B per token — delivering near-GPT-5/Claude Opus-level performance at a fraction of the inference cost and with native 1M token context support.
Hardware Side: The Ascend 950PR offers 1.56 PFLOPS FP4 compute, 112 GB HiBL memory, and advanced heterogeneous architecture optimized for MoE’s “cold-hot expert” routing patterns.
Performance: Day-0 full adaptation means real-world inference performance already matches or exceeds the best NVIDIA H20 solutions available in China, at significantly lower total cost of ownership.
This is the first time a truly frontier open-weight model has been proven at scale on fully domestic AI silicon.
Industry-Wide Impact
The ripple effects are already visible:
LLM Companies (MiniMax, 智谱, etc.): Inference costs drop sharply → token pricing becomes competitive → faster path from “token demand” to real revenue.
Semiconductor Sector: Strong validation for Ascend and other domestic players — the +6.33% movement in semiconductor stocks was just the beginning.
Enterprises: Top-tier AI can now be deployed safely and cost-effectively on-premise or in private clouds without relying on restricted foreign GPUs.
Global AI Landscape: The world is moving toward a true “dual-track” infrastructure — CUDA in the West, CANN + heterogeneous chips in China.
From Breakthrough to Business Value: A*Crew by GeniiAI
For enterprises, the most exciting part is not the benchmark numbers — it’s the ability to turn this technology into daily productivity.
A*Crew by GeniiAI is purpose-built for exactly this moment.
Powered by the latest frontier LLMs (including DeepSeek-class models), A*Crew is an enterprise-grade AI Agent Office Assistant that delivers:
Multi-agent collaboration across HR, finance, marketing, IT, and operations
Secure, privacy-first deployment with full company knowledge base integration
Automated workflows that actually save hours every day
Whether you run V4 on Ascend chips in your own data center or in the cloud, A*Crew makes the leap from raw model power to real business impact seamless and secure.
👉 Ready to bring V4-level AI into your organization? Explore A*Crew by GeniiAI today: https://www.geniiai.co/
