Moore Threads unveils Huagang GPU architecture at developer conference

Chinese GPU developer Moore Threads has introduced its Huagang architecture, promising significant advances in gaming and AI performance. Set for a 2026 launch, the design targets self-reliance in semiconductors amid global export curbs. While details remain sparse, the company highlighted ambitious benchmarks for upcoming products.

At the recent MUSA Developer Conference, Moore Threads presented its next-generation Huagang architecture, dubbed "Flowerpot" in some translations. This platform aims to power both gaming and artificial intelligence applications, with a full rollout planned for 2026. The announcement focused on performance projections rather than in-depth technical breakdowns, underscoring China's efforts to build domestic GPU capabilities in the face of international restrictions.

Central to the reveal is the Lushan gaming GPU, which will replace the existing MTT S80 and S90 models. Moore Threads asserts that Lushan will deliver a 15-fold increase in AAA game rendering speed and a 50-fold enhancement in ray tracing capabilities. It incorporates a second-generation hardware ray tracing engine and complete DirectX 12 Ultimate compatibility for improved software integration. Memory capacity is expected to reach 64 GB, a quadrupling from the current 16 GB GDDR6 in prior models. Additional touted gains include 64 times faster AI computations, 16 times better geometry processing, four times higher texture fill rates, and eight times quicker atomic memory operations. The architecture introduces UniTE, a unified rendering system with an integrated AI processing unit.

Complementing this, the Huashan AI GPU features a dual-chiplet configuration equipped with nine HBM modules. The firm claims its performance rivals Nvidia's Hopper and Blackwell series, with memory bandwidth surpassing that of the Nvidia B200. Huashan supports a range of precision formats from FP4 to FP64, including proprietary MTFP4, MTFP6, and MTFP8 options. Scalability extends to clusters exceeding 100,000 units through MTLink 4.0, offering 1,314 GB/s interconnect speed. Compared to current offerings, it promises a 50 percent rise in compute density and tenfold efficiency improvements.

Although no gaming demonstrations were shown, a benchmark on the forthcoming MTT S5000 GPU—unrelated to Huashan—ran the DeepSeek V3 model at 1,000 tokens per second in decoding and 4,000 in prefill phases, edging out Nvidia's Hopper performance. These developments reflect Beijing's drive toward technological independence, though the claims await validation as products near market.

Related Articles

Nvidia CEO Jensen Huang announces full production of Vera Rubin AI superchips at CES 2026, with futuristic chip visuals on stage.
Image generated by AI

Nvidia's Vera Rubin AI chips enter full production

Reported by AI Image generated by AI

Nvidia CEO Jensen Huang announced at CES 2026 that the company's next-generation AI superchip platform, Vera Rubin, is now in full production. The platform, first revealed in 2024, promises to reduce costs for training and running AI models. Customers can expect deliveries later this year.

Maxsun has introduced a compact Mini-ITX motherboard equipped with four DDR5 slots, positioning it as a competitor to NVIDIA’s high-performance Petaflop AI mini PC setups. The design emphasizes enhanced memory capabilities for AI applications. Testing shows its airflow optimization lowers system temperatures by about 10°C.

Reported by AI

Shanghai-based GPU maker MetaX made a spectacular debut on the Star Market, with shares opening at 700 yuan, a 569% increase from the IPO price of 104.66 yuan. The stock climbed further to 824.50 yuan intraday.

Nvidia unveiled NemoClaw for easier OpenClaw AI agent creation, DLSS 5 for advanced gaming graphics, and a new Vera CPU during its GTC 2026 keynote in San Jose. CEO Jensen Huang highlighted agentic AI advancements, comparing OpenClaw to HTML and Linux, while teasing partnerships like Disney robotics and space computing.

Reported by AI

China’s Zhipu AI launched its new major model GLM-5 on February 12, 2026, positioning it as a challenge to rivals. The model marks a shift in AI development from ‘vibe coding’ to ‘agentic engineering’ for enhanced performance.

Chinese researchers have introduced photonic AI chips that promise significant speed advantages in specific generative tasks. These chips use photons instead of electrons, enabling greater parallelism through optical interference. The development could mark a step forward in AI hardware, though claims are limited to narrowly defined applications.

Reported by AI

Several mainland Chinese suppliers of memory chips and storage solutions are pursuing listings in Hong Kong, signaling a strategic shift to fuel the sector's global ambitions. The most watched is Shanghai-based Montage Technology, set to debut on the Hong Kong stock exchange next week and raise up to US$896 million. Analysts view this wave as a key move for international growth in cloud computing and AI.

 

 

 

This website uses cookies

We use cookies for analytics to improve our site. Read our privacy policy for more information.
Decline