The "mutual promotion" between domestic large models and domestic computing power has once again achieved significant progress. On February 12, 2026, Molyneux officially announced that its flagship AI training and inference integrated GPU MTT S5000 has successfully completed full-process adaptation and verification for Zhipu's new generation large model GLM-5.

Core hardware: Full-function GPU smart computing card MTT S5000

As the main character of this adaptation, the MTT S5000 is a hardware foundation specifically designed for large model training, inference, and high-performance computing:

Self-developed architecture: Based on Molyneux's fourth-generation MUSA architecture—**"Pinghu"**.

Peak computing power: The single-card AI computing power can reach up to 1000 TFLOPS.

All-round positioning: It not only supports efficient training of large-scale models but also meets the real-time inference needs in complex scenarios.

Technical significance: Accelerating the "full-chain domestication" of the AI industry

Zhipu's GLM-5, as a leading domestic large model, has a massive parameter scale and complex logic, which places high demands on the computing power platform. The successful completion of this adaptation means:

Ecosystem integration: Domestic self-developed GPUs can stably and efficiently support the full-process operation of top domestic large models.

Cost reduction and efficiency improvement: It provides enterprise users with AI solutions based on domestic computing power, further reducing the barriers and cost risks of deploying large models.

Industry observation:

With the popularity of products like DeepSeek, the market's demand for high-performance computing has reached an unprecedented level. Molyneux and Zhipu's deep collaboration demonstrate a dual breakthrough in hardware performance and model compatibility within the domestic intelligent computing ecosystem, providing solid hardware support for the accelerated development of AGI in China.