Fu Sheng Unveils Orion-14B Large Model with 14 Billion Parameters


Tencent has released the HY-Embodied-0.5 foundational model specifically designed for robots, aiming to address the shortcomings of general visual language models in 3D spatial perception and physical interaction, and to advance large models toward the field of robot control. The series of models have been restructured in both architecture and training, and are accompanied by the release of main models such as MoT-2B.
MiniMax launched the MMX-CLI command line tool, specifically designed for AI Agent, simplifying the process of calling multi-modal models. The tool solves issues such as complicated interface adaptation and code redundancy, allowing Agents to easily schedule various AI capabilities as if they were native applications. Users can invoke programming, video generation, and other functions with one click in mainstream development environments, without the need to write additional MCP Servers or adapt complex interfaces.
OpenAI officially adapts to Apple CarPlay, allowing iPhone users to interact with ChatGPT through voice on the car's center display. To ensure driving safety, ChatGPT on CarPlay only supports full voice interaction and prohibits displaying text or images.
Alibaba released the new generation cross-modal large model Qwen3.5-Omni, achieving seamless cross-modal interaction with images, videos, voice, and text. It performed excellently in 215 tasks, marking a significant breakthrough for domestic large models in the field of audiovisual interaction, reaching global top levels.
ByteDance's Doubao ranks among top global AI models, competing with leaders. Xiaomi's MiMo debuts, highlighting phone makers' R&D. Chinese models advance in Chinese comprehension, closing in on GPT-4.....