Kunlun Tech: Multi-Modal Large Model Has Entered Experimental Training Phase


According to Kunlun Tech's 2025 annual report, the company's revenue reached 8.198 billion yuan, an increase of 44.78% year-on-year, with overseas revenue reaching 7.723 billion yuan, up 49.91%. The company introduced the "4+3 Strategy", clearly defining the development direction of AI-driven content production, covering both technological and business layout.
On June 29, 2025, the Alibaba International AI Team officially released the new multi-modal large model **Ovis-U1**, marking another major breakthrough in the field of multi-modal artificial intelligence. As the latest masterpiece of the Ovis series, Ovis-U1 integrates multi-modal understanding, image generation, and image editing functions, demonstrating powerful cross-modal processing capabilities, providing new possibilities for developers, researchers, and industry applications. This is a detailed report on Ovis-U1 by AIbase. Ovis-U1
The latest release from the Alibaba team, mPLUG-Owl3 is a general-purpose multi-modal large model, with its core capability being the understanding of long image sequences. By introducing a hyper attention module, mPLUG-Owl3 can efficiently process visual and language information, achieving in-depth understanding and communication of multi-modal data such as images and videos. This model has made significant breakthroughs in inference efficiency, image processing capabilities, and the application of multi-modal knowledge, particularly in video understanding, where it can 'watch' a 2-hour movie in 4 seconds and accurately answer related questions.
ChinaZ.com (Site Editor's Home) July 9 News: Poe recently launched an innovative feature called Previews, offering users an unprecedented interactive experience. This feature allows users to directly view and operate web applications generated by AI during chats, marking a new era for AI social interaction.The core highlight of the Previews feature lies in its intuitiveness and ease of use. Users can view and interact with AI-generated web applications in real-time within the chat interface, jus
Groq has recently launched a lightning-fast LLM engine on its website, allowing developers to perform quick queries and task execution on large language models directly.This engine utilizes Meta's open-source Llama3-8b-8192 LLM, supports other models by default, and its speed is stunning. According to test results, Groq's engine can process 1256.54 tokens per second, far surpassing the speed of GPU chips from companies like Nvidia. This has garnered widespread attention from both developers and