Hugging Face Launches Open Source Multimodal AI Model IDEFIX


Multimodal AI company ElevenLabs launches an integrated content creation platform, combining image generation, video production, voice synthesis, music creation, and sound design features, enabling a complete production cycle from script to final video. It helps creators and marketers avoid switching between multiple platforms, efficiently completing commercial video production.
Meituan's open-source multimodal large model, LongCat-Flash-Omni, achieves a technological breakthrough, surpassing closed-source competitors in multiple benchmark tests, reaching industry-leading levels. The model supports real-time integration processing of text, speech, images, and video, with near-zero latency in interaction, pushing locally developed multimodal AI applications to a new level.
Google has launched the StreetReaderAI prototype system, helping blind and low-vision users to independently explore Google Street View through natural language interaction. The system integrates computer vision, geographic information systems, and large language models, enabling a multimodal AI-driven real-time conversational street view experience, breaking through the limitations of traditional voice announcements and enhancing the freedom of accessible urban exploration.
Recently, Shengshu Technology, a leading company in the field of multimodal AI, announced the successful completion of an A-round funding round worth several billion yuan. This round was led by Bohua Capital, with existing investors such as Baidu's strategic investment division and the Beijing Artificial Intelligence Industry Investment Fund continuing to participate, demonstrating strong market recognition of Shengshu Technology. The company plans to use the funds to further advance model R&D and technological innovation, explore the potential of multimodal large models, and accelerate product expansion and user services. Multimodal technology, especially in the field of video generation, is currently experiencing rapid development.
Liquid AI launches LFM2-VL vision-language models optimized for low-latency and edge devices. The 450M and 1.6B variants offer speed-accuracy balance for mobile/embedded use, doubling GPU inference speed versus existing models.....