ZhiYuan Research Institute Releases Emu2: A New Generation Generative Multimodal Foundation Model


ZhiYuan Research Institute has open-sourced the JudgeLM evaluation model, which can efficiently assess various large models and provide scores. Compared to GPT-4, JudgeLM's cost is only 1/120, with a consistency rate of over 90% for evaluation results. JudgeLM can be applied in various assessment scenarios including pure text and multimodal contexts, generating scores and justifying reasons. The consistency of JudgeLM with reference answers exceeds 90%, approaching human performance. ZhiYuan Research Institute has also released datasets for training and validation samples for in-depth research on large models.
["ZhiYuan Research Institute has recently open-sourced the Uni3D model with 1 billion parameters, designed for general 3D vision tasks.", "The model can process point cloud data and has achieved breakthroughs in mainstream 3D vision tasks.", "Uni3D employs a unified Transformer architecture and introduces a multimodal alignment training method.", "The model has achieved state-of-the-art results across various 3D vision tasks.", "ZhiYuan Research Institute states that the open-source release of Uni3D will contribute to the future of 3D computing."]
ZhiYuan Research Institute has unveiled the new open-source bilingual model Wudao・Tianying 34 Billion Aquila2-34B, which excels in reasoning, generalization, and more. The institute has also released a comprehensive open-source toolkit to promote collaborative innovation in large model research. Aquila2-34B surpasses other open-source foundational models in overall capabilities, with the ZhiYuan team developing the NLPE method to enhance the model's extension capabilities.
The ZhiYuan Research Institute has released the world's largest Chinese-English semantic vector model training dataset, MTP, with a data scale of 300 million pairs. MTP is the largest open-source dataset of Chinese-English related text pairs, providing an important foundation for training semantic vector models. The dataset includes Chinese-English text pairs from multiple sources, covering various types such as Q&A, comments, and news. The ZhiYuan Research Institute stated that this data plays a crucial role in training large models and will promote collaborative innovation in artificial intelligence. The release of this dataset is expected to address the shortage of training datasets for Chinese models.
Israeli AI platform Wonderful has completed a 100 million USD Series A funding round, bringing total funding to 134 million USD. Unlike GPT shell products, it rapidly gains traction in the global enterprise market through deep integration and localized deployment, attracting the attention of top-tier venture capital firms and demonstrating strong commercial application potential.