Recently, the NLP Lab at Tsinghua University, OpenBMB, and Miga Intelligence jointly released and open-sourced UltraEval-Audio, a evaluation framework specifically designed for audio models. UltraEval-Audio not only establishes a complete set of evaluation methodologies for the field of audio large models, but also concretizes this system into an out-of-the-box engineering framework, thereby completing the overall structure of audio evaluation.

The latest version of UltraEval-Audio, v1.1.0, adds the capability to reproduce popular audio models with one click based on the existing "one-click evaluation" function, and expands support for specialized models such as Text-to-Speech (TTS), Automatic Speech Recognition (ASR), and Codec. In addition, this version introduces an isolated inference operation mechanism, aiming to lower the threshold for model reproduction and improve the controllability and portability of the evaluation process.
Notably, UltraEval-Audio v1.1.0 has become an essential evaluation tool for many high-impact audio and multimodal models such as MiniCPM-o2.6 and VoxCPM. The open-source release of this framework will significantly improve the efficiency of researchers in the development of audio models and promote progress in the relevant fields.
The open-source address is also public, and researchers can obtain more information through GitHub. The release of UltraEval-Audio marks an important step forward in the standardization of audio model evaluation, helping to accelerate the development of audio technology.
Open source address:https://github.com/OpenBMB/UltraEval-Audio
Key points:
🌟 UltraEval-Audio is an evaluation framework for audio models, jointly released by the NLP Lab at Tsinghua University, OpenBMB, and Miga Intelligence.
🚀 The latest version v1.1.0 adds the one-click reproduction function and supports the evaluation of more specialized models.
📈 The open-source release will significantly improve the development efficiency of researchers and promote progress in the field of audio models.
