Recently, the École Polytechnique Fédérale de Lausanne (EPFL), ETH Zurich, and the Swiss National Supercomputing Centre (CSCS) jointly released a large-scale open-source language model named "Apertus." The name "Apertus" means "open" in Latin, and its development philosophy perfectly embodies this spirit.
Compared to major models in the current market such as OpenAI's GPT series, Meta's Llama, and Anthropic's Claude, Apertus not only focuses on technology itself but has also gained widespread attention for its transparency. All aspects of this model, including model weights, architecture design, training code, and data sources, are fully open. Even the complete documentation of the training process is available without any restrictions. This "open" strategy makes Apertus a promising new choice.
In this era of rapid AI development, the phenomenon of "data black boxes" remains prevalent. Many large tech companies often withhold significant key details when releasing their models, leading to a lack of trust in the industry. Apertus breaks this norm by striving to promote technology sharing and collaboration through open source. Researchers and developers can build upon it to innovate and improve, driving progress across the entire industry.
The research team stated that the launch of Apertus is not only aimed at facilitating research but also hopes to inspire AI researchers and developers around the world to participate in the open-source ecosystem. Through this approach, Apertus aims to lower the barriers to AI technology, allowing more people to get involved in AI development and application, thus accelerating technological advancement.
The emergence of Apertus marks an important attempt by Switzerland in the global AI field, indicating that more similar open-source projects may emerge in the future, promoting the development of the entire industry. As this project continues to progress, we look forward to seeing more innovations and breakthroughs.