The Allen Institute for AI (AI2) has unveiled OLMo, a large language model touted as “truly open source,” giving companies confidence in its underlying architecture and development process. Accompanied by its model code, weights, training code, and an evaluation suite, OLMo enables users to fully understand its design, training, and evaluation, fostering transparency and innovation in AI applications.
Built on a robust dataset known as Dolma, which consists of three trillion tokens developed by AI2, OLMo is available in four configurations, each featuring approximately seven billion parameters. This positions OLMo as a formidable competitor to models such as Meta's Llama 2-7B and Mistral’s Mixtral 8x7B.
Founded by the late Paul Allen, co-founder of Microsoft, AI2 aims to empower academics and researchers by providing an open-source model that facilitates collaborative exploration of language model science. By making the training aspects of OLMo fully accessible, AI2 emphasizes that this open approach can significantly reduce carbon emissions often associated with fine-tuning AI models. The organization argues that such transparency diminishes developmental redundancies, which is essential for the ongoing decarbonization endeavors within the AI field. Furthermore, it enhances research efficiency, as scholars are no longer constrained by speculative assumptions regarding model performance.
Notable figures in the AI community, including Eric Horvitz, Microsoft’s chief scientific officer and a founding member of the AI2 Scientific Advisory Board, have expressed enthusiasm about OLMo's potential to propel AI research forward. Similarly, Yann LeCun, Chief AI Scientist at Meta, remarked on the benefits of open-source collaboration, claiming it is the “fastest and most effective way to build the future of AI.”
In conjunction with OLMo's release, AI2 also introduced Paloma, a benchmarking tool designed to evaluate open language models. Paloma assesses models on various natural language processing tasks across diverse domains, including niche artistic communities and mental health discussions on platforms like Reddit. This comprehensive testing approach opens new avenues for evaluating model effectiveness, beyond traditional metrics typically considered by model testers.
Additionally, companies can access Dolma, the trillion-token pretraining dataset that formed the foundation for OLMo. This dataset, derived from a variety of sources including web content, academic publications, and books, is available for commercial applications through platforms like Hugging Face Hub, enabling widespread use and exploration in the field of AI.
With OLMo and its related resources, AI2 is poised to drive significant advancements in language model research and application, reinforcing the critical role of openness, collaboration, and sustainability in the evolution of artificial intelligence.