Toggle light / dark theme

The Allen Institute for AI created the Open Language Model, or OLMo, which is an open-source large language model with the aim of advancing the science of language models through open research.


AI2 has partnered with organizations such as Surge AI and MosaicML for data and training code. These partnerships are crucial for providing the diverse datasets and sophisticated training methodologies that underpin OLMo’s capabilities. The collaboration with the Paul G. Allen School of Computer Science and Engineering at the University of Washington and Databricks Inc. has also been pivotal in realizing the OLMo project.

It is important to note that the current architecture of OLMo is not the same as the models that power chatbots or AI assistants, which use instruction-based models. However, that’s on the roadmap. According to AI2, there will be multiple enhancements made to the model in the future. In the coming months, there are plans to iterate on OLMo by introducing different model sizes, modalities, datasets, and capabilities into the OLMo family. This iterative process is aimed at continuously improving the model’s performance and utility for the research community.

OLMo’s open and transparent approach, along with its advanced capabilities and commitment to continuous improvement, make it a major milestone in the evolution of LLMs.

In a famous line over 60 years ago, early AI pioneer Norbert Wiener summed up one of the core challenges that humanity faces in building artificial intelligence: If we use, to achieve our purposes, a mechanical agency with whose operation we cannot interfere effectively…we had better be quite sure…


The answer is a technology known as reinforcement learning from human feedback (RLHF).

RLHF has become the dominant method by which human developers control and steer the behavior of AI models, especially language models. It impacts how millions of people around the world experience artificial intelligence today. It is impossible to understand how today’s most advanced AI systems work without understanding RLHF.

At the same time, newer methods are quickly emerging that seek to improve upon and displace RLHF in the AI development process. The technological, commercial and societal implications are profound: at stake is how humans shape the way that AI behaves. Few areas of AI research are more active or important today.