First install PyTorch according to the instructions specific to your operating system and hardware.
For development, we recommend installing from source:
pip install uv
git clone https://github.com/allenai/OLMo-core.git
cd OLMo-core
uv sync --all-extras
Or you can install from PyPI with:
uv pip install ai2-olmo-core
There are a number of optional dependencies that must be installed to use certain functionality as well, including:
The published Docker images contain all core and optional dependencies, and are regularly tested on our in-house H100 clusters. But there are several things to keep in mind if you intend to use these images:
If the published images do not work for your use-case for any of the above reasons, you could adapt our Dockerfile to build your own images.
Official training scripts for released models can be found in src/scripts/official/.
These scripts are meant to be launched with torchrun, or with OLMo-core's Beaker launch CLI if you have access to Beaker.
For example:
torchrun --nproc-per-node=8 src/scripts/official/OLMo2/OLMo-2-0325-32B-train.py \ --save-folder=/path/to/save/checkpoints
You can override most configuration options from the command-line. For example, to override the learning rate you could launch the script like this:
torchrun --nproc-per-node=8 src/scripts/official/OLMo2/OLMo-2-0325-32B-train.py \ --save-folder=/path/to/save/checkpoints \ --train_module.optim.lr=6e-3
To continue annealing from a checkpoint, we use a separate script which can be launched like this:
torchrun --nproc-per-node=8 src/scripts/official/OLMo2/OLMo-2-0325-32B-anneal.py \ --save-folder=/path/to/save/checkpoints \ --checkpoint=https://olmo-checkpoints.org/ai2-llm/peteish32/step721901
| Model Family | Directory | Description |
|---|---|---|
| OLMo-2 | src/scripts/official/OLMo2/ | Training scripts and model card for OLMo-2 32B models |
| OLMo-3 | src/scripts/official/OLMo3/ | Training scripts and model cards for OLMo-3 7B and 32B models |
You can use our Hugging Face transformers integration to run inference on the OLMo checkpoints:
pip install transformers>=4.57.0
from transformers import AutoModelForCausalLM, AutoTokenizer
olmo = AutoModelForCausalLM.from_pretrained("allenai/Olmo-3-1125-32B")
tokenizer = AutoTokenizer.from_pretrained("allenai/Olmo-3-1125-32B")
message = ["Language modeling is "]
inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
# inputs = {k: v.to('cuda') for k,v in inputs.items()} # optional verifying cuda
# olmo = olmo.to('cuda')
response = olmo.generate(**inputs, max_new_tokens=100, do_sample=True, temperature=1.0, top_p=0.7)
print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])
Alternatively, with the Hugging Face pipeline abstraction:
from transformers import pipeline
olmo_pipe = pipeline("text-generation", model="allenai/Olmo-3-1125-32B")
print(olmo_pipe("Language modeling is"))
vLLM provides high-throughput inference for OLMo models. You can use it for offline batched inference:
pip install vllm>=0.11.0
from vllm import LLM, SamplingParams
llm = LLM(model="allenai/Olmo-3-1125-32B")
sampling_params = SamplingParams(temperature=1.0, top_p=0.7)
prompts = ["Language modeling is"]
outputs = llm.generate(prompts, sampling_params)
for output in outputs:
prompt = output.prompt
generated_text = output.outputs[0].text
print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
For more details, see the vLLM documentation.
Autoregressive generation is supported directly in Olmo-core. Using this capability, we provide a chat-loop demo that can be used to interact with models in an interactive chat session:
python -m olmo_core.generate.chat https://olmo-checkpoints.org/ai2-llm/Olmo-3-1025-7B/stage3/step11921/ --max-new-tokens 512
Additional tools for evaluating OLMo models are available at the OLMo Eval and olmes repositories.
The Python library source code is located in src/olmo_core. The corresponding tests are located in src/test. The library docs are located in docs. You can build the docs locally with make docs.
Code checks:
pytest to run tests. You can run all tests with pytest -v src/test. You can also point pytest at a specific test file to run it individually.isort and black for code formatting. Ideally you should integrate these into your editor, but you can also run them manually or configure them with a pre-commit hook. To validate that all files are formatted correctly, run make style-check.ruff as our primary linter. You can run it with make lint-check.mypy as our type checker. You can run it with make type-check.@misc{olmo20242olmo2furious, title={{2 OLMo 2 Furious}}, author={{Team OLMo} and Pete Walsh and Luca Soldaini and Dirk Groeneveld and Kyle Lo and Shane Arora and Akshita Bhagia and Yuling Gu and Shengyi Huang and Matt Jordan and Nathan Lambert and Dustin Schwenk and Oyvind Tafjord and Taira Anderson and David Atkinson and Faeze Brahman and Christopher Clark and Pradeep Dasigi and Nouha Dziri and Michal Guerquin and Hamish Ivison and Pang Wei Koh and Jiacheng Liu and Saumya Malik and William Merrill and Lester James V. Miranda and Jacob Morrison and Tyler Murray and Crystal Nam and Valentina Pyatkin and Aman Rangapur and Michael Schmitz and Sam Skjonsberg and David Wadden and Christopher Wilhelm and Michael Wilson and Luke Zettlemoyer and Ali Farhadi and Noah A. Smith and Hannaneh Hajishirzi}, year={2024}, eprint={2501.00656}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2501.00656}, }