File size: 1,096 Bytes
e183f8f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
<img src="https://allenai.org/olmo/olmo-7b-animation.gif" alt="OLMo Logo" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
# Model Card for OLMo 7B
OLMo 7B November 2024 is an updated version of the original [OLMo 7B](https://huggingface.co./allenai/OLMo-7B) model rocking a ____ point increase in ____, among other evaluations improvements, from an improved version of the Dolma dataset and staged training.
**This version is for direct use with HuggingFace Transformers** from v4.40 on.
**For transformers versions v4.40.0 or newer, we suggest using [OLMo 7B HF](https://huggingface.co./allenai/OLMo-7B-hf) instead.**
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
The OLMo models are trained on the [Dolma](https://huggingface.co./datasets/allenai/dolma) dataset.
We release all code, checkpoints, logs (coming soon), and details involved in training these models.
<!-- *A new version of this model with a 24 point improvement on MMLU is available [here](https://huggingface.co./allenai/OLMo-1.7-7B)*. --> |