license: apache-2.0 | |
language: | |
- en | |
tags: | |
- moe | |
- olmo | |
- olmoe | |
- molmo | |
- molmoe | |
co2_eq_emissions: 1 | |
datasets: | |
- allenai/OLMoE-mix-0924 | |
library_name: transformers | |
<img alt="Molmo Logo." src="molmo_logo.png" width="250px"> | |
# Model Summary | |
> MolmoE-1B is a multimodal Mixture-of-Experts LLM with 1.5B active and 7.2B total parameters released in September 2024 (0924) based on [OLMoE-1B-7B-0924](https://huggingface.co./allenai/OLMoE-1B-7B-0924). It yields state-of-the-art performance among multimodal models with a similar size while being fully open-source. | |
- **Paper:** WIP | |
- **Code:** WIP | |
# Use | |
WIP | |
# Evaluation Snapshot | |
WIP | |
# Citation | |
WIP |