File size: 1,705 Bytes
81dadc1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
---
license: mit
widget:
 - text: "MQIFVKTLTGKTITLEVEPS<mask>TIENVKAKIQDKEGIPPDQQRLIFAGKQLEDGRTLSDYNIQKESTLHLVLRLRGG"
---

## ESM-2

ESM-2 is a state-of-the-art protein model trained on a masked language modelling objective. It is suitable for fine-tuning on a wide range of tasks that take protein sequences as input. For detailed information on the model architecture and training data, please refer to the [accompanying paper](https://www.biorxiv.org/content/10.1101/2022.07.20.500902v2). You may also be interested in some demo notebooks ([PyTorch](https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/protein_language_modeling.ipynb), [TensorFlow](https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/protein_language_modeling-tf.ipynb)) which demonstrate how to fine-tune ESM-2 models on your tasks of interest.

Several ESM-2 checkpoints are available in the Hub with varying sizes. Larger sizes generally have somewhat better accuracy, but require much more memory and time to train:

| Checkpoint name | Num layers | Num parameters |
|------------------------------|----|----------|
| [esm2_t48_15B_UR50D](https://huggingface.co./facebook/esm2_t48_15B_UR50D) | 48 | 15B     |
| [esm2_t36_3B_UR50D](https://huggingface.co./facebook/esm2_t36_3B_UR50D) | 36 | 3B      | 
| [esm2_t33_650M_UR50D](https://huggingface.co./facebook/esm2_t33_650M_UR50D) | 33 | 650M    | 
| [esm2_t30_150M_UR50D](https://huggingface.co./facebook/esm2_t30_150M_UR50D) | 30 | 150M    | 
| [esm2_t12_35M_UR50D](https://huggingface.co./facebook/esm2_t12_35M_UR50D) | 12 | 35M     | 
| [esm2_t6_8M_UR50D](https://huggingface.co./facebook/esm2_t6_8M_UR50D)  | 6  | 8M      |