jacobfulano
commited on
Commit
•
af1b522
1
Parent(s):
aa53cd9
Update README.md
Browse files
README.md
CHANGED
@@ -49,11 +49,12 @@ model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b-story
|
|
49 |
model.to(device='cuda:0', dtype=torch.bfloat16)
|
50 |
```
|
51 |
|
52 |
-
Although the model was trained with a sequence length of 2048
|
|
|
53 |
|
54 |
```python
|
55 |
config = transformers.AutoConfig.from_pretrained('mosaicml/mpt-7b-storywriter', trust_remote_code=True)
|
56 |
-
config.update({"max_seq_len":
|
57 |
model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b-storywriter', config=config, trust_remote_code=True)
|
58 |
```
|
59 |
|
|
|
49 |
model.to(device='cuda:0', dtype=torch.bfloat16)
|
50 |
```
|
51 |
|
52 |
+
Although the model was trained with a sequence length of 2048 and finetuned with a sequence length of 65536,
|
53 |
+
ALiBi enables users to increase the maximum sequence length during finetuning and/or inference. For example:
|
54 |
|
55 |
```python
|
56 |
config = transformers.AutoConfig.from_pretrained('mosaicml/mpt-7b-storywriter', trust_remote_code=True)
|
57 |
+
config.update({"max_seq_len": 83968})
|
58 |
model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b-storywriter', config=config, trust_remote_code=True)
|
59 |
```
|
60 |
|