Text Generation
Transformers
PyTorch
mpt
Composer
MosaicML
llm-foundry
custom_code
text-generation-inference
jacobfulano commited on
Commit
af1b522
1 Parent(s): aa53cd9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -49,11 +49,12 @@ model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b-story
49
  model.to(device='cuda:0', dtype=torch.bfloat16)
50
  ```
51
 
52
- Although the model was trained with a sequence length of 2048, ALiBi enables users to increase the maximum sequence length during finetuning and/or inference. For example:
 
53
 
54
  ```python
55
  config = transformers.AutoConfig.from_pretrained('mosaicml/mpt-7b-storywriter', trust_remote_code=True)
56
- config.update({"max_seq_len": 4096})
57
  model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b-storywriter', config=config, trust_remote_code=True)
58
  ```
59
 
 
49
  model.to(device='cuda:0', dtype=torch.bfloat16)
50
  ```
51
 
52
+ Although the model was trained with a sequence length of 2048 and finetuned with a sequence length of 65536,
53
+ ALiBi enables users to increase the maximum sequence length during finetuning and/or inference. For example:
54
 
55
  ```python
56
  config = transformers.AutoConfig.from_pretrained('mosaicml/mpt-7b-storywriter', trust_remote_code=True)
57
+ config.update({"max_seq_len": 83968})
58
  model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b-storywriter', config=config, trust_remote_code=True)
59
  ```
60