Muhammadreza commited on
Commit
4b199d7
1 Parent(s): efd0231

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -86,6 +86,13 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
86
 
87
  ### Inference on a small GPU (Consumer Hardware/Free Colab)
88
 
 
 
 
 
 
 
 
89
  ## Known Issues
90
 
91
  ## Special Thanks
 
86
 
87
  ### Inference on a small GPU (Consumer Hardware/Free Colab)
88
 
89
+ The code is pretty much the same as above, but with a slight diferrence.
90
+
91
+ * Make sure `bitsandbytes` is installed correctly.
92
+ * Your model loading must be `model = AutoModelForCausalLM.from_pretrained(model_name_or_id, load_in_8bit=True, torch_dtype=torch.float16, device_map="auto")`
93
+
94
+ On _free version_ of Google Colab, you may face RAM problems. I guess using `low_cpu_mem_usage=True` in model loading would help.
95
+
96
  ## Known Issues
97
 
98
  ## Special Thanks