dahara1 commited on
Commit
0657532
1 Parent(s): 5cdfc95

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -50,6 +50,8 @@ ALMA-7B-Ja(Ours) | 26.41/83.13 | 34.39/83.50 | 24.77/81.12 | 20.60/78.54 | 15.57
50
  mmnga made llama.cpp(gguf) version [webbigdata-ALMA-7B-Ja-gguf](https://huggingface.co/mmnga/webbigdata-ALMA-7B-Ja-gguf). Thank you!
51
  llama.cpp is a tool used primarily on Macs, and gguf is its latest version format. It can be used without gpu.
52
 
 
 
53
 
54
  ### ALMA-7B-Ja-GPTQ-Ja-En
55
  GPTQ is quantized(reduce the size of the model) method and ALMA-7B-Ja-GPTQ has GPTQ quantized version that reduces model size(3.9GB) and memory usage.
 
50
  mmnga made llama.cpp(gguf) version [webbigdata-ALMA-7B-Ja-gguf](https://huggingface.co/mmnga/webbigdata-ALMA-7B-Ja-gguf). Thank you!
51
  llama.cpp is a tool used primarily on Macs, and gguf is its latest version format. It can be used without gpu.
52
 
53
+ [ALMA-7B-Ja-gguf Free Colab sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_gguf_Free_Colab_sample.ipynb)
54
+
55
 
56
  ### ALMA-7B-Ja-GPTQ-Ja-En
57
  GPTQ is quantized(reduce the size of the model) method and ALMA-7B-Ja-GPTQ has GPTQ quantized version that reduces model size(3.9GB) and memory usage.