ericsorides commited on
Commit
d332e6d
·
verified ·
1 Parent(s): c6fdfab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -34,7 +34,7 @@ pip3 install huggingface-hub>=0.17.1
34
  Then you can download any individual model file to the current directory, at high speed, with a command like this:
35
 
36
  ```shell
37
- huggingface-cli download Esperanto/granite-3.0-2b-instruct-fp16-onnx --local-dir granite-3.0-2b-instruct-fp16-onnx --local-dir-use-symlinks False
38
  ```
39
 
40
  For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli).
@@ -130,8 +130,8 @@ def generate_text(model_path, prompt, tokenizer, max_gen_tokens, total_sequence,
130
  We now run the inferences:
131
 
132
  ```python
133
- tokenizer = AutoTokenizer.from_pretrained("Esperanto/granite-3.0-2b-instruct-fp16-onnx")
134
- model_path = "granite-3.0-2b-instruct-fp16-onnx/model.onnx"
135
 
136
  max_gen_tokens = 20 #number of tokens we want tog eneral
137
  total_sequence = 128 #total sequence_length
 
34
  Then you can download any individual model file to the current directory, at high speed, with a command like this:
35
 
36
  ```shell
37
+ huggingface-cli download Esperanto/granite-3.0-2b-instruct-kvc-fp16-onnx --local-dir granite-3.0-2b-instruct-kvc-fp16-onnx --local-dir-use-symlinks False
38
  ```
39
 
40
  For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli).
 
130
  We now run the inferences:
131
 
132
  ```python
133
+ tokenizer = AutoTokenizer.from_pretrained("Esperanto/granite-3.0-2b-instruct-kvc-fp16-onnx")
134
+ model_path = "granite-3.0-2b-instruct-kvc-fp16-onnx/model.onnx"
135
 
136
  max_gen_tokens = 20 #number of tokens we want tog eneral
137
  total_sequence = 128 #total sequence_length