Update README.md
Browse files
README.md
CHANGED
@@ -97,7 +97,7 @@ The model is fake quantized which means each weight has <= 16 (2<sup>4</sup>) un
|
|
97 |
```python
|
98 |
from transformers import AutoModelForCausalLM
|
99 |
|
100 |
-
model = AutoModelForCausalLM.from_pretrained("smpanaro/Llama-2-7b-NuGPTQ")
|
101 |
linear_layers = ["k_proj", "q_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"]
|
102 |
count = 0
|
103 |
for key, tensor in model.state_dict().items():
|
|
|
97 |
```python
|
98 |
from transformers import AutoModelForCausalLM
|
99 |
|
100 |
+
model = AutoModelForCausalLM.from_pretrained("smpanaro/Llama-2-7b-NuGPTQ", trust_remote_code=True)
|
101 |
linear_layers = ["k_proj", "q_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"]
|
102 |
count = 0
|
103 |
for key, tensor in model.state_dict().items():
|