Update README.md
Browse files
README.md
CHANGED
@@ -19,6 +19,6 @@ inference: false
|
|
19 |
# GLM 4 9B Chat - Abliterated
|
20 |
Check out the <a href="https://huggingface.co/byroneverson/glm-4-9b-chat-abliterated/blob/main/abliterate-glm-4-9b-chat.ipynb">jupyter notebook</a> for details of how this model was abliterated from glm-4-9b-chat.
|
21 |
|
22 |
-
The python package "tiktoken" is required to quantize the model into gguf format. So I had to create <a href="https://huggingface.co/spaces/byroneverson/gguf-my-repo-plus-tiktoken">a fork of GGUF My Repo (+tiktoken)</a
|
23 |
|
24 |
![Logo](https://huggingface.co/byroneverson/internlm2_5-7b-chat-abliterated/resolve/main/logo.png "Logo")
|
|
|
19 |
# GLM 4 9B Chat - Abliterated
|
20 |
Check out the <a href="https://huggingface.co/byroneverson/glm-4-9b-chat-abliterated/blob/main/abliterate-glm-4-9b-chat.ipynb">jupyter notebook</a> for details of how this model was abliterated from glm-4-9b-chat.
|
21 |
|
22 |
+
The python package "tiktoken" is required to quantize the model into gguf format. So I had to create <a href="https://huggingface.co/spaces/byroneverson/gguf-my-repo-plus-tiktoken">a fork of GGUF My Repo (+tiktoken)</a>.
|
23 |
|
24 |
![Logo](https://huggingface.co/byroneverson/internlm2_5-7b-chat-abliterated/resolve/main/logo.png "Logo")
|