winglian commited on
Commit
946b2e7
·
1 Parent(s): d90dc27

base model in config is unnecessary, and fix grammar

Browse files
Files changed (2) hide show
  1. chat.py +1 -1
  2. config.yml +0 -2
chat.py CHANGED
@@ -86,7 +86,7 @@ with blocks:
86
  gr.Markdown(f"""
87
  ### brought to you by OpenAccess AI Collective
88
  - This is the [{config["repo"]}](https://huggingface.co/{config["repo"]}) model file [{config["file"]}](https://huggingface.co/{config["repo"]}/blob/main/{config["file"]})
89
- - This Space uses GGML with GPU support, so it can run larger models on smaller GPUs & VRAM quickly.
90
  - This is running on a smaller, shared GPU, so it may take a few seconds to respond.
91
  - [Duplicate the Space](https://huggingface.co/spaces/openaccess-ai-collective/ggml-ui?duplicate=true) to skip the queue and run in a private space or to use your own GGML models.
92
  - When using your own models, simply update the [config.yml](https://huggingface.co/spaces/openaccess-ai-collective/ggml-ui/blob/main/config.yml)")
 
86
  gr.Markdown(f"""
87
  ### brought to you by OpenAccess AI Collective
88
  - This is the [{config["repo"]}](https://huggingface.co/{config["repo"]}) model file [{config["file"]}](https://huggingface.co/{config["repo"]}/blob/main/{config["file"]})
89
+ - This Space uses GGML with GPU support, so it can quickly run larger models on smaller GPUs & VRAM.
90
  - This is running on a smaller, shared GPU, so it may take a few seconds to respond.
91
  - [Duplicate the Space](https://huggingface.co/spaces/openaccess-ai-collective/ggml-ui?duplicate=true) to skip the queue and run in a private space or to use your own GGML models.
92
  - When using your own models, simply update the [config.yml](https://huggingface.co/spaces/openaccess-ai-collective/ggml-ui/blob/main/config.yml)")
config.yml CHANGED
@@ -1,8 +1,6 @@
1
  ---
2
  repo: TheBloke/wizard-vicuna-13B-GGML
3
  file: wizard-vicuna-13B.ggml.q5_1.bin
4
- # if the repo above doesn't include the tokenizer set the base repo it was based on with a valid tokenizer model
5
- base_model: junelee/wizard-vicuna-13b
6
  llama_cpp:
7
  n_ctx: 2048
8
  n_gpu_layers: 40 # llama 13b has 40 layers
 
1
  ---
2
  repo: TheBloke/wizard-vicuna-13B-GGML
3
  file: wizard-vicuna-13B.ggml.q5_1.bin
 
 
4
  llama_cpp:
5
  n_ctx: 2048
6
  n_gpu_layers: 40 # llama 13b has 40 layers