jarradh commited on
Commit
a643903
1 Parent(s): 626c051

Better attribution

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -1,17 +1,21 @@
1
  ---
2
- license: other
3
  datasets:
4
  - ehartford/wizard_vicuna_70k_unfiltered
 
 
5
  ---
6
 
7
  # Overview
8
  Fine-tuned [Llama-2 70B](https://huggingface.co/TheBloke/Llama-2-70B-fp16) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered).
9
  [QLoRA](https://arxiv.org/abs/2305.14314) was used for fine-tuning. The model was trained for three epochs on a single NVIDIA A100 80GB GPU instance, taking ~1 week to train.
10
 
 
 
11
  The version here is the fp16 HuggingFace model.
12
 
13
- In 8 bit mode, the model fits into 84% of A100 80GB (~67.2GB) 68747MiB
14
- In 4 bit mode, the model fits into 51% of A100 80GB (~40.8GB) 41559MiB
15
  500gb of RAM/Swap was required to merge the model.
16
 
17
  ## GGML & GPTQ versions
@@ -89,8 +93,6 @@ The road to hell is paved with good intentions, the current approach to AI Safet
89
 
90
 
91
  # Training code
92
- Special thanks to [George Sung](https://huggingface.co/georgesung) for creating [llama2_7b_chat_uncensored](https://huggingface.co/georgesung/llama2_7b_chat_uncensored).
93
-
94
  Code used to train the model is available [here](https://github.com/georgesung/llm_qlora).
95
 
96
  To reproduce the results:
@@ -131,6 +133,4 @@ model_output_dir: models/ # model saved in {model_output_dir}/{model_name}
131
  ```
132
 
133
  # Fine-tuning guide
134
- https://georgesung.github.io/ai/qlora-ift/
135
-
136
-
 
1
  ---
2
+ license: llama2
3
  datasets:
4
  - ehartford/wizard_vicuna_70k_unfiltered
5
+ tags:
6
+ - uncensored
7
  ---
8
 
9
  # Overview
10
  Fine-tuned [Llama-2 70B](https://huggingface.co/TheBloke/Llama-2-70B-fp16) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered).
11
  [QLoRA](https://arxiv.org/abs/2305.14314) was used for fine-tuning. The model was trained for three epochs on a single NVIDIA A100 80GB GPU instance, taking ~1 week to train.
12
 
13
+ Special thanks to [George Sung](https://huggingface.co/georgesung) for creating [llama2_7b_chat_uncensored](https://huggingface.co/georgesung/llama2_7b_chat_uncensored), and to [Eric Hartford](https://huggingface.co/ehartford/) for creating [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered)
14
+
15
  The version here is the fp16 HuggingFace model.
16
 
17
+ In 8 bit mode, the model fits into 84% of A100 80GB (67.2GB) 68747MiB
18
+ In 4 bit mode, the model fits into 51% of A100 80GB (40.8GB) 41559MiB
19
  500gb of RAM/Swap was required to merge the model.
20
 
21
  ## GGML & GPTQ versions
 
93
 
94
 
95
  # Training code
 
 
96
  Code used to train the model is available [here](https://github.com/georgesung/llm_qlora).
97
 
98
  To reproduce the results:
 
133
  ```
134
 
135
  # Fine-tuning guide
136
+ https://georgesung.github.io/ai/qlora-ift/