stephenlzc
commited on
Commit
•
a25de57
1
Parent(s):
1c73e76
Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,4 @@
|
|
1 |
---
|
2 |
-
|
3 |
datasets:
|
4 |
- Minami-su/toxic-sft-zh
|
5 |
- llm-wizard/alpaca-gpt4-data-zh
|
@@ -12,19 +11,27 @@ tags:
|
|
12 |
- text-generation-inference
|
13 |
- code
|
14 |
- unsloth
|
|
|
15 |
task_categories:
|
16 |
- conversational
|
17 |
base_model: shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat
|
18 |
-
|
19 |
widget:
|
20 |
-
- text:
|
21 |
-
|
22 |
-
|
23 |
-
example_title:
|
24 |
-
- text:
|
25 |
-
|
26 |
-
|
27 |
-
example_title:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
---
|
29 |
|
30 |
|
@@ -32,35 +39,34 @@ widget:
|
|
32 |
|
33 |
### Model Description
|
34 |
|
35 |
-
Using **shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat** as base model, and finetune the dataset as mentioned.
|
36 |
|
37 |
-
## Training
|
38 |
-
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1K9stY8LMVcySG0jDMYZdWQCFPfoDFBL-?usp=sharing)
|
39 |
|
40 |
## Training Procedure Raw Files
|
41 |
|
42 |
-
ALL the procedure are training on
|
43 |
-
|
|
|
|
|
44 |
|
45 |
-
**Hardware**:
|
46 |
**GPU**: 1x A100 SXM4 80GB
|
|
|
47 |
**CPU**: AMD EPYC 7513 32-Core Processor
|
|
|
48 |
**RAM**: 129 GB
|
|
|
49 |
**Docker Image**: pytorch/pytorch:2.2.0-cuda12.1-cudnn8-devel
|
50 |
-
|
51 |
-
```python
|
52 |
-
{
|
53 |
-
"DATA_DIRECTORY": "/workspace/",
|
54 |
-
"JUPYTER_DIR": "/"
|
55 |
-
}
|
56 |
-
```
|
57 |
**[ipynb file](https://huggingface.co/stephenlzc/dolphin-llama3-zh-cn-uncensored/blob/main/config.json)**
|
58 |
|
|
|
59 |
### Training Data
|
60 |
**Base Model**
|
61 |
-
[shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat](https://huggingface.co/shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat)
|
62 |
|
63 |
**Dataset**
|
64 |
-
- [Minami-su/toxic-sft-zh](https://huggingface.co/datasets/Minami-su/toxic-sft-zh)
|
65 |
-
- [llm-wizard/alpaca-gpt4-data-zh](https://huggingface.co/datasets/llm-wizard/alpaca-gpt4-data-zh)
|
66 |
-
- [stephenlzc/stf-alpaca](https://huggingface.co/datasets/stephenlzc/stf-alpaca)
|
|
|
1 |
---
|
|
|
2 |
datasets:
|
3 |
- Minami-su/toxic-sft-zh
|
4 |
- llm-wizard/alpaca-gpt4-data-zh
|
|
|
11 |
- text-generation-inference
|
12 |
- code
|
13 |
- unsloth
|
14 |
+
- uncensored
|
15 |
task_categories:
|
16 |
- conversational
|
17 |
base_model: shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat
|
|
|
18 |
widget:
|
19 |
+
- text: >-
|
20 |
+
Is this review positive or negative? Review: Best cast iron skillet you will
|
21 |
+
ever buy.
|
22 |
+
example_title: Sentiment analysis
|
23 |
+
- text: >-
|
24 |
+
Barack Obama nominated Hilary Clinton as his secretary of state on Monday.
|
25 |
+
He chose her because she had ...
|
26 |
+
example_title: Coreference resolution
|
27 |
+
- text: >-
|
28 |
+
On a shelf, there are five books: a gray book, a red book, a purple book, a
|
29 |
+
blue book, and a black book ...
|
30 |
+
example_title: Logic puzzles
|
31 |
+
- text: >-
|
32 |
+
The two men running to become New York City's next mayor will face off in
|
33 |
+
their first debate Wednesday night ...
|
34 |
+
example_title: Reading comprehension
|
35 |
---
|
36 |
|
37 |
|
|
|
39 |
|
40 |
### Model Description
|
41 |
|
42 |
+
Using **shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat** as base model, and finetune the dataset as mentioned. Makes the model uncensored.
|
43 |
|
44 |
+
## Training Code
|
45 |
+
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1K9stY8LMVcySG0jDMYZdWQCFPfoDFBL-?usp=sharing)
|
46 |
|
47 |
## Training Procedure Raw Files
|
48 |
|
49 |
+
ALL the procedure are training on **[Vast.ai](https://cloud.vast.ai/?ref_id=138637)**
|
50 |
+
|
51 |
+
|
52 |
+
**Hardware in Vast.ai**:
|
53 |
|
|
|
54 |
**GPU**: 1x A100 SXM4 80GB
|
55 |
+
|
56 |
**CPU**: AMD EPYC 7513 32-Core Processor
|
57 |
+
|
58 |
**RAM**: 129 GB
|
59 |
+
|
60 |
**Docker Image**: pytorch/pytorch:2.2.0-cuda12.1-cudnn8-devel
|
61 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
**[ipynb file](https://huggingface.co/stephenlzc/dolphin-llama3-zh-cn-uncensored/blob/main/config.json)**
|
63 |
|
64 |
+
|
65 |
### Training Data
|
66 |
**Base Model**
|
67 |
+
[shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat](https://huggingface.co/shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat)
|
68 |
|
69 |
**Dataset**
|
70 |
+
- [Minami-su/toxic-sft-zh](https://huggingface.co/datasets/Minami-su/toxic-sft-zh)
|
71 |
+
- [llm-wizard/alpaca-gpt4-data-zh](https://huggingface.co/datasets/llm-wizard/alpaca-gpt4-data-zh)
|
72 |
+
- [stephenlzc/stf-alpaca](https://huggingface.co/datasets/stephenlzc/stf-alpaca)
|