quantumaikr commited on
Commit
d61eccf
โ€ข
1 Parent(s): e9c2e17

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +71 -0
README.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ko
4
+ - en
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - orca
8
+ - llama-2
9
+ ---
10
+
11
+
12
+ <p align="center" width="100%">
13
+ <img src="https://i.imgur.com/snFDU0P.png" alt="KoreanLM icon" style="width: 500px; display: block; margin: auto; border-radius: 10%;">
14
+ </p>
15
+
16
+
17
+
18
+ # quantumaikr/llama-2-70b-fb16-korean
19
+
20
+ ## Model Description
21
+
22
+ `quantumaikr/llama-2-70b-fb16-korean` is a Llama2 70B model finetuned the Korean Dataset
23
+
24
+ ## Usage
25
+
26
+ Start chatting with `quantumaikr/llama-2-70b-fb16-korean` using the following code snippet:
27
+
28
+ ```python
29
+ import torch
30
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
31
+
32
+ tokenizer = AutoTokenizer.from_pretrained("quantumaikr/llama-2-70b-fb16-korean")
33
+ model = AutoModelForCausalLM.from_pretrained("quantumaikr/llama-2-70b-fb16-korean", torch_dtype=torch.float16, device_map="auto")
34
+
35
+ system_prompt = "### System:\n๊ท€ํ•˜๋Š” ์ง€์‹œ๋ฅผ ๋งค์šฐ ์ž˜ ๋”ฐ๋ฅด๋Š” AI์ธ QuantumLM์ž…๋‹ˆ๋‹ค. ์ตœ๋Œ€ํ•œ ๋งŽ์ด ๋„์™€์ฃผ์„ธ์š”. ์•ˆ์ „์— ์œ ์˜ํ•˜๊ณ  ๋ถˆ๋ฒ•์ ์ธ ํ–‰๋™์€ ํ•˜์ง€ ๋งˆ์„ธ์š”.\n\n"
36
+
37
+ message = "์ธ๊ณต์ง€๋Šฅ์ด๋ž€ ๋ฌด์—‡์ธ๊ฐ€์š”?"
38
+ prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
39
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
40
+ output = model.generate(**inputs, do_sample=True, temperature=0.9, top_p=0.75, max_new_tokens=4096)
41
+
42
+ print(tokenizer.decode(output[0], skip_special_tokens=True))
43
+ ```
44
+
45
+ QuantumLM should be used with this prompt format:
46
+ ```
47
+ ### System:
48
+ This is a system prompt, please behave and help the user.
49
+
50
+ ### User:
51
+ Your prompt here
52
+
53
+ ### Assistant
54
+ The output of QuantumLM
55
+ ```
56
+
57
+
58
+
59
+ ## Use and Limitations
60
+
61
+ ### Intended Use
62
+
63
+ These models are intended for research only, in adherence with the [CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license.
64
+
65
+ ### Limitations and bias
66
+
67
+ Although the aforementioned dataset helps to steer the base language models into "safer" distributions of text, not all biases and toxicity can be mitigated through fine-tuning. We ask that users be mindful of such potential issues that can arise in generated responses. Do not treat model outputs as substitutes for human judgment or as sources of truth. Please use it responsibly.
68
+
69
+
70
+
71
+ Contact us : [email protected]