Taishi-N324 commited on
Commit
dd5f2ac
1 Parent(s): da45946

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -13
README.md CHANGED
@@ -6,9 +6,12 @@ library_name: transformers
6
  pipeline_tag: text-generation
7
  license: llama3.1
8
  model_type: llama
 
 
 
9
  ---
10
 
11
- # Llama3.1 Swallow
12
 
13
  Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models.
14
  Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities.
@@ -197,20 +200,23 @@ print(output[0].outputs[0].text)
197
 
198
  ### Instruction Tuning
199
 
200
- The following datasets were used for the instruction tuning.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
 
202
- - lmsys-chat-1m-synth-ja-wo-pii
203
-
204
- - Japanese translation of the lmsys-chat-1m dataset using DeepL, with synthetic instruction data created using the Llama-3.1-405B model.
205
- - 'wo-pii' indicates removal of personally identifiable information.
206
 
207
- - filtered magpie-ultra
208
- - Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
209
-
210
- - gemma-magpie
211
- - Japanese dataset.
212
- - Generated using prompts for specific category words.
213
-
214
  ## Risks and Limitations
215
 
216
  The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
 
6
  pipeline_tag: text-generation
7
  license: llama3.1
8
  model_type: llama
9
+ datasets:
10
+ - lmsys/lmsys-chat-1m
11
+ - argilla/magpie-ultra-v0.1
12
  ---
13
 
14
+ # Llama 3.1 Swallow
15
 
16
  Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models.
17
  Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities.
 
200
 
201
  ### Instruction Tuning
202
 
203
+ The following instruction datasets were used for the instruction tuning.
204
+ - Japanese
205
+ - `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`
206
+ - Single-turn Japanese synthetic instruction dataset derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) dataset [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)).
207
+ The first-turn user instructions were translated into Japanese via DeepL machine translation, and the assistant responses were generated using the Llama 3.1 405B Instruct model. Rejection sampling (n=6) was applied, with Llama 3.1 70B Instruct serving as a judge.
208
+ - As implied by the dataset name, conversations that contain personally identifiable information (PII) or template-based user instructions have been removed. Duplicate instuctions have also been removed.
209
+ - `filtered-magpie-ultra-ja`
210
+ - A Japanese variant of the `filtered-magpie-ultra-en` dataset, machine-translated into Japanese using the Gemma 2 27B IT.
211
+ - `gemma-magpie`
212
+ - Japanese Q&A dataset on diverse topics, generated using prompts with specific category words, with answers by Gemma 2 27B IT, heuristically filtered for quality and length.
213
+ - English
214
+ - `lmsys-chat-1m-synth-en-wo-pii-and-template-instructions`
215
+ - Similar to the `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied in this version.
216
+ - `filtered-magpie-ultra-en`
217
+ - A subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, developed following the MAGPIE recipe [\[Xu+, arXiv24\]](https://arxiv.org/abs/2406.08464) using Llama 3.1 405B Instruct. This subset includes only samples rated as 'average,' 'good,' or 'excellent.'
218
 
 
 
 
 
219
 
 
 
 
 
 
 
 
220
  ## Risks and Limitations
221
 
222
  The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.