Taishi-N324 commited on
Commit
ce4b414
1 Parent(s): 2ffaf2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -184,7 +184,7 @@ The following instruction datasets were used for the instruction tuning.
184
  - Japanese
185
  - `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`
186
  - Single-turn Japanese synthetic instruction dataset derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) dataset [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)).
187
- The first-turn user instructions were translated into Japanese via DeepL machine translation, and the assistant responses were generated using the Llama 3.1 405B Instruct model. Rejection sampling (n=6) was applied, with Llama 3.1 70B Instruct serving as a judge.
188
  - As implied by the dataset name, conversations that contain personally identifiable information (PII) or template-based user instructions have been removed. Duplicate instuctions have also been removed.
189
  - `filtered-magpie-ultra-ja`
190
  - A Japanese variant of the `filtered-magpie-ultra-en` dataset, machine-translated into Japanese using the [gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it).
@@ -194,7 +194,7 @@ The following instruction datasets were used for the instruction tuning.
194
  - `lmsys-chat-1m-synth-en-wo-pii-and-template-instructions`
195
  - Similar to the `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied in this version.
196
  - `filtered-magpie-ultra-en`
197
- - A subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, developed following the MAGPIE recipe [\[Xu+, arXiv24\]](https://arxiv.org/abs/2406.08464) using Llama 3.1 405B Instruct. This subset includes only samples rated as 'average,' 'good,' or 'excellent.'
198
 
199
 
200
  ## Risks and Limitations
 
184
  - Japanese
185
  - `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`
186
  - Single-turn Japanese synthetic instruction dataset derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) dataset [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)).
187
+ The first-turn user instructions were translated into Japanese via DeepL machine translation, and the assistant responses were generated using the [Llama-3.1-405B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct) model. Rejection sampling (n=6) was applied, with [Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) serving as a judge.
188
  - As implied by the dataset name, conversations that contain personally identifiable information (PII) or template-based user instructions have been removed. Duplicate instuctions have also been removed.
189
  - `filtered-magpie-ultra-ja`
190
  - A Japanese variant of the `filtered-magpie-ultra-en` dataset, machine-translated into Japanese using the [gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it).
 
194
  - `lmsys-chat-1m-synth-en-wo-pii-and-template-instructions`
195
  - Similar to the `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied in this version.
196
  - `filtered-magpie-ultra-en`
197
+ - A subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, developed following the MAGPIE recipe [\[Xu+, arXiv24\]](https://arxiv.org/abs/2406.08464) using [Llama-3.1-405B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct). This subset includes only samples rated as 'average,' 'good,' or 'excellent.'
198
 
199
 
200
  ## Risks and Limitations