--- license: mit task_categories: - text2text-generation language: - en - zh configs: - config_name: chatgpt-2000 default: true data_files: - split: train path: "chatgpt-train-2000.jsonl" - split: test path: "chatgpt-test.jsonl" - config_name: chatgpt-8192 data_files: - split: train path: "chatgpt-train-8192.jsonl" - split: test path: "chatgpt-test-8192.jsonl" --- # Introduction This repository holds the data file for translating TechLinked, which talks about mostly technology and science news. Raw data is in the data/ folder. Scripts generate OpenAI's ChatCompletion Fine-tuning API formatted training data in `jsonl` format. `-2000` variants are designed to be used with GPT-3 with 8192 tokens context length limit. `-8192` variants are designed to be used with GPT-4o mini with 128000 context window and 16384 max output tokens. # How to add data to this repository 0. `pip install ass` 1. Convert ASS file into `.en.txt` and `.cn.txt` files: `python ./ass_extract.py [ASS Filename]` This step will generate two files: `Extracted - [Filename].en.txt` and `Extracted - [Filename].cn.txt` 2. Move them into the `data/` folder. You may want to rename them also, but keep their filenames the same except `.en` and `.cn`. 3. Run script: `python ./generate_chatgpt_varlen data --maxlen MAXLEN --test-ratio TEST_RATIO` - `data` is the data directory. - `MAXLEN` is recommended to be a quarter of the context window, or a little bit less than maximum output tokens, whichever is smaller. - `TEST_RATIO` is the ratio of data to be reserved for testing. A decimal number. Default is 0.2 This will generate three files: - `combined-{MAXLEN}.jsonl`: Test+Train data. - `chatgpt-train-{MAXLEN}.jsonl`: Train data. - `chatgpt-test-{MAXLEN}.jsonl`: Test data. The other scripts are deprecated.