Dave17 commited on
Commit
229ba2a
·
verified ·
1 Parent(s): cf126e5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -115
README.md CHANGED
@@ -1,115 +0,0 @@
1
- ---
2
- language:
3
- - en
4
- license: cc-by-nc-4.0
5
- datasets:
6
- - facebook/asset
7
- - wi_locness
8
- - GEM/wiki_auto_asset_turk
9
- - discofuse
10
- - zaemyung/IteraTeR_plus
11
- - jfleg
12
- - grammarly/coedit
13
- metrics:
14
- - sari
15
- - bleu
16
- - accuracy
17
- widget:
18
- - text: 'Fix the grammar: When I grow up, I start to understand what he said is quite
19
- right.'
20
- example_title: Fluency
21
- - text: 'Make this text coherent: Their flight is weak. They run quickly through the
22
- tree canopy.'
23
- example_title: Coherence
24
- - text: 'Rewrite to make this easier to understand: A storm surge is what forecasters
25
- consider a hurricane''s most treacherous aspect.'
26
- example_title: Simplification
27
- - text: 'Paraphrase this: Do you know where I was born?'
28
- example_title: Paraphrase
29
- - text: 'Write this more formally: omg i love that song im listening to it right now'
30
- example_title: Formalize
31
- - text: 'Write in a more neutral way: The authors'' exposé on nutrition studies.'
32
- example_title: Neutralize
33
- ---
34
- # Model Card for CoEdIT-Large
35
-
36
- This model was obtained by fine-tuning the corresponding `google/flan-t5-large` model on the CoEdIT dataset. Details of the dataset can be found in our paper and repository.
37
-
38
- **Paper:** CoEdIT: Text Editing by Task-Specific Instruction Tuning
39
-
40
- **Authors:** Vipul Raheja, Dhruv Kumar, Ryan Koo, Dongyeop Kang
41
-
42
- ## Model Details
43
-
44
- ### Model Description
45
-
46
- - **Language(s) (NLP)**: English
47
- - **Finetuned from model:** google/flan-t5-large
48
-
49
- ### Model Sources
50
-
51
- - **Repository:** https://github.com/vipulraheja/coedit
52
- - **Paper:** https://arxiv.org/abs/2305.09857
53
-
54
- ## How to use
55
- We make available the models presented in our paper.
56
-
57
- <table>
58
- <tr>
59
- <th>Model</th>
60
- <th>Number of parameters</th>
61
- </tr>
62
- <tr>
63
- <td>CoEdIT-large</td>
64
- <td>770M</td>
65
- </tr>
66
- <tr>
67
- <td>CoEdIT-xl</td>
68
- <td>3B</td>
69
- </tr>
70
- <tr>
71
- <td>CoEdIT-xxl</td>
72
- <td>11B</td>
73
- </tr>
74
- </table>
75
-
76
-
77
- ## Uses
78
-
79
- ## Text Revision Task
80
- Given an edit instruction and an original text, our model can generate the edited version of the text.<br>
81
-
82
- ![task_specs](https://huggingface.co/grammarly/coedit-xl/resolve/main/task_examples.png)
83
-
84
- ## Usage
85
- ```python
86
- from transformers import AutoTokenizer, T5ForConditionalGeneration
87
-
88
- tokenizer = AutoTokenizer.from_pretrained("grammarly/coedit-large")
89
- model = T5ForConditionalGeneration.from_pretrained("grammarly/coedit-large")
90
- input_text = 'Fix grammatical errors in this sentence: When I grow up, I start to understand what he said is quite right.'
91
- input_ids = tokenizer(input_text, return_tensors="pt").input_ids
92
- outputs = model.generate(input_ids, max_length=256)
93
- edited_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
94
- ```
95
-
96
-
97
- #### Software
98
- https://github.com/vipulraheja/coedit
99
-
100
- ## Citation
101
-
102
- **BibTeX:**
103
- ```
104
- @article{raheja2023coedit,
105
- title={CoEdIT: Text Editing by Task-Specific Instruction Tuning},
106
- author={Vipul Raheja and Dhruv Kumar and Ryan Koo and Dongyeop Kang},
107
- year={2023},
108
- eprint={2305.09857},
109
- archivePrefix={arXiv},
110
- primaryClass={cs.CL}
111
- }
112
- ```
113
-
114
- **APA:**
115
- Raheja, V., Kumar, D., Koo, R., & Kang, D. (2023). CoEdIT: Text Editing by Task-Specific Instruction Tuning. ArXiv. /abs/2305.09857