Update README.md
Browse files
README.md
CHANGED
@@ -8,6 +8,7 @@ thumbnail: https://huggingface.co/front/thumbnails/microsoft.png
|
|
8 |
license: mit
|
9 |
widget:
|
10 |
- text: "[CLS] I love you. [SEP] I like you. [SEP]"
|
|
|
11 |
---
|
12 |
|
13 |
## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
|
@@ -40,9 +41,7 @@ We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks.
|
|
40 |
```bash
|
41 |
cd transformers/examples/text-classification/
|
42 |
export TASK_NAME=mrpc
|
43 |
-
python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge
|
44 |
-
--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\
|
45 |
-
--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16
|
46 |
```
|
47 |
|
48 |
### Citation
|
|
|
8 |
license: mit
|
9 |
widget:
|
10 |
- text: "[CLS] I love you. [SEP] I like you. [SEP]"
|
11 |
+
pipeline_tag: zero-shot-classification
|
12 |
---
|
13 |
|
14 |
## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
|
|
|
41 |
```bash
|
42 |
cd transformers/examples/text-classification/
|
43 |
export TASK_NAME=mrpc
|
44 |
+
python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\\n--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\\n--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16
|
|
|
|
|
45 |
```
|
46 |
|
47 |
### Citation
|