Update README.md
Browse files
README.md
CHANGED
@@ -37,7 +37,7 @@ For an automated evaluation of this model, see the *Sentence Embeddings Benchmar
|
|
37 |
The model was trained with the parameters:
|
38 |
|
39 |
**공통**
|
40 |
-
- **do_lower_case=1, correct_bios=0, polling_mode=
|
41 |
|
42 |
**1.STS**
|
43 |
- 말뭉치 : korsts(5,749) + kluestsV1.1(11,668) + stsb_multi_mt(5,749) + mteb/sickr-sts(9,927) + glue stsb(5,749) (총:38,842)
|
@@ -47,7 +47,7 @@ The model was trained with the parameters:
|
|
47 |
**2.distilation**
|
48 |
- 교사 모델 : paraphrase-multilingual-mpnet-base-v2(max_token_len:128)
|
49 |
- 말뭉치 : news_talk_en_ko_train.tsv (영어-한국어 대화-뉴스 병렬 말뭉치 : 1.38M)
|
50 |
-
- Param : **lr: 5e-5, epochs: 10, train_batch: 128, eval/test_batch: 64, max_token_len: 128(교사모델이 128이므로 맟춰줌)**
|
51 |
- 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/sbert/sbert-distillaton.ipynb) 참조
|
52 |
|
53 |
**3.NLI**
|
|
|
37 |
The model was trained with the parameters:
|
38 |
|
39 |
**공통**
|
40 |
+
- **do_lower_case=1, correct_bios=0, polling_mode=mean**
|
41 |
|
42 |
**1.STS**
|
43 |
- 말뭉치 : korsts(5,749) + kluestsV1.1(11,668) + stsb_multi_mt(5,749) + mteb/sickr-sts(9,927) + glue stsb(5,749) (총:38,842)
|
|
|
47 |
**2.distilation**
|
48 |
- 교사 모델 : paraphrase-multilingual-mpnet-base-v2(max_token_len:128)
|
49 |
- 말뭉치 : news_talk_en_ko_train.tsv (영어-한국어 대화-뉴스 병렬 말뭉치 : 1.38M)
|
50 |
+
- Param : **lr: 5e-5, eps: 1e-8, epochs: 10, train_batch: 128, eval/test_batch: 64, max_token_len: 128(교사모델이 128이므로 맟춰줌)**
|
51 |
- 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/sbert/sbert-distillaton.ipynb) 참조
|
52 |
|
53 |
**3.NLI**
|