Netta1994 commited on
Commit
f77d898
1 Parent(s): cb7b7fd

Add SetFit model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,321 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: BAAI/bge-base-en-v1.5
3
+ library_name: setfit
4
+ metrics:
5
+ - accuracy
6
+ pipeline_tag: text-classification
7
+ tags:
8
+ - setfit
9
+ - sentence-transformers
10
+ - text-classification
11
+ - generated_from_setfit_trainer
12
+ widget:
13
+ - text: 'Reasoning:
14
+
15
+
16
+ **Why the answer may be good:**
17
+
18
+ - Context Grounding: The document provides specific information that the College
19
+ of Arts and Letters was established in 1842. The answer given in the response
20
+ is directly supported by the document.
21
+
22
+ - Relevance: The answer addresses the specific question asked by providing the
23
+ year the college was created.
24
+
25
+ - Conciseness: The answer is clear, precise, and straight to the point.
26
+
27
+
28
+ **Why the answer may be bad:**
29
+
30
+ - There does not appear to be any reasons why the answer may be bad based on the
31
+ criteria specified.
32
+
33
+
34
+ Final result: ****'
35
+ - text: 'The answer provided is:
36
+
37
+
38
+ "The average student at Notre Dame travels more than 750 miles to study there."
39
+
40
+
41
+ Reasoning:
42
+
43
+
44
+ **Good points:**
45
+
46
+ 1. **Context Grounding**: The answer is supported by information present in the
47
+ document, which states, "the average student traveled more than 750 miles to Notre
48
+ Dame".
49
+
50
+ 2. **Relevance**: The answer directly addresses the specific question asking about
51
+ the number of miles the average student travels to study at Notre Dame.
52
+
53
+ 3. **Conciseness**: The answer is clear and to the point without any unnecessary
54
+ information.
55
+
56
+
57
+ **Bad points:**
58
+
59
+ - There are no bad points in this case as the answer aligns perfectly with all
60
+ the evaluation criteria.
61
+
62
+
63
+ Final Result: ****'
64
+ - text: 'Reasoning why the answer may be good:
65
+
66
+ - The answer correctly identifies Mick LaSalle as the writer for the San Francisco
67
+ Chronicle.
68
+
69
+ - The answer states that Mick LaSalle awarded "Spectre" a perfect score, which
70
+ is supported by the document.
71
+
72
+
73
+ Reasoning why the answer may be bad:
74
+
75
+ - The answer is concise and to the point, fulfilling the criteria for conciseness
76
+ and relevance.
77
+
78
+ - The document provided confirms that Mick LaSalle gave "Spectre" a perfect score
79
+ of 100.
80
+
81
+ - There is no deviation into unrelated topics, maintaining focus on the question
82
+ asked.
83
+
84
+
85
+ Final result:'
86
+ - text: 'Reasoning why the answer may be good:
87
+
88
+ 1. Context Grounding: The document does mention that The Review of Politics was
89
+ inspired by German Catholic journals.
90
+
91
+ 2. Relevance: The answer addresses the specific question about what inspired The
92
+ Review of Politics.
93
+
94
+
95
+ Reasoning why the answer may be bad:
96
+
97
+ 1. Context Grounding: The document does not support the claim that it predominantly
98
+ featured articles written by Karl Marx. In fact, none of the intellectual leaders
99
+ mentioned in the document are Karl Marx, and the document emphasizes a Catholic
100
+ intellectual revival, which is inconsistent with Marx''s philosophy.
101
+
102
+ 2. Conciseness: The additional information about Karl Marx is not needed and is
103
+ misleading, detracting from the core answer.
104
+
105
+
106
+ Final Result: Bad
107
+
108
+
109
+ The overall response, despite having a relevant and correct part, is ultimately
110
+ flawed due to significant inaccuracies and irrelevant information.'
111
+ - text: 'Reasoning why the answer may be good:
112
+
113
+ - The answer directly addresses the question by providing the specific position
114
+ Forbes.com placed Notre Dame among US research universities.
115
+
116
+ - It uses information directly from the provided document to support the claim.
117
+
118
+
119
+ Reasoning why the answer may be bad:
120
+
121
+ - There are no apparent reasons why the answer would be considered bad, as it
122
+ adheres to all evaluation criteria.
123
+
124
+
125
+ Final result:'
126
+ inference: true
127
+ model-index:
128
+ - name: SetFit with BAAI/bge-base-en-v1.5
129
+ results:
130
+ - task:
131
+ type: text-classification
132
+ name: Text Classification
133
+ dataset:
134
+ name: Unknown
135
+ type: unknown
136
+ split: test
137
+ metrics:
138
+ - type: accuracy
139
+ value: 0.95
140
+ name: Accuracy
141
+ ---
142
+
143
+ # SetFit with BAAI/bge-base-en-v1.5
144
+
145
+ This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
146
+
147
+ The model has been trained using an efficient few-shot learning technique that involves:
148
+
149
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
150
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
151
+
152
+ ## Model Details
153
+
154
+ ### Model Description
155
+ - **Model Type:** SetFit
156
+ - **Sentence Transformer body:** [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5)
157
+ - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
158
+ - **Maximum Sequence Length:** 512 tokens
159
+ - **Number of Classes:** 2 classes
160
+ <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
161
+ <!-- - **Language:** Unknown -->
162
+ <!-- - **License:** Unknown -->
163
+
164
+ ### Model Sources
165
+
166
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
167
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
168
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
169
+
170
+ ### Model Labels
171
+ | Label | Examples |
172
+ |:------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
173
+ | 1 | <ul><li>"Reasoning why the answer may be good:\n1. **Context Grounding**: The answer is well-supported by the provided document and directly quotes relevant information about Patricia Wallace's roles and responsibilities.\n2. **Relevance**: The answer specifically addresses the question asked, detailing the roles and responsibilities of Patricia Wallace without deviating into unrelated topics.\n3. **Conciseness**: The answer is clear, concise, and focuses on the main points relevant to the question, avoiding unnecessary information.\n\nReasoning why the answer may be bad:\n- There is no significant reason to consider the answer bad based on the given criteria. It comprehensively covers the roles and responsibilities of Patricia Wallace as mentioned in the document.\n\nFinal Result:"</li><li>'### Reasoning:\n**Why the answer may be good:**\n1. **Context Grounding:** The answer is directly taken from the document, which states that a dime is one-tenth of a dollar.\n2. **Relevance:** The answer addresses the specific question asked about the monetary value of a dime.\n3. **Conciseness:** The answer is clear and to the point, providing no more information than necessary.\n\n**Why the answer may be bad:**\n1. **Context Grounding:** The document provides additional context and details about the U.S. dollar system which were not included in the answer. However, these details are not directly necessary to answer the question.\n2. **Relevance:** No deviation or unrelated topics are present in the answer. \n3. **Conciseness:** The answer avoids unnecessary information, maintaining itsclarity and brevity. \n\n### Final Result:\n****'</li><li>'Reasoning why the answer may be good:\n- Context Grounding: The answer refers to symptoms like flu-like signs, which are detailed in the provided document. It also mentions the connection with tampon use, the presence of rashes, and the seriousness of seeking medical help, all of which are discussed in the document.\n- Relevance: The answer addresses the question by listing symptoms and highlighting the importance of recognizing them, which directly corresponds to the question asked.\n- Conciseness: The answer is relatively concise while covering most of the essential details related to recognizing TSS.\n\nReasoning why the answer may be bad:\n- Context Grounding: While the answer does mention flu-like symptoms and the association with tampon use, it lacks specific details like fever and other visible signs mentioned in the document.\n- Relevance: The mention of treatment with antibiotics is somewhat relevant but moves slightly away from the specific focus of how to recognize TSS.\n- Conciseness: The answer could be streamlined further by focusing more on the core question of identifying symptoms rather than mentioning treatment.\n\nFinal Result:'</li></ul> |
174
+ | 0 | <ul><li>'**Reasoning:**\n\n**Why the answer may be good:**\n1. **Context Grounding:** The answer does affirm Gregory Johnson as the CEO of Franklin Templeton Investments, which is supported by the provided document.\n2. **Relevance:** The answer directly addresses the question regarding the CEO of Franklin Templeton Investments.\n3. **Conciseness:** The answer is relatively clear and to the point, providing the name of the CEO as requested.\n\n**Why the answer may be bad:**\n1. **Context Grounding:** The statement about Gregory Johnson inheriting the position from his father, Rupert H. Johnson, Sr., is not mentioned in the provided document.\n2. **Relevance:** While the primary answer is correct and relevant, the additional information about the inheritance is not relevant to the specific question asked.\n3. **Conciseness:** The answer includes unnecessary information about the inheritance of the position, which was not part of the question.\n\n**Final result:**'</li><li>'Reasoning why the answer may be good:\n1. The answer is well-supported by the provided document, mentioning key steps in diagnosis and treatment such as taking the cat to the vet, using topical antibiotics and anti-inflammatory medications, completing the full course of treatment, and isolating the infected cat.\n2. It directly addresses the specific question of how to treat conjunctivitis in cats.\n3. The answer is clear and to the point, providing practical advice on treatment.\n\nReasoning why the answer may be bad:\n1. The mention of conjunctivitis in cats often resulting from exposure to a rare type of pollen found only in the Amazon rainforest is not supported by the document. This statement is factually incorrect and detracts from the overall accuracy.\n2. It could be more concise by avoiding unnecessary information and focusing solely on the mostcritical points of treatment.\n\nFinal result:'</li><li>"Reasoning why the answer may be good: \n- The answer correctly identifies the College of Arts and Letters as Notre Dame's first college, founded in 1842, which is directly related to the question asked.\n\nReasoning why the answer may be bad:\n- The answer includes an incorrect and unsupported statement about the curriculum for time travel studies, which is not mentioned in the provided document andis irrelevant to the question.\n\nFinal result:"</li></ul> |
175
+
176
+ ## Evaluation
177
+
178
+ ### Metrics
179
+ | Label | Accuracy |
180
+ |:--------|:---------|
181
+ | **all** | 0.95 |
182
+
183
+ ## Uses
184
+
185
+ ### Direct Use for Inference
186
+
187
+ First install the SetFit library:
188
+
189
+ ```bash
190
+ pip install setfit
191
+ ```
192
+
193
+ Then you can load this model and run inference.
194
+
195
+ ```python
196
+ from setfit import SetFitModel
197
+
198
+ # Download from the 🤗 Hub
199
+ model = SetFitModel.from_pretrained("Netta1994/setfit_baai_squad_gpt-4o_improved-cot-instructions_two_reasoning_remove_final_evaluat")
200
+ # Run inference
201
+ preds = model("Reasoning why the answer may be good:
202
+ - The answer directly addresses the question by providing the specific position Forbes.com placed Notre Dame among US research universities.
203
+ - It uses information directly from the provided document to support the claim.
204
+
205
+ Reasoning why the answer may be bad:
206
+ - There are no apparent reasons why the answer would be considered bad, as it adheres to all evaluation criteria.
207
+
208
+ Final result:")
209
+ ```
210
+
211
+ <!--
212
+ ### Downstream Use
213
+
214
+ *List how someone could finetune this model on their own dataset.*
215
+ -->
216
+
217
+ <!--
218
+ ### Out-of-Scope Use
219
+
220
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
221
+ -->
222
+
223
+ <!--
224
+ ## Bias, Risks and Limitations
225
+
226
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
227
+ -->
228
+
229
+ <!--
230
+ ### Recommendations
231
+
232
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
233
+ -->
234
+
235
+ ## Training Details
236
+
237
+ ### Training Set Metrics
238
+ | Training set | Min | Median | Max |
239
+ |:-------------|:----|:---------|:----|
240
+ | Word count | 50 | 125.2071 | 274 |
241
+
242
+ | Label | Training Sample Count |
243
+ |:------|:----------------------|
244
+ | 0 | 95 |
245
+ | 1 | 103 |
246
+
247
+ ### Training Hyperparameters
248
+ - batch_size: (16, 16)
249
+ - num_epochs: (1, 1)
250
+ - max_steps: -1
251
+ - sampling_strategy: oversampling
252
+ - num_iterations: 20
253
+ - body_learning_rate: (2e-05, 2e-05)
254
+ - head_learning_rate: 2e-05
255
+ - loss: CosineSimilarityLoss
256
+ - distance_metric: cosine_distance
257
+ - margin: 0.25
258
+ - end_to_end: False
259
+ - use_amp: False
260
+ - warmup_proportion: 0.1
261
+ - l2_weight: 0.01
262
+ - seed: 42
263
+ - eval_max_steps: -1
264
+ - load_best_model_at_end: False
265
+
266
+ ### Training Results
267
+ | Epoch | Step | Training Loss | Validation Loss |
268
+ |:------:|:----:|:-------------:|:---------------:|
269
+ | 0.0020 | 1 | 0.1499 | - |
270
+ | 0.1010 | 50 | 0.2586 | - |
271
+ | 0.2020 | 100 | 0.2524 | - |
272
+ | 0.3030 | 150 | 0.1409 | - |
273
+ | 0.4040 | 200 | 0.0305 | - |
274
+ | 0.5051 | 250 | 0.015 | - |
275
+ | 0.6061 | 300 | 0.0097 | - |
276
+ | 0.7071 | 350 | 0.0108 | - |
277
+ | 0.8081 | 400 | 0.0054 | - |
278
+ | 0.9091 | 450 | 0.0047 | - |
279
+
280
+ ### Framework Versions
281
+ - Python: 3.10.14
282
+ - SetFit: 1.1.0
283
+ - Sentence Transformers: 3.1.1
284
+ - Transformers: 4.44.0
285
+ - PyTorch: 2.4.0+cu121
286
+ - Datasets: 3.0.0
287
+ - Tokenizers: 0.19.1
288
+
289
+ ## Citation
290
+
291
+ ### BibTeX
292
+ ```bibtex
293
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
294
+ doi = {10.48550/ARXIV.2209.11055},
295
+ url = {https://arxiv.org/abs/2209.11055},
296
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
297
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
298
+ title = {Efficient Few-Shot Learning Without Prompts},
299
+ publisher = {arXiv},
300
+ year = {2022},
301
+ copyright = {Creative Commons Attribution 4.0 International}
302
+ }
303
+ ```
304
+
305
+ <!--
306
+ ## Glossary
307
+
308
+ *Clearly define terms in order to be accessible across audiences.*
309
+ -->
310
+
311
+ <!--
312
+ ## Model Card Authors
313
+
314
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
315
+ -->
316
+
317
+ <!--
318
+ ## Model Card Contact
319
+
320
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
321
+ -->
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "BAAI/bge-base-en-v1.5",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "pad_token_id": 0,
26
+ "position_embedding_type": "absolute",
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.44.0",
29
+ "type_vocab_size": 2,
30
+ "use_cache": true,
31
+ "vocab_size": 30522
32
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.1.1",
4
+ "transformers": "4.44.0",
5
+ "pytorch": "2.4.0+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
config_setfit.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "normalize_embeddings": false,
3
+ "labels": null
4
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a59b4985c598f59cff9a9d0be78f4b0ce7817698195eb8fc888f1ea5420e6b46
3
+ size 437951328
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe9332b93f18bf96d6102132cc8f2a14c6d8c8b7040d09e4ca26676a2419c1fb
3
+ size 7007
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": true
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff