File size: 12,778 Bytes
0a2e1be |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 |
---
library_name: setfit
tags:
- setfit
- sentence-transformers
- text-classification
- generated_from_setfit_trainer
metrics:
- accuracy
widget:
- text: 'The Alavas worked themselves to the bone in the last period , and English
and San Emeterio ( 65-75 ) had already made it clear that they were not going
to let anyone take away what they had earned during the first thirty minutes . '
- text: 'To break the uncomfortable silence , Haney began to talk . '
- text: 'For the treatment of non-small cell lung cancer , the effects of Alimta were
compared with those of docetaxel ( another anticancer medicine ) in one study
involving 571 patients with locally advanced or metastatic disease who had received
chemotherapy in the past . '
- text: 'As we all know , a few minutes before the end of the game ( that their team
had already won ) , both players deliberately wasted time which made the referee
show the second yellow card to both of them . '
- text: 'In contrast , patients whose cancer was affecting squamous cells had shorter
survival times if they received Alimta . '
pipeline_tag: text-classification
inference: true
base_model: sentence-transformers/paraphrase-mpnet-base-v2
model-index:
- name: SetFit with sentence-transformers/paraphrase-mpnet-base-v2
results:
- task:
type: text-classification
name: Text Classification
dataset:
name: Unknown
type: unknown
split: test
metrics:
- type: accuracy
value: 0.14172185430463577
name: Accuracy
---
# SetFit with sentence-transformers/paraphrase-mpnet-base-v2
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co./sentence-transformers/paraphrase-mpnet-base-v2) as the Sentence Transformer embedding model. A [SetFitHead](huggingface.co/docs/setfit/reference/main#setfit.SetFitHead) instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
## Model Details
### Model Description
- **Model Type:** SetFit
- **Sentence Transformer body:** [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co./sentence-transformers/paraphrase-mpnet-base-v2)
- **Classification head:** a [SetFitHead](huggingface.co/docs/setfit/reference/main#setfit.SetFitHead) instance
- **Maximum Sequence Length:** 512 tokens
- **Number of Classes:** 7 classes
<!-- - **Training Dataset:** [Unknown](https://huggingface.co./datasets/unknown) -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->
### Model Sources
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co./blog/setfit)
### Model Labels
| Label | Examples |
|:------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| 1 | <ul><li>'Eventually little French farmers and their little French farmwives came out of their stone houses and put their hands above their tiny eyes and squinted at us . '</li><li>'Mr. Neuberger realized that , although of Italian ancestry , Mr. Mariotta still could qualify as a minority person since he was born in Puerto Rico . '</li><li>"Biggest trouble was scared family who could n't get a phone line through , and spent a really horrible hour not knowing . "</li></ul> |
| 4 | <ul><li>'`` To ring for even one service at this tower , we have to scrape , `` says Mr. Hammond , a retired water-authority worker . `` '</li><li>"`` It 's my line of work `` , he said "</li><li>'One writer , signing his letter as `` Red-blooded , balanced male , `` remarked on the `` frequency of women fainting in peals , `` and suggested that they `` settle back into their traditional role of making tea at meetings . `` '</li></ul> |
| 5 | <ul><li>'Of course On Thursday , Haney mailed the monthly check for separate maintenance to his wife Lolly , and wished the stranger could do something about her '</li><li>"On the Right , the tone was set by Jacques Chirac , who declared in 1976 that `` 900,000 unemployed would not become a problem in a country with 2 million of foreign workers , '' and on the Left by Michel Rocard explaining in 1990 that France `` can not accommodate all the world 's misery . '' "</li><li>"But the council 's program to attract and train ringers is only partly successful , says Mr. Baldwin . "</li></ul> |
| 6 | <ul><li>'3 -RRB- Republican congressional representatives , because of their belief in a minimalist state , are less willing to engage in local benefit-seeking than are Democratic members of Congress . '</li><li>'As we know , voters tend to favor Republicans more in races for president than in those for Congress . '</li><li>'That is the way the system works . '</li></ul> |
| 2 | <ul><li>'-- Students should move up the educational ladder as their academic potential allows . '</li><li>'The next day , Sunday , the hangover reminded Haney where he had been the night before . '</li><li>'-- In most states , increasing expenditures on education , in our current circumstances , will probably make things worse , not better . '</li></ul> |
| 0 | <ul><li>'Then your focus will go to an input text box where you can type your function . '</li><li>"I might have got hit by that truck if it was n't for you . "</li><li>"Second , it explains why voters hold Congress in disdain but generally love their own congressional representatives : Any individual legislator 's constituents appreciate the specific benefits that the legislator wins for them but not the overall cost associated with every other legislator doing likewise for his own constituency . "</li></ul> |
| 3 | <ul><li>"It was the most exercise we 'd had all morning and it was followed by our driving immediately to the nearest watering hole . "</li><li>'Alimta is used together with cisplatin ( another anticancer medicine ) when the cancer is unresectable ( cannot be removed by surgery alone ) and malignant ( has spread , or is likely to spread easily , to other parts of the body ) , in patients who have not received chemotherapy ( medicines for cancer ) before advanced or metastatic non-small cell lung cancer that is not affecting the squamous cells . '</li><li>'If it is , it will be treated as an operator , if it is not , it will be treated as a user function . '</li></ul> |
## Evaluation
### Metrics
| Label | Accuracy |
|:--------|:---------|
| **all** | 0.1417 |
## Uses
### Direct Use for Inference
First install the SetFit library:
```bash
pip install setfit
```
Then you can load this model and run inference.
```python
from setfit import SetFitModel
# Download from the 🤗 Hub
model = SetFitModel.from_pretrained("HelgeKn/SemEval-multi-class-4")
# Run inference
preds = model("To break the uncomfortable silence , Haney began to talk . ")
```
<!--
### Downstream Use
*List how someone could finetune this model on their own dataset.*
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Set Metrics
| Training set | Min | Median | Max |
|:-------------|:----|:--------|:----|
| Word count | 4 | 27.1786 | 74 |
| Label | Training Sample Count |
|:------|:----------------------|
| 0 | 4 |
| 1 | 4 |
| 2 | 4 |
| 3 | 4 |
| 4 | 4 |
| 5 | 4 |
| 6 | 4 |
### Training Hyperparameters
- batch_size: (16, 16)
- num_epochs: (2, 2)
- max_steps: -1
- sampling_strategy: oversampling
- num_iterations: 20
- body_learning_rate: (2e-05, 2e-05)
- head_learning_rate: 2e-05
- loss: CosineSimilarityLoss
- distance_metric: cosine_distance
- margin: 0.25
- end_to_end: False
- use_amp: False
- warmup_proportion: 0.1
- seed: 42
- eval_max_steps: -1
- load_best_model_at_end: False
### Training Results
| Epoch | Step | Training Loss | Validation Loss |
|:------:|:----:|:-------------:|:---------------:|
| 0.0143 | 1 | 0.2446 | - |
| 0.7143 | 50 | 0.0612 | - |
| 1.4286 | 100 | 0.0078 | - |
### Framework Versions
- Python: 3.9.13
- SetFit: 1.0.1
- Sentence Transformers: 2.2.2
- Transformers: 4.36.0
- PyTorch: 2.1.1+cpu
- Datasets: 2.15.0
- Tokenizers: 0.15.0
## Citation
### BibTeX
```bibtex
@article{https://doi.org/10.48550/arxiv.2209.11055,
doi = {10.48550/ARXIV.2209.11055},
url = {https://arxiv.org/abs/2209.11055},
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Efficient Few-Shot Learning Without Prompts},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> |