dipteshkanojia's picture
Librarian Bot: Add base_model information to model (#1)
0f4d23d
|
raw
history blame
2.78 kB
metadata
tags:
  - generated_from_trainer
metrics:
  - accuracy
  - precision
  - recall
  - f1
model_creators:
  - Jordan Painter, Diptesh Kanojia
widget:
  - text: wow, i mean who would have thought
base_model: vinai/bertweet-base
model-index:
  - name: bertweet-base-finetuned-SARC-combined-DS
    results: []

Utilising Weak Supervision to Create S3D: A Sarcasm Annotated Dataset

This is the repository for the S3D dataset published at EMNLP 2022. The dataset can help build sarcasm detection models.

bertweet-base-finetuned-SARC-combined-DS

This model is a fine-tuned version of vinai/bertweet-base on our combined sarcasm dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4624
  • Accuracy: 0.7611
  • Precision: 0.7611
  • Recall: 0.7611
  • F1: 0.7611

Model description

The given description for BERTweet by VinAI is as follows:
BERTweet is the first public large-scale language model pre-trained for English Tweets. BERTweet is trained based on the RoBERTa pre-training procedure. The corpus used to pre-train BERTweet consists of 850M English Tweets (16B word tokens ~ 80GB), containing 845M Tweets streamed from 01/2012 to 08/2019 and 5M Tweets related to the COVID-19 pandemic.

Training and evaluation data

More information neededThis vinai/bertweet-base model was finetuned on our combined sarcasm dataset. This dataset was created to aid the building of sarcasm detection models

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 64
  • eval_batch_size: 16
  • seed: 43
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 30

Training results

Training Loss Epoch Step Validation Loss Accuracy Precision Recall F1
0.4319 4.0 44819 0.5049 0.7790 0.7796 0.7789 0.7789
0.2835 8.0 89638 0.6475 0.7663 0.7664 0.7663 0.7663
0.1797 12.0 134457 0.8746 0.7638 0.7639 0.7637 0.7637
0.1219 16.0 179276 1.0595 0.7585 0.7597 0.7587 0.7583
0.0905 20.0 224095 1.2115 0.7611 0.7612 0.7612 0.7611
0.0728 24.0 268914 1.3644 0.7628 0.7629 0.7627 0.7627
0.0612 28.0 313733 1.4624 0.7611 0.7611 0.7611 0.7611

Framework versions

  • Transformers 4.20.1
  • Pytorch 1.10.1+cu111
  • Datasets 2.3.2
  • Tokenizers 0.12.1