varun-v-rao commited on
Commit
f953973
1 Parent(s): 2a7757d

Upload GPT2ForSequenceClassification

Browse files
Files changed (3) hide show
  1. README.md +23 -55
  2. pytorch_adapter.bin +1 -1
  3. pytorch_model_head.bin +1 -1
README.md CHANGED
@@ -1,74 +1,42 @@
1
  ---
2
- license: mit
3
- base_model: openai-community/gpt2-large
4
  tags:
5
- - generated_from_trainer
 
6
  datasets:
7
- - stanfordnlp/snli
8
- metrics:
9
- - accuracy
10
- model-index:
11
- - name: gpt2-large-bn-adapter-7.42M-snli-model1
12
- results:
13
- - task:
14
- name: Text Classification
15
- type: text-classification
16
- dataset:
17
- name: snli
18
- type: stanfordnlp/snli
19
- metrics:
20
- - name: Accuracy
21
- type: accuracy
22
- value: 0.9016460069091649
23
  ---
24
 
25
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
26
- should probably proofread and complete it, then remove this comment. -->
27
 
28
- # gpt2-large-bn-adapter-7.42M-snli-model1
29
 
30
- This model is a fine-tuned version of [openai-community/gpt2-large](https://huggingface.co/openai-community/gpt2-large) on the snli dataset.
31
- It achieves the following results on the evaluation set:
32
- - Loss: 0.2719
33
- - Accuracy: 0.9016
34
 
35
- ## Model description
36
 
37
- More information needed
38
 
39
- ## Intended uses & limitations
 
 
40
 
41
- More information needed
42
 
43
- ## Training and evaluation data
 
44
 
45
- More information needed
 
 
46
 
47
- ## Training procedure
48
 
49
- ### Training hyperparameters
50
 
51
- The following hyperparameters were used during training:
52
- - learning_rate: 2e-05
53
- - train_batch_size: 32
54
- - eval_batch_size: 32
55
- - seed: 2
56
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
57
- - lr_scheduler_type: linear
58
- - num_epochs: 3
59
 
60
- ### Training results
61
 
62
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
63
- |:-------------:|:-----:|:-----:|:---------------:|:--------:|
64
- | 0.3654 | 1.0 | 17168 | 0.3073 | 0.8859 |
65
- | 0.3239 | 2.0 | 34336 | 0.2859 | 0.8967 |
66
- | 0.3041 | 3.0 | 51504 | 0.2719 | 0.9016 |
67
 
68
-
69
- ### Framework versions
70
-
71
- - Transformers 4.35.2
72
- - Pytorch 2.1.1+cu121
73
- - Datasets 2.15.0
74
- - Tokenizers 0.15.0
 
1
  ---
 
 
2
  tags:
3
+ - adapter-transformers
4
+ - gpt2
5
  datasets:
6
+ - snli
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  ---
8
 
9
+ # Adapter `varun-v-rao/gpt2-large-bn-adapter-7.42M-snli-model1` for openai-community/gpt2-large
 
10
 
11
+ An [adapter](https://adapterhub.ml) for the `openai-community/gpt2-large` model that was trained on the [snli](https://huggingface.co/datasets/snli/) dataset.
12
 
13
+ This adapter was created for usage with the **[Adapters](https://github.com/Adapter-Hub/adapters)** library.
 
 
 
14
 
15
+ ## Usage
16
 
17
+ First, install `adapters`:
18
 
19
+ ```
20
+ pip install -U adapters
21
+ ```
22
 
23
+ Now, the adapter can be loaded and activated like this:
24
 
25
+ ```python
26
+ from adapters import AutoAdapterModel
27
 
28
+ model = AutoAdapterModel.from_pretrained("openai-community/gpt2-large")
29
+ adapter_name = model.load_adapter("varun-v-rao/gpt2-large-bn-adapter-7.42M-snli-model1", source="hf", set_active=True)
30
+ ```
31
 
32
+ ## Architecture & Training
33
 
34
+ <!-- Add some description here -->
35
 
36
+ ## Evaluation results
 
 
 
 
 
 
 
37
 
38
+ <!-- Add some description here -->
39
 
40
+ ## Citation
 
 
 
 
41
 
42
+ <!-- Add some description here -->
 
 
 
 
 
 
pytorch_adapter.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:207afe89af3015147aaa29ff00fa8ae3543d4d5745a2bc2f898b204419d27552
3
  size 29739378
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6e6574ab8d2b146d1cf515c3cfb9cff2deece941d7fbdee69caf47154925fef
3
  size 29739378
pytorch_model_head.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7041212c0430d1f3a1ab5c0560ade9d9d6c95d9d324cb16cd25304c680540cb4
3
  size 16659
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8cd8018d957c26c21f8e5cf1ed2270d73c4f415085f7917f56bff84f08d3de6
3
  size 16659