mtyrrell commited on
Commit
c0df130
1 Parent(s): 194039b

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -48
README.md CHANGED
@@ -8,14 +8,6 @@ metrics:
8
  model-index:
9
  - name: IKT_classifier_transport_ghg_best
10
  results: []
11
-
12
- widget:
13
- - text: "Unconditional Contribution In the unconditional scenario, GHG emissions would be reduced by 27.56 Mt CO2e (6.73%) below BAU in 2030 in the respective sectors. 26.3 Mt CO2e (95.4%) of this emission reduction will be from the Energy sector while 0.64 (2.3%) and 0.6 (2.2%) Mt CO2e reduction will be from AFOLU (agriculture) and waste sector respectively. There will be no reduction in the IPPU sector. Conditional Contribution In the conditional scenario, GHG emissions would be reduced by 61.9 Mt CO2e (15.12%) below BAU in 2030 in the respective sectors."
14
- example_title: GHG
15
- - text: "Key Long-Term Climate Actions Cleaner and greener vehicles on our roads Singapore is working to enhance the overall carbon efficiency of our land transport system through the large-scale adoption of green vehicles. By 2040, we aim to phase out internal combustion engine vehicles and have all vehicles running on cleaner energy. We will introduce policies and initiatives to encourage the adoption of EVs. The public sector itself will take the lead and progressively procure and use cleaner vehicles."
16
- example_title: NOT_GHG
17
- - text: "This includes installation of rooftop PV panels for electricity generation, 5,300 solar water heaters, and expand the use of LED lighting in residential sector by 2030. • Expanding on energy efficiency labels and specifications for appliances programme, elimination of non-energy efficient equipment, and raising awareness among consumers on purchasing alternative energy efficient home appliances."
18
- example_title: NEGATIVE
19
  ---
20
 
21
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -23,54 +15,30 @@ should probably proofread and complete it, then remove this comment. -->
23
 
24
  # IKT_classifier_transport_ghg_best
25
 
26
- This model is a fine-tuned version of [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) on the [GIZ/policy_qa_v0_1](https://huggingface.co/datasets/GIZ/policy_qa_v0_1) dataset.
27
  It achieves the following results on the evaluation set:
28
- - Loss: 0.4763
29
- - Precision Macro: 0.8974
30
- - Precision Weighted: 0.8696
31
- - Recall Macro: 0.8974
32
- - Recall Weighted: 0.8696
33
- - F1-score: 0.8974
34
- - Accuracy: 0.8696
35
 
36
  ## Model description
37
 
38
- The model is a multi-class text classifier based on [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) and fine-tuned on text sourced from national climate policy documents.
39
 
40
  ## Intended uses & limitations
41
 
42
- The classifier assigns a class of 'GHG','NOT_GHG', or 'NEGATIVE' to denote alignment with **GHG-related transport targets** in extracted passages from the documents. The 'NEGATIVE' class in this case relates to negative samples not aligning with targets.
43
-
44
- The intended use is for climate policy researchers and analysts seeking to automate the process of reviewing lengthy, non-standardized PDF documents to produce summaries and reports.
45
-
46
- The performance of the classifier is middle of the road. On training, the classifier exhibited very good overall performance (F1 ~ 0.9). This performance was evenly balanced between precise identification of true positive classifications (precision ~ 0.9) and a wide net to capture as many true positives as possible (recall ~ 0.9). By contrast, when tested on real world unseen test data, the performance was mediocre (F1 ~ 0.5). However, testing was based on a small out-of-sample dataset. Therefore classification performance may differ in the wild.
47
-
48
 
49
  ## Training and evaluation data
50
 
51
- The training dataset is comprised of labelled passages from 2 sources:
52
- - [ClimateWatch NDC Sector data](https://www.climatewatchdata.org/data-explorer/historical-emissions?historical-emissions-data-sources=climate-watch&historical-emissions-gases=all-ghg&historical-emissions-regions=All%20Selected&historical-emissions-sectors=total-including-lucf%2Ctotal-including-lucf&page=1).
53
- - [IKI TraCS Climate Strategies for Transport Tracker](https://changing-transport.org/wp-content/uploads/20220722_Tracker_Database.xlsx) implemented by GIZ and funded by theInternational Climate Initiative (IKI) of the German Federal Ministry for Economic Affairs and Climate Action (BMWK). Here we utilized the QA dataset (CW_NDC_data_Sector).
54
-
55
- The combined dataset[GIZ/policy_qa_v0_1](https://huggingface.co/datasets/GIZ/policy_qa_v0_1) contains ~85k rows. Each row is duplicated twice, to provide varying sequence lengths (denoted by the values 'small', 'medium', and 'large', which correspond to sequence lengths of 60, 85, and 150 respectively - indicated in the 'strategy' column). This effectively means the dataset is reduced by 1/3 in useful size, and the 'strategy' value should be selected based on the use case. For this training, we utilized the 'medium' samples Furthermore, for each row, the 'context' column contains 3 samples of varying quality. The approach used to assess quality and select samples is described below.
56
-
57
- The pre-processing operations used to produce the final training dataset were as follows:
58
-
59
- 1. Dataset is filtered based on 'medium' value in 'strategy' column (sequence length = 85).
60
- 2. For ClimateWatch, all rows are removed as there was assessed to be no taxonomical alignment with the IKITracs labels inherent to the dataset. For IKITracs, labels are assigned based on the presence of certain substrings based on 'parameter' values which correspond to assessments of Net-Zero targets by human annotaters. The specific assignments are as follows:
61
- > - 'GHG': target_labels_ghg_yes = ['T_Transport_Unc','T_Transport_C']
62
- > - 'NOT_GHG': target_labels_ghg_no = ['T_Adaptation_Unc', 'T_Adaptation_C', 'T_Transport_O_Unc', 'T_Transport_O_C']
63
- > - 'NEGATIVE': random sample of other labeled data omitting above labels
64
- 3. If 'context_translated' is available and the 'language' is not English, 'context' is replaced with 'context_translated'.
65
- 4. The dataset is "exploded" - i.e., the text samples in the 'context' column, which are lists, are converted into separate rows - and labels are merged to align with the associated samples.
66
- 5. The 'match_onanswer' and 'answerWordcount' are used conditionally to select high quality samples (prefers high % of word matches in 'match_onanswer', but will take lower if there is a high 'answerWordcount')
67
- 6. Data is then augmented using sentence shuffle from the ```albumentations``` library and NLP-based insertions using ```nlpaug```.
68
-
69
 
70
  ## Training procedure
71
 
72
- The model hyperparameters were tuned using ```optuna``` over 10 trials on a truncated training and validation dataset. The model was then trained over 5 epochs using the best hyperparameters identified.
73
-
74
  ### Training hyperparameters
75
 
76
  The following hyperparameters were used during training:
@@ -87,11 +55,11 @@ The following hyperparameters were used during training:
87
 
88
  | Training Loss | Epoch | Step | Validation Loss | Precision Macro | Precision Weighted | Recall Macro | Recall Weighted | F1-score | Accuracy |
89
  |:-------------:|:-----:|:----:|:---------------:|:---------------:|:------------------:|:------------:|:---------------:|:--------:|:--------:|
90
- | No log | 1.0 | 52 | 0.2945 | 0.9149 | 0.8923 | 0.9149 | 0.8913 | 0.9145 | 0.8913 |
91
- | No log | 2.0 | 104 | 0.5662 | 0.8361 | 0.7928 | 0.8307 | 0.7826 | 0.8280 | 0.7826 |
92
- | No log | 3.0 | 156 | 0.5846 | 0.8201 | 0.8059 | 0.8474 | 0.8043 | 0.8300 | 0.8043 |
93
- | No log | 4.0 | 208 | 0.4326 | 0.9149 | 0.8923 | 0.9149 | 0.8913 | 0.9145 | 0.8913 |
94
- | No log | 5.0 | 260 | 0.4763 | 0.8974 | 0.8696 | 0.8974 | 0.8696 | 0.8974 | 0.8696 |
95
 
96
 
97
  ### Framework versions
 
8
  model-index:
9
  - name: IKT_classifier_transport_ghg_best
10
  results: []
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
15
 
16
  # IKT_classifier_transport_ghg_best
17
 
18
+ This model is a fine-tuned version of [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.4963
21
+ - Precision Macro: 0.9175
22
+ - Precision Weighted: 0.8942
23
+ - Recall Macro: 0.9156
24
+ - Recall Weighted: 0.8936
25
+ - F1-score: 0.9162
26
+ - Accuracy: 0.8936
27
 
28
  ## Model description
29
 
30
+ More information needed
31
 
32
  ## Intended uses & limitations
33
 
34
+ More information needed
 
 
 
 
 
35
 
36
  ## Training and evaluation data
37
 
38
+ More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39
 
40
  ## Training procedure
41
 
 
 
42
  ### Training hyperparameters
43
 
44
  The following hyperparameters were used during training:
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision Macro | Precision Weighted | Recall Macro | Recall Weighted | F1-score | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------------:|:------------------:|:------------:|:---------------:|:--------:|:--------:|
58
+ | No log | 1.0 | 53 | 0.3979 | 0.8806 | 0.8800 | 0.8964 | 0.8723 | 0.8819 | 0.8723 |
59
+ | No log | 2.0 | 106 | 0.7787 | 0.8428 | 0.8005 | 0.7377 | 0.7872 | 0.7695 | 0.7872 |
60
+ | No log | 3.0 | 159 | 0.4507 | 0.9028 | 0.8747 | 0.8981 | 0.8723 | 0.8990 | 0.8723 |
61
+ | No log | 4.0 | 212 | 0.7270 | 0.9019 | 0.8752 | 0.8680 | 0.8723 | 0.8830 | 0.8723 |
62
+ | No log | 5.0 | 265 | 0.4963 | 0.9175 | 0.8942 | 0.9156 | 0.8936 | 0.9162 | 0.8936 |
63
 
64
 
65
  ### Framework versions