HannaAbiAkl commited on
Commit
96c3044
1 Parent(s): dd510aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +80 -3
README.md CHANGED
@@ -1,3 +1,80 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: flan-t5-small
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: shadow
8
+ results: []
9
+ language:
10
+ - en
11
+ library_name: transformers
12
+ pipeline_tag: question-answering
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # SHADOW ![alt text](https://github.com/HannaAbiAkl/SHADOW/blob/main/figures/shadow.jpg)
19
+
20
+ SHADOW (**S**ymbolic **H**igher-order **A**ssociative **D**eductive reasoning **O**n **W**ikidata) is a model that is a fine-tuned version of [flan-t5-small](https://huggingface.co/google/flan-t5-small) on the LM-KBC 2024 dataset. It achieves the following results on the validation set:
21
+ - Loss: 0.0003
22
+
23
+ ## Model description
24
+
25
+ The model is trained to learn specific Wikidata patterns (e.g. type of relation, type of subject entity, etc.) to identify the correct template id that matches the SPARQL query that finds the appropriate object entities for the triple (subject entity, relation, object entity). The SPARQL queries are dynamically pre-defined for each template id and the model is not exposed to the queries themselves. The training of the model follows the associative reasoning pattern based on higher-order reasoning that deductively tries to associate a symbol (here, a template id) to data (here, the (subject entity, relation) tuple).
26
+
27
+ ## Intended uses & limitations
28
+
29
+ This model is intended to be used with a (subject entity, relation) Wikidata pair to determine the best template id for the SPARQL query that fetches the matching object entities.
30
+
31
+ ## Training and evaluation data
32
+
33
+ The LM-KBC dataset is used for training and evaluation.
34
+
35
+ ## Templates
36
+
37
+ The templates are frames that define dynamic SPARQL queries to find the object entities to complete the triple (subject entity, relation, object entities). The model is currently trained on the following relations:
38
+
39
+ - **countryLandBordersCountry:** Null values possible (e.g., Iceland)
40
+ - **personHasCityOfDeath:** Null values possible
41
+ - **seriesHasNumberOfEpisodes:** Object is numeric
42
+ - **awardWonBy:** Many objects per subject (e.g., 224 Physics Nobel prize winners)
43
+ - **companyTradesAtStockExchange:** Null values possible
44
+
45
+ ## Training procedure
46
+
47
+ ### Training hyperparameters
48
+
49
+ The following hyperparameters were used during training:
50
+ - learning_rate: 1e-04
51
+ - train_batch_size: 4
52
+ - eval_batch_size: 4
53
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
+ - lr_scheduler_type: linear
55
+ - num_epochs: 20
56
+
57
+ ### Training results
58
+
59
+ | Training Loss | Epoch | Step | Validation Loss |
60
+ |:-------------:|:-----:|:----:|:---------------:|
61
+ | 0.6679 | 1.0 | 1000 | 0.0005 |
62
+ | 0.3425 | 2.0 | 2000 | 0.0002 |
63
+ | 0.2303 | 3.0 | 3000 | 0.0001 |
64
+ | 0.1735 | 4.0 | 4000 | 0.000 |
65
+ | 0.1394 | 5.0 | 5000 | 0.0001 |
66
+ | 0.1167 | 6.0 | 6000 | 0.00009 |
67
+ | 0.1006 | 7.0 | 7000 | 0.00008 |
68
+ | 0.0882 | 8.0 | 8000 | 0.00007 |
69
+ | 0.0785 | 9.0 | 9000 | 0.00006 |
70
+ | 0.0707 | 10.0 | 10000| 0.00006 |
71
+ | 0.0643 | 11.0 | 11000| 0.00005 |
72
+ | 0.0590 | 12.0 | 12000| 0.00005 |
73
+ | 0.0545 | 13.0 | 13000| 0.00004 |
74
+ | 0.0506 | 14.0 | 14000| 0.00004 |
75
+ | 0.0473 | 15.0 | 15000| 0.00004 |
76
+ | 0.0443 | 16.0 | 16000| 0.00003 |
77
+ | 0.0417 | 17.0 | 17000| 0.00003 |
78
+ | 0.0394 | 18.0 | 18000| 0.00003 |
79
+ | 0.0374 | 19.0 | 19000| 0.00003 |
80
+ | 0.0355 | 20.0 | 20000| 0.00003 |