justinqbui
commited on
Commit
•
7ddc081
1
Parent(s):
23b3273
Update README.md
Browse files
README.md
CHANGED
@@ -22,10 +22,9 @@ Alternatively, to run locally
|
|
22 |
```
|
23 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
24 |
|
25 |
-
tokenizer = AutoTokenizer.from_pretrained("justinqbui/bertweet-
|
26 |
-
|
27 |
-
model = AutoModelForSequenceClassification.from_pretrained("justinqbui/bertweet-pretraining-covid-vaccine-tweets-finetuned")
|
28 |
|
|
|
29 |
```
|
30 |
|
31 |
## Model description
|
@@ -43,8 +42,19 @@ The tokenizer requires the emoji library to be installed.
|
|
43 |
|
44 |
The intended use of this model is to detect if the contents of a covid tweet is potentially false or misleading. This model is not an end all be all. It has many limitations. For example, if someone makes a post containing an image, but has attached a satirical image, this model would not be able to distinguish this. If a user links a website, the tokenizer allocates a special token for links, meaning the contents of the linked website is completely lost. If someone tweets a reply, this model can't look at the parent tweets, and will lack context.
|
45 |
|
46 |
-
This model's dataset relies on the crowd-sourcing annotations being accurate.
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
|
|
|
|
|
|
|
|
|
|
|
48 |
|
49 |
|
50 |
|
|
|
22 |
```
|
23 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
24 |
|
25 |
+
tokenizer = AutoTokenizer.from_pretrained("justinqbui/bertweet-covid-vaccine-tweets-finetuned")
|
|
|
|
|
26 |
|
27 |
+
model = AutoModelForSequenceClassification.from_pretrained("justinqbui/bertweet-covid-vaccine-tweets-finetuned")
|
28 |
```
|
29 |
|
30 |
## Model description
|
|
|
42 |
|
43 |
The intended use of this model is to detect if the contents of a covid tweet is potentially false or misleading. This model is not an end all be all. It has many limitations. For example, if someone makes a post containing an image, but has attached a satirical image, this model would not be able to distinguish this. If a user links a website, the tokenizer allocates a special token for links, meaning the contents of the linked website is completely lost. If someone tweets a reply, this model can't look at the parent tweets, and will lack context.
|
44 |
|
45 |
+
This model's dataset relies on the crowd-sourcing annotations being accurate. This data is only accurate of up until early December 2021. For example, it probably wouldn't do very ell with tweets regarded the new omicron variant.
|
46 |
+
|
47 |
+
Example true inputs:
|
48 |
+
```
|
49 |
+
Covid vaccines are safe and effective. -> 97% true
|
50 |
+
Vaccinations are safe and help prevent covid. -> 97% true
|
51 |
+
```
|
52 |
|
53 |
+
Example false inputs:
|
54 |
+
```
|
55 |
+
Covid vaccines will kill you. -> 97% false
|
56 |
+
covid vaccines make you infertile. -> 97% false
|
57 |
+
```
|
58 |
|
59 |
|
60 |
|