File size: 2,940 Bytes
23d5fd6
 
 
 
 
 
 
 
 
 
eddd71e
 
58bebe5
c84e354
75c448d
58bebe5
75c448d
 
 
58bebe5
23d5fd6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cd04579
 
 
 
23d5fd6
c101ae7
ddd8a2e
 
23d5fd6
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
---
language:
- en
tags:
- roberta
- marketing mix
- multi-label
- classification
- microblog
- tweets

widget:
- text: "Best cushioning ever!!! 🤗🤗🤗 my zoom vomeros are the bomb🏃🏽‍♀️💨!!! @nike #run #training"
- text: "Why are they always sold-out of Apple's new airpods in their online shop 🤯😡?"
- text: "They’re closing the @Aldo at the Lehigh Vally Mall and KOP 😭"
- text: "@Sony’s XM3’s ain’t as sweet as my bro’s airpod pros but got a real steal 🤑 the other day #deal #headphonez" 
- text: "Nike needs to sponsor more e-sports atheletes with Air Jordans! #nike #esports"
- text: "Say what you want about @Abercrombie's 90s shirtless males ads, they made dang good woll sweaters back in the day. This is one of 3 I have from the late 90s."
- text: "To celebrate this New Year, @Nordstrom is DOUBLING all donations up to $25,000! 🎉 Your donation will help us answer 2X the calls, texts, and chats that come in, allow us to train 2X more volunteers!"
- text: "It's inspiring to see religious leaders speaking up for workers' rights and fair wages. Every voice matters in the #FightFor15! 💪🏽✊🏼 #Solidarity #WorkersRights"
---
# Model Card for: mmx_classifier_microblog_ENv02

Multi-label classifier that identifies which marketing mix variable(s) a microblog post pertains to.

## Model Details

You can use this classifier to determine which of the 4P's of marketing, also known as marketing mix variables, a microblog post (e.g., Tweet) pertains to:

1. Product
2. Place
3. Price
4. Promotion

### Model Description

This classifier is a fine-tuned checkpoint of [cardiffnlp/twitter-roberta-large-2022-154m] (https://huggingface.co./cardiffnlp/twitter-roberta-large-2022-154m). 
It was trained on 15K Tweets that mentioned at least one of 699 brands. The Tweets were cleaned and labeled using OpenAI's GPT4. 

Because this is a multi-label classification problem, we use binary cross-entropy (BCE) with logits loss for the fine-tuning. We basically combine a sigmoid layer with BCELoss in a single class.
To obtain the probabilities for each label (i.e., marketing mix variable), you need to "push" the predictions through a sigmoid function. This is already done in the accompanying python notebook.

IMPORTANT: At the time of writing this description, Huggingface's pipeline did not support multi-label classifiers.

### Citation
For attribution, please cite the following reference if you use this model:
```
Ringel, Daniel, Creating Synthetic Experts with Generative Artificial Intelligence (July 15, 2023). Available at SSRN: https://ssrn.com/abstract=4542949
```

### Working Paper
Download the working paper from SSRN: ["Creating Synthetic Experts with Generative AI"](https://papers.ssrn.com/abstract_id=4542949)

### Additional Ressources
[www.synthetic-experts.ai](http://www.synthetic-experts.ai)  
[GitHub Repository](https://github.com/dringel/Synthetic-Experts)