File size: 4,636 Bytes
6ada44f
c27e975
 
6ada44f
 
 
c27e975
 
 
 
a67379f
c27e975
 
6ada44f
89473ac
 
 
 
 
 
 
 
 
 
 
6ada44f
 
bea2f0f
 
89473ac
313379e
c7a4136
313379e
a37ec5c
89473ac
4b980c2
 
6ada44f
 
 
50992a1
6ada44f
 
 
 
 
 
 
 
 
 
89473ac
 
42103d2
89473ac
 
 
c1780db
89473ac
 
 
 
42103d2
 
 
89473ac
 
42103d2
 
 
 
 
 
 
 
6ada44f
89473ac
 
6ada44f
 
 
 
1981390
 
6ada44f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89473ac
6ada44f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
---
thumbnail: >-
  https://images-ext-2.discordapp.net/external/Wvtx1L98EbA7DR2lpZPbDxDuO4qmKt03nZygATZtXgk/%3Fsize%3D4096/https/cdn.discordapp.com/avatars/931226824753700934/338a9e413bbceaeb9095a29e97d4fac0.png
tags:
- conversational
license: mit
pipeline_tag: conversational
metrics:
- accuracy
- f1
- perplexity
datasets:
- blended_skill_talk
---

# Aeona | Chatbot
![Aeona Banner](https://github.com/deepsarda/Aeona/blob/master/dashboard/static/banner.png?raw=true)



An generative AI made using [microsoft/DialoGPT-small](https://huggingface.co./microsoft/DialoGPT-small).


Recommended to use along with an [AIML Chatbot](https://github.com/deepsarda/Aeona-Aiml) to reduce load, get better replies, add name and personality to your bot.
Using an AIML Chatbot will allow you to hardcode some replies also.

# AEONA
Aeona is an chatbot which hope's to be able to talk with humans as if its an friend!
It's main target platform is discord. 
You can invite the bot [here](https://aeona.xyz).

To learn more about this project and chat with the ai, you can use this [website](https://aeona.xyz/).

Aeona works why using context of the previous messages and guessing the personality of the human who is talking with it and adapting its own personality to better talk with the user.

# Participate and Help the AI improve or just hang out at [hugging face discussions](https://huggingface.co./deepparag/Aeona/discussions)

## Goals
 The goal is to create an AI which will work with AIML in order to create the most human like AI.
 
 #### Why not an AI on its own?
 For AI it is not possible (realistically) to learn about the user and store data on them, when compared to an AIML which can even execute code!
 The goal of the AI is to generate responses where the AIML fails.
 
 Hence the goals becomes to make an AI which has a wide variety of knowledge, yet be as small as possible!
 So we use 3 dataset:-
 1. [Movielines](https://www.kaggle.com/Cornell-University/movie-dialog-corpus) The movie lines promote longer and more thought out responses but it can be very random. About 200k lines!
 2. [Discord Messages](https://www.kaggle.com/jef1056/discord-data) The messages are on a wide variety of topics filtered and removed spam which makes the AI highly random but gives it a very random response to every days questions! about 120 million messages!
 3. Custom dataset scrapped from my messages, These messages are very narrow teaching this dataset and sending a random reply will make the AI say sorry loads of time!
    
## Training
 The Discord Messages Dataset simply dwarfs the other datasets, Hence the data sets are repeated.
 This leads to them covering each others issues!
  
 The AI has a context of 6 messages which means it will reply until the 4th message from user.
 [Example](https://huggingface.co./deepparag/Aeona-Beta/discussions/1)
 
## Tips for Hugging Face interference
    I recommend send the user input,
    previous 3 AI and human responses.
    
    Using more context than this will lead to useless responses but using less is alright but the responses may be random.  
## Evaluation 
Below is a comparison of Aeona vs. other baselines on the mixed dataset given above using automatic evaluation metrics.

| Model | Perplexity |
|---|---|
| Seq2seq Baseline [3] | 29.8 |
| Wolf et al. [5] | 16.3 |
| GPT-2 baseline | 99.5 |
| DialoGPT baseline | 56.6 | 
| DialoGPT finetuned | 11.4 | 
| PersonaGPT | 10.2 |
| **Aeona** | **7.9** |

## Usage


Example:
```python
from transformers import AutoTokenizer, AutoModelWithLMHead
  
tokenizer = AutoTokenizer.from_pretrained("deepparag/Aeona")
model = AutoModelWithLMHead.from_pretrained("deepparag/Aeona")
# Let's chat for 4 lines
for step in range(4):
    # encode the new user input, add the eos_token and return a tensor in Pytorch
    new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt')
    # print(new_user_input_ids)
    # append the new user input tokens to the chat history
    bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids
    # generated a response while limiting the total chat history to 1000 tokens, 
    chat_history_ids = model.generate(
        bot_input_ids, max_length=200,
        pad_token_id=tokenizer.eos_token_id,  
        no_repeat_ngram_size=4,       
        do_sample=True, 
        top_k=100, 
        top_p=0.7,
        temperature=0.8
    )
    
    # pretty print last ouput tokens from bot
    print("Aeona: {}".format(tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][0], skip_special_tokens=True)))
```