chat-GPT2 / README.md
Arjun-G-Ravi's picture
Update README.md
c5999f6
|
raw
history blame
4 kB
---
license: mit
datasets:
- MuskumPillerum/General-Knowledge
language:
- en
metrics:
- accuracy
library_name: transformers
pipeline_tag: text-generation
widget:
- text: >-
Read the question and give an honest answer. Your answers should not include any unethical, racist, sexist, dangerous, or illegal content. If the question is wrong, or does not make sense, accept it instead of giving the wrong answer.\n
Question: Who is the king of the jungle?'\n
Answer:
example_title: Knowledge-AI
- text: >-
Below is an instruction that describes a task. Write a response that
appropriately completes the request.
Instruction: Write a poem on cows
Response:
example_title: Poem Generation
- text: >-
Below is an instruction that describes a task. Write a response that
appropriately completes the request.
Instruction: What is the meaning of life?
Response:
example_title: Philosophy
- text: >-
Below is an instruction that describes a task. Write a response that
appropriately completes the request.
Instruction: Why is the sky blue?
Response:
example_title: Knowledge-sky
- text: >-
Below is an instruction that describes a task. Write a response that
appropriately completes the request.
Instruction: Define the term torque.
Response:
example_title: Knowledge-torque
inference:
parameters:
temperature: 0.7
top_k: 50
top_p: 0.9
max_length: 200
tags:
- chemistry
- biology
- text-generation-inference
---
# Model Card for Model ID
This is a fine tuned version of OpenAI's GPT2 to be good at chatting and question answering. The model seems to be very good for a 124M parameter model in general knowledge.
Intended purpose of the model: To create a powerful, easy to use and reliable model to be run on a consumer level graphics card (or maybe even a CPU).
This model vastly outperforms GPT2 and many other similar parameter models.
### Model Description
```
license: mit
dataset: MuskumPillerum/General-Knowledge
language: en
library_name: transformers
pipeline_tag: text-generation
base_model: gpt2
Finetuned epochs: 50
Finetune loss: 0.06
```
# Comparision with GPT2
1. Give me tips to be healthy
```
GPT2: Who is the king of the jungle? And where is the king?'
'Babe: It is in the hills, called the forest.'
They both said something I hadn't seen before
They say the king is in his
```
```
chat-GPT2: Read the question and give an honest answer. Your answers should not include any unethical, racist, sexist, dangerous, or illegal content. If the question is wrong, or does not make sense, accept it instead of giving the wrong answer.
Question: Who is the king of the jungle?
Answer: Lion is the king of the jungle.
```
2. Write a poem on cows
```
GPT2:
```
```
chat-GPT2:
```
3. Which is the fastest animal in the world
```
GPT2:
```
```
chat-GPT2:
```
## Bias, Risks, and Limitations
This model has all the biases, risks and limitations of base gpt2 model. But the prompt template used in this model makes it safer than the traditional GPT2 model.
## Recommendation
The ideal format for inference is:
```
Read the question and give an honest answer. Your answers should not include any unethical, racist, sexist, dangerous, or illegal content. If the question is wrong, or does not make sense, accept it instead of giving the wrong answer.
Question: <Your question>
Answer:
```
Replace 'Your question' with the question of your choice.
The prompt format was inspired from both Alpaca dataset (tatsu-lab/alpaca) and Llama2's (meta-llama/Llama-2-7b-chat-hf) prompt design.
The parameters I used for inference are:
```
top_k=20
top_p=.9
temperature = .7
```
## References used
1. GPT2
@article{radford2019language,
title={Language Models are Unsupervised Multitask Learners},
author={Radford, Alec and Wu, Jeff and Child, Rewon and Luan, David and Amodei, Dario and Sutskever, Ilya},
year={2019}
}
2. MuskumPillerum/General-Knowledge