DarwinAnim8or commited on
Commit
5ef678b
1 Parent(s): af15558

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md CHANGED
@@ -1,3 +1,61 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - DarwinAnim8or/greentext
5
+ language:
6
+ - en
7
+ tags:
8
+ - fun
9
+ - greentext
10
+ widget:
11
+ - text: '>be me'
12
+ example_title: be me
13
+ - text: '>be zoo keeper'
14
+ co2_eq_emissions:
15
+ emissions: 10
16
+ source: https://mlco2.github.io/impact/#compute
17
+ training_type: fine-tuning
18
+ geographical_location: Oregon, USA
19
+ hardware_used: 1x T4, Google Colab
20
  ---
21
+
22
+ # Pythia-Greentext-1.4b
23
+ A finetuned version of [Pythia-1.4b](https://huggingface.co/gpt2-xl) on the 'greentext' dataset.
24
+ A demo is available [here](https://huggingface.co/spaces/DarwinAnim8or/Pythia-Greentext-Playground)
25
+ The demo playground is recommended over the inference box on the right.
26
+
27
+ This is an alternate take on my "GPT-Greentext" releases.
28
+
29
+ # Training Procedure
30
+ This was trained on the 'greentext' dataset, on Google Colab.
31
+ This model was trained for 1 epoch with learning rate 1e-2.
32
+ Notably this uses the "prompt" and "completion" style jsonl file, rather than the plain text file found in the greentext dataset.
33
+ This nets somewhat better, mostly more consistent results.
34
+
35
+ # Biases & Limitations
36
+ This likely contains the same biases and limitations as the original model that it is based on, and additionally heavy biases from the greentext dataset.
37
+ It should be noted that offensive or not PG-output is definitely possible and likely will happen.
38
+
39
+ # Intended Use
40
+ This model is meant for fun, nothing else.
41
+
42
+ # Noteworthy differences between this model and the others
43
+ This model tends to like no_repeat_ngram_size values of 1 or 2; whereas the other models in this series tend to prefer 3.
44
+
45
+ # Sample Use
46
+ ```python
47
+ #Import model:
48
+ from happytransformer import HappyGeneration
49
+ happy_gen = HappyGeneration("GPTNEO", "DarwinAnim8or/Pythia-Greentext-1.4b")
50
+
51
+ #Set generation settings:
52
+ from happytransformer import GENSettings
53
+ args_top_k = GENSettingsGENSettings(no_repeat_ngram_size=2, do_sample=True, top_k=80, temperature=0.1, max_length=150, early_stopping=False)
54
+
55
+ #Generate a response:
56
+ result = happy_gen.generate_text(""">be me
57
+ >""", args=args_top_k)
58
+
59
+ print(result)
60
+ print(result.text)
61
+ ```