File size: 1,995 Bytes
f664d12
f38ce2d
0b5398e
5155e15
 
 
0b5398e
9c2c62a
5155e15
 
 
 
0b5398e
5155e15
 
9c2c62a
 
 
f0a43dd
9c2c62a
5155e15
9f84e26
0b5398e
9c2c62a
f664d12
0b5398e
9c2c62a
0b5398e
0241489
0b5398e
 
0241489
 
 
 
 
0b5398e
9c2c62a
 
0b5398e
9c2c62a
1fea721
0241489
0b5398e
 
 
 
 
 
 
 
 
0241489
9f84e26
 
0b5398e
 
 
 
 
 
 
1fea721
0b5398e
 
 
 
 
 
 
 
 
 
9c2c62a
1fea721
9c2c62a
0b5398e
 
 
 
 
 
9c2c62a
f664d12
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
{
  "_name_or_path": "ayjays132/CustomGPT2Conversational",
  "activation_function": "gelu_new",
  "architectures": [
    "GPT2LMHeadModel"
  ],
  "attn_pdrop": 0.1,
  "bos_token_id": 50256,
  "config": {
    "activation_function": "gelu_new",
    "attn_pdrop": 0.1,
    "embd_pdrop": 0.1,
    "gradient_checkpointing": true,
    "initializer_range": 0.02,
    "layer_norm_epsilon": 1e-05,
    "n_ctx": 2048,
    "n_embd": 2048,
    "n_head": 16,
    "n_layer": 36,
    "n_positions": 2048,
    "resid_pdrop": 0.1,
    "scale_attn_weights": true,
    "use_cache": true,
    "vocab_size": 50257
  },
  "embd_pdrop": 0.1,
  "eos_token_id": 50256,
  "initializer_range": 0.02,
  "language": "en",
  "layer_norm_epsilon": 1e-05,
  "library_name": "transformers",
  "license": "apache-2.0",
  "metrics": [
    "perplexity",
    "accuracy"
  ],
  "model_type": "gpt2",
  "n_embd": 768,
  "n_head": 12,
  "n_inner": null,
  "n_layer": 12,
  "n_positions": 1024,
  "pipeline_tag": "conversational",
  "reorder_and_upcast_attn": false,
  "resid_pdrop": 0.1,
  "scale_attn_by_inverse_layer_idx": false,
  "scale_attn_weights": true,
  "summary_activation": null,
  "summary_first_dropout": 0.1,
  "summary_proj_to_labels": true,
  "summary_type": "cls_index",
  "summary_use_proj": true,
  "tags": [
    "conversational",
    "state-of-the-art"
  ],
  "task_specific_params": {
    "conversational": {
      "do_sample": true,
      "early_stopping": true,
      "frequency_penalty": 0.5,
      "length_penalty": 2.0,
      "max_length": 1024,
      "min_length": 20,
      "no_repeat_ngram_size": 3,
      "num_beams": 5,
      "presence_penalty": 0.5,
      "temperature": 0.7,
      "top_k": 40,
      "top_p": 0.95
    }
  },
  "tokenizer_config": {
    "bos_token_id": 50256,
    "eos_token_id": 50256,
    "n_positions": 2048,
    "padding_side": "left",
    "truncation_side": "right"
  },
  "torch_dtype": "float32",
  "transformers_version": "4.37.2",
  "use_cache": true,
  "vocab_size": 50257
}