k4rth33k commited on
Commit
cb3409d
·
verified ·
1 Parent(s): 1c45008

Add fields to `config.json` needed to be supported by HF TEI server

Browse files

HuggingFace's [text-embeddings-inference](https://github.com/huggingface/text-embeddings-inference) server expects a certain [format](https://github.com/huggingface/text-embeddings-inference/blob/a059696a33f3b2cd28ce5e69d3195d5b03189d96/router/src/lib.rs#L326) for the `config.json`.
This PR adds the 2 fields missing from the expected config:
- `pad_token_id` (0 because of bert tokenizer)
- `max_position_embeddings` (768)
Any suggestions and corrections are welcome :)

Files changed (1) hide show
  1. config.json +7 -5
config.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
  "activation_function": "swiglu",
3
  "architectures": [
4
- "NomicBertModel"
5
  ],
6
- "attn_pdrop": 0.0,
7
  "auto_map": {
8
- "AutoConfig": "nomic-ai/nomic-embed-text-v1-unsupervised--configuration_hf_nomic_bert.NomicBertConfig",
9
- "AutoModel": "nomic-ai/nomic-embed-text-v1-unsupervised--modeling_hf_nomic_bert.NomicBertModel"
10
  },
11
  "bos_token_id": null,
12
  "causal": false,
@@ -17,6 +17,7 @@
17
  "fused_dropout_add_ln": true,
18
  "initializer_range": 0.02,
19
  "layer_norm_epsilon": 1e-12,
 
20
  "max_trained_positions": 2048,
21
  "mlp_fc1_bias": false,
22
  "mlp_fc2_bias": false,
@@ -26,6 +27,7 @@
26
  "n_inner": 3072,
27
  "n_layer": 12,
28
  "n_positions": 8192,
 
29
  "pad_vocab_size_multiple": 64,
30
  "parallel_block": false,
31
  "parallel_block_tied_norm": false,
@@ -34,7 +36,7 @@
34
  "reorder_and_upcast_attn": false,
35
  "resid_pdrop": 0.1,
36
  "rotary_emb_base": 1000,
37
- "rotary_emb_fraction": 1.0,
38
  "rotary_emb_interleaved": false,
39
  "rotary_emb_scale_base": null,
40
  "rotary_scaling_factor": null,
 
1
  {
2
  "activation_function": "swiglu",
3
  "architectures": [
4
+ "NomicBertModel"
5
  ],
6
+ "attn_pdrop": 0,
7
  "auto_map": {
8
+ "AutoConfig": "nomic-ai/nomic-embed-text-v1-unsupervised--configuration_hf_nomic_bert.NomicBertConfig",
9
+ "AutoModel": "nomic-ai/nomic-embed-text-v1-unsupervised--modeling_hf_nomic_bert.NomicBertModel"
10
  },
11
  "bos_token_id": null,
12
  "causal": false,
 
17
  "fused_dropout_add_ln": true,
18
  "initializer_range": 0.02,
19
  "layer_norm_epsilon": 1e-12,
20
+ "max_position_embeddings": 768,
21
  "max_trained_positions": 2048,
22
  "mlp_fc1_bias": false,
23
  "mlp_fc2_bias": false,
 
27
  "n_inner": 3072,
28
  "n_layer": 12,
29
  "n_positions": 8192,
30
+ "pad_token_id": 0,
31
  "pad_vocab_size_multiple": 64,
32
  "parallel_block": false,
33
  "parallel_block_tied_norm": false,
 
36
  "reorder_and_upcast_attn": false,
37
  "resid_pdrop": 0.1,
38
  "rotary_emb_base": 1000,
39
+ "rotary_emb_fraction": 1,
40
  "rotary_emb_interleaved": false,
41
  "rotary_emb_scale_base": null,
42
  "rotary_scaling_factor": null,