Divyasreepat commited on
Commit
2c68b0f
1 Parent(s): 501e2e1

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: keras-hub
3
+ ---
4
+ This is a [`Bert` model](https://keras.io/api/keras_hub/models/bert) uploaded using the KerasHub library and can be used with JAX, TensorFlow, and PyTorch backends.
5
+ Model config:
6
+ * **name:** bert_backbone
7
+ * **trainable:** True
8
+ * **vocabulary_size:** 30522
9
+ * **num_layers:** 24
10
+ * **num_heads:** 16
11
+ * **hidden_dim:** 1024
12
+ * **intermediate_dim:** 4096
13
+ * **dropout:** 0.1
14
+ * **max_sequence_length:** 512
15
+ * **num_segments:** 2
16
+
17
+ This model card has been generated automatically and should be completed by the model author. See [Model Cards documentation](https://huggingface.co/docs/hub/model-cards) for more information.
assets/tokenizer/vocabulary.txt ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_nlp.src.models.bert.bert_backbone",
3
+ "class_name": "BertBackbone",
4
+ "config": {
5
+ "name": "bert_backbone",
6
+ "trainable": true,
7
+ "vocabulary_size": 30522,
8
+ "num_layers": 24,
9
+ "num_heads": 16,
10
+ "hidden_dim": 1024,
11
+ "intermediate_dim": 4096,
12
+ "dropout": 0.1,
13
+ "max_sequence_length": 512,
14
+ "num_segments": 2
15
+ },
16
+ "registered_name": "keras_nlp>BertBackbone",
17
+ "assets": [],
18
+ "weights": "model.weights.h5"
19
+ }
metadata.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "keras_version": "3.0.1",
3
+ "keras_nlp_version": "0.7.0",
4
+ "parameter_count": 335141888,
5
+ "date_saved": "2023-12-27@02:04:23"
6
+ }
model.weights.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6db467aae557bb71a9ff4ec6d65145247b3e6febd9b83ed4bccf610cb7c2b7f2
3
+ size 1341357928
tokenizer.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_nlp.src.models.bert.bert_tokenizer",
3
+ "class_name": "BertTokenizer",
4
+ "config": {
5
+ "name": "bert_tokenizer",
6
+ "trainable": true,
7
+ "dtype": "int32",
8
+ "vocabulary": null,
9
+ "sequence_length": null,
10
+ "lowercase": true,
11
+ "strip_accents": false,
12
+ "split": true,
13
+ "suffix_indicator": "##",
14
+ "oov_token": "[UNK]"
15
+ },
16
+ "registered_name": "keras_nlp>BertTokenizer",
17
+ "assets": [
18
+ "assets/tokenizer/vocabulary.txt"
19
+ ],
20
+ "weights": null
21
+ }