Update configuration_Llamoe.py
Browse files- configuration_Llamoe.py +7 -4
configuration_Llamoe.py
CHANGED
@@ -1,18 +1,20 @@
|
|
1 |
-
|
2 |
from transformers.configuration_utils import PretrainedConfig
|
3 |
from transformers.utils import logging
|
4 |
|
5 |
|
6 |
logger = logging.get_logger(__name__)
|
7 |
|
8 |
-
|
9 |
"damerajee/Llamoe-test": "https://huggingface.co/damerajee/Llamoe-test/resolve/main/config.json",
|
10 |
}
|
11 |
|
12 |
|
13 |
class LlamoeConfig(PretrainedConfig):
|
14 |
-
|
|
|
|
|
15 |
keys_to_ignore_at_inference = ["past_key_values"]
|
|
|
16 |
def __init__(
|
17 |
self,
|
18 |
vocab_size=32000,
|
@@ -47,8 +49,9 @@ class LlamoeConfig(PretrainedConfig):
|
|
47 |
self.num_hidden_layers = num_hidden_layers
|
48 |
self.num_attention_heads = num_attention_heads
|
49 |
self.head_dim = head_dim
|
50 |
-
self.num_key_value_heads = num_key_value_heads
|
51 |
self.hidden_act = hidden_act
|
|
|
|
|
52 |
self.initializer_range = initializer_range
|
53 |
self.rms_norm_eps = rms_norm_eps
|
54 |
self.use_cache = use_cache
|
|
|
|
|
1 |
from transformers.configuration_utils import PretrainedConfig
|
2 |
from transformers.utils import logging
|
3 |
|
4 |
|
5 |
logger = logging.get_logger(__name__)
|
6 |
|
7 |
+
GEMMOE_PRETRAINED_CONFIG_ARCHIVE_MAP = {
|
8 |
"damerajee/Llamoe-test": "https://huggingface.co/damerajee/Llamoe-test/resolve/main/config.json",
|
9 |
}
|
10 |
|
11 |
|
12 |
class LlamoeConfig(PretrainedConfig):
|
13 |
+
|
14 |
+
|
15 |
+
model_type = "llamoe"
|
16 |
keys_to_ignore_at_inference = ["past_key_values"]
|
17 |
+
|
18 |
def __init__(
|
19 |
self,
|
20 |
vocab_size=32000,
|
|
|
49 |
self.num_hidden_layers = num_hidden_layers
|
50 |
self.num_attention_heads = num_attention_heads
|
51 |
self.head_dim = head_dim
|
|
|
52 |
self.hidden_act = hidden_act
|
53 |
+
self.hidden_activation = hidden_activation
|
54 |
+
self.num_key_value_heads = num_key_value_heads
|
55 |
self.initializer_range = initializer_range
|
56 |
self.rms_norm_eps = rms_norm_eps
|
57 |
self.use_cache = use_cache
|