ejbejaranos
commited on
Commit
β’
d95e646
1
Parent(s):
daddbb5
Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,40 @@
|
|
11 |
- **Bitlinear Layers:** Reduces weights to values of 1, 0, and -1. β
|
12 |
- **Optimized for:** Fast inference and memory efficiency β‘
|
13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
## Requirements π¦
|
15 |
Make sure you have the following libraries installed:
|
16 |
|
|
|
11 |
- **Bitlinear Layers:** Reduces weights to values of 1, 0, and -1. β
|
12 |
- **Optimized for:** Fast inference and memory efficiency β‘
|
13 |
|
14 |
+
## Architecture
|
15 |
+
```bash
|
16 |
+
Model size: 1.604B parameters
|
17 |
+
2024-10-08 14:53:07 - INFO - π’ Number of parameters in the model after extracting weights: 1
|
18 |
+
2024-10-08 14:53:07 - INFO - π Reduced model structure:
|
19 |
+
LlamaForCausalLM(
|
20 |
+
(model): LlamaModel(
|
21 |
+
(embed_tokens): Embedding(128256, 4096)
|
22 |
+
(layers): ModuleList(
|
23 |
+
(0-5): 6 x LlamaDecoderLayer(
|
24 |
+
(self_attn): LlamaSdpaAttention(
|
25 |
+
(q_proj): BitLinear(in_features=4096, out_features=4096, bias=False)
|
26 |
+
(k_proj): BitLinear(in_features=4096, out_features=4096, bias=False)
|
27 |
+
(v_proj): BitLinear(in_features=4096, out_features=4096, bias=False)
|
28 |
+
(o_proj): BitLinear(in_features=4096, out_features=4096, bias=False)
|
29 |
+
(rotary_emb): LlamaRotaryEmbedding()
|
30 |
+
)
|
31 |
+
(mlp): LlamaMLP(
|
32 |
+
(gate_proj): BitLinear(in_features=4096, out_features=2048, bias=False)
|
33 |
+
(up_proj): BitLinear(in_features=4096, out_features=2048, bias=False)
|
34 |
+
(down_proj): BitLinear(in_features=2048, out_features=4096, bias=False)
|
35 |
+
(act_fn): SiLU()
|
36 |
+
)
|
37 |
+
(input_layernorm): Identity()
|
38 |
+
(post_attention_layernorm): LlamaRMSNorm((4096,), eps=1e-05)
|
39 |
+
)
|
40 |
+
)
|
41 |
+
(norm): LlamaRMSNorm((4096,), eps=1e-05)
|
42 |
+
(rotary_emb): LlamaRotaryEmbedding()
|
43 |
+
)
|
44 |
+
(lm_head): Linear(in_features=4096, out_features=128256, bias=False)
|
45 |
+
)
|
46 |
+
```
|
47 |
+
|
48 |
## Requirements π¦
|
49 |
Make sure you have the following libraries installed:
|
50 |
|