Update base_model in README.md from FPHam-L3-8B-Everything-COT to FPHam/L3-8B-Everything-COT
Browse files
README.md
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
---
|
2 |
-
base_model: FPHam
|
3 |
pipeline_tag: text-generation
|
4 |
quantized_by: featherless-ai-quants
|
5 |
---
|
6 |
|
7 |
-
# FPHam
|
8 |
|
9 |
![Featherless AI Quants](./featherless-quants.png)
|
10 |
|
|
|
1 |
---
|
2 |
+
base_model: FPHam/L3-8B-Everything-COT
|
3 |
pipeline_tag: text-generation
|
4 |
quantized_by: featherless-ai-quants
|
5 |
---
|
6 |
|
7 |
+
# FPHam/L3-8B-Everything-COT GGUF Quantizations 🚀
|
8 |
|
9 |
![Featherless AI Quants](./featherless-quants.png)
|
10 |
|