Update README.md
Browse files
README.md
CHANGED
@@ -5,10 +5,11 @@ tags:
|
|
5 |
- Mistral
|
6 |
- merge
|
7 |
- moe
|
|
|
8 |
---
|
9 |
|
10 |
<img src="https://huggingface.co/aigeek0x0/radiantloom-mixtral-8x7b-fusion/resolve/main/Radiantloom-Mixtral-8x7B-Fusion.png" alt="Radiantloom Mixtral 8X7B Fusion" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
|
11 |
|
12 |
## Radiantloom Mixtral 8X7B Fusion DPO
|
13 |
|
14 |
-
This model is a finetuned version of [Radiantloom Mixtral 8X7B Fusion](https://huggingface.co/Radiantloom/radiantloom-mixtral-8x7b-fusion). It was finetuned using Direct Preference Optimization (DPO).
|
|
|
5 |
- Mistral
|
6 |
- merge
|
7 |
- moe
|
8 |
+
license: apache-2.0
|
9 |
---
|
10 |
|
11 |
<img src="https://huggingface.co/aigeek0x0/radiantloom-mixtral-8x7b-fusion/resolve/main/Radiantloom-Mixtral-8x7B-Fusion.png" alt="Radiantloom Mixtral 8X7B Fusion" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
|
12 |
|
13 |
## Radiantloom Mixtral 8X7B Fusion DPO
|
14 |
|
15 |
+
This model is a finetuned version of [Radiantloom Mixtral 8X7B Fusion](https://huggingface.co/Radiantloom/radiantloom-mixtral-8x7b-fusion). It was finetuned using Direct Preference Optimization (DPO).
|