Update README.md
Browse files
README.md
CHANGED
@@ -20,7 +20,7 @@ An initial foray into the world of fine-tuning. The goal of this release was to
|
|
20 |
### Notes & Methodology
|
21 |
* [Excalibur-7b](https://huggingface.co/InferenceIllusionist/Excalibur-7b) fine-tuned with Direct Preference Optimization (DPO) using Intel/orca_dpo_pairs
|
22 |
* This is a quick experiment to determine the impact of DPO finetuning on the original base model
|
23 |
-
*
|
24 |
|
25 |
|
26 |
|
|
|
20 |
### Notes & Methodology
|
21 |
* [Excalibur-7b](https://huggingface.co/InferenceIllusionist/Excalibur-7b) fine-tuned with Direct Preference Optimization (DPO) using Intel/orca_dpo_pairs
|
22 |
* This is a quick experiment to determine the impact of DPO finetuning on the original base model
|
23 |
+
* Ran for a little over an hour on a single A100
|
24 |
|
25 |
|
26 |
|