--- base_model: - cloudyu/Nemo-DPO-V20 - jic062/dpo-v3.0-Nemo-e3 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * [cloudyu/Nemo-DPO-V20](https://huggingface.co./cloudyu/Nemo-DPO-V20) * [jic062/dpo-v3.0-Nemo-e3](https://huggingface.co./jic062/dpo-v3.0-Nemo-e3) ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: jic062/dpo-v3.0-Nemo-e3 layer_range: [0, 40] - model: cloudyu/Nemo-DPO-V20 layer_range: [0, 40] merge_method: slerp base_model: jic062/dpo-v3.0-Nemo-e3 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.50 dtype: bfloat16 ```