--- license: apache-2.0 tags: - merge - mergekit - mlabonne/AlphaMonarch-7B - bardsai/jaskier-7b-dpo-v5.6 --- # neurotic-crown-clown-7B-ties neurotic-crown-clown-7B-ties is a TRIM, ELECT SIGN & MERGE (TIES) merge of the following models using [mergekit](https://github.com/cg123/mergekit): * [mlabonne/AlphaMonarch-7B](https://huggingface.co./mlabonne/AlphaMonarch-7B) * [bardsai/jaskier-7b-dpo-v5.6](https://huggingface.co./bardsai/jaskier-7b-dpo-v5.6) See the paper [TIES-Merging: Resolving Interference When Merging Models](https://arxiv.org/abs/2306.01708) for more on the method. ## 🧩 Configuration ```yaml models: - model: mlabonne/NeuralMonarch-7B # no parameters necessary for base model - model: mlabonne/AlphaMonarch-7B parameters: density: 0.5 weight: 0.5 - model: bardsai/jaskier-7b-dpo-v5.6 parameters: density: 0.5 weight: 0.3 merge_method: ties base_model: mlabonne/NeuralMonarch-7B parameters: normalize: true dtype: float16 ```