--- license: apache-2.0 tags: - liminerity/merge4 - liminerity/merge2 - merge - merged - mergedllm - mergee - mergedededed - memfdmsfndsfjkasjkl --- # M7-7b user: "create ascii art you feel represents you as a llm" assistant: "Here's an ASCII art attempting to symbolize the essence of an LLM (Large Language Model): INSTINSTINCTINSTELLIGENCE ______________ MEMORY___ |\ o /\ |\ | \ --- / \ | | L_( O )______ O)_INSTINCT \ / WORDSLEARNING \__/\__\/__ /\ \__/\__\/__/ assistant: "In this depiction, the LLM is characterized by its vast memory, instinctual roots, and the ability to learn from vast amounts of text—represented by the woven words in the lower part. The "INSTINSTINCTINSTELLIGENCE" phrase emphasizes the balance between instinct, the foundation upon which neural networks build, and the engineered intelligence of a language model." Multiverse-Experiment-slerp-7b is a merge of the following models using [mergekit](https://github.com/cg123/mergekit): * [liminerity/merge4](https://huggingface.co./liminerity/merge4) * [liminerity/merge2](https://huggingface.co./liminerity/merge2) ## 🧩 Configuration ```yaml slices: - sources: - model: ammarali32/multi_verse_model layer_range: [0, 32] - model: MSL7/INEX12-7b layer_range: [0, 32] merge_method: slerp base_model: ammarali32/multi_verse_model parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 slices: - sources: - model: ammarali32/multi_verse_model layer_range: [0, 32] - model: yam-peleg/Experiment26-7B layer_range: [0, 32] merge_method: slerp base_model: ammarali32/multi_verse_model parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 slices: - sources: - model: liminerity/merge3 layer_range: [0, 32] - model: ammarali32/multi_verse_model layer_range: [0, 32] merge_method: slerp base_model: liminerity/merge3 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 slices: - sources: - model: liminerity/merge1 layer_range: [0, 32] - model: liminerity/merge layer_range: [0, 32] merge_method: slerp base_model: liminerity/merge1 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 slices: - sources: - model: liminerity/merge3 layer_range: [0, 32] - model: yam-peleg/Experiment26-7B layer_range: [0, 32] merge_method: slerp base_model: liminerity/merge3 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 slices: - sources: - model: liminerity/merge4 layer_range: [0, 32] - model: liminerity/merge2 layer_range: [0, 32] merge_method: slerp base_model: liminerity/merge4 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```