File size: 2,258 Bytes
c813a4a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
license: other
license_name: yi-license
license_link: https://huggingface.co./01-ai/Yi-34B-200K/blob/main/LICENSE
tags:
- yi
- moe
---
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/mLH2E0dk9On_LcFX9yhuS.jpeg)

# Bagel-Hermes-2x34B

This is the model for Bagel-Hermes-2x34B. I used [mergekit](https://github.com/cg123/mergekit) to make this MOE model.

# Prompt Template(s):

Since [bagel-dpo-34b-v0.2](https://huggingface.co./jondurbin/bagel-dpo-34b-v0.2) uses many prompt templates, and [Nous-Hermes-2-Yi-34B](https://huggingface.co./NousResearch/Nous-Hermes-2-Yi-34B) uses ChatML, you can utilize ChatML and other prompt templates provided by bagel.

**Note:** I currently do not know which prompt template is best.

### ChatML:

```
<|im_start|>system
{system}<|im_end|>
<|im_start|>user
{user}<|im_end|>
<|im_start|>assistant
{asistant}<|im_end|>
```

### Alpaca (sort of)

```
Below is an instruction that describes a task.  Write a response that appropriately completes the request.

### Instruction:
{system}
{instruction}

### Response:
```
### Vicuna

```
{system}
USER: {instruction}
ASSISTANT: 
```

Visit [bagel-dpo-34b-v0.2](https://huggingface.co./jondurbin/bagel-dpo-34b-v0.2) to try more prompt templates. 

# Yaml Config to reproduce

```yaml
base_model: nontoxic-bagel-34b-v0.2
gate_mode: hidden
dtype: bfloat16

experts:
  - source_model: bagel-dpo-34b-v0.2
    positive_prompts: ["question answering", "Q:", science", "biology", "chemistry", "physics"]

  - source_model: Nous-Hermes-2-Yi-34B
    positive_prompts: ["chat", "math", "reason", "mathematics", "solve", "count", "python", "javascript", "programming", "algorithm", "tell me", "assistant"]
```

# Quantizationed versions

Quantizationed versions of this model is available thanks to [TheBloke](https://hf.co/TheBloke).

##### GPTQ

- [TheBloke/Bagel-Hermes-2x34B-GPTQ](https://huggingface.co./TheBloke/Bagel-Hermes-2x34B-GPTQ)

##### GGUF

- [TheBloke/Bagel-Hermes-2x34B-GGUF](https://huggingface.co./TheBloke/Bagel-Hermes-2x34B-GGUF)

##### AWQ

- [TheBloke/Bagel-Hermes-2x34B-AWQ](https://huggingface.co./TheBloke/Bagel-Hermes-2x34B-AWQ)

If you would like to support me:

[☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)