the last ones
Browse files- .gitattributes +2 -0
- README.md +88 -0
- model_config.json +51 -0
- vikhr-7b-instruct_0.2.Q6_K.gguf +3 -0
- vikhr-7b-instruct_0.2.Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -45,3 +45,5 @@ vikhr-7b-instruct_0.2.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
45 |
vikhr-7b-instruct_0.2.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
vikhr-7b-instruct_0.2.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
vikhr-7b-instruct_0.2.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
45 |
vikhr-7b-instruct_0.2.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
vikhr-7b-instruct_0.2.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
vikhr-7b-instruct_0.2.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
vikhr-7b-instruct_0.2.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
49 |
+
vikhr-7b-instruct_0.2.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,88 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Quantized version of [Vikhr-7B-instruct_0.2](https://huggingface.co/Vikhrmodels/Vikhr-7B-instruct_0.2) using llama.cpp.
|
2 |
+
|
3 |
+
See [Files and Versions](https://huggingface.co/pirbis/Vikhr-7B-instruct_0.2-GGUF/tree/main) for quantized GGUF files.
|
4 |
+
|
5 |
+
model_config.json contains an example of settings to use with LMStudio.
|
6 |
+
|
7 |
+
---
|
8 |
+
language:
|
9 |
+
- ru
|
10 |
+
- en
|
11 |
+
datasets:
|
12 |
+
- zjkarina/Vikhr_instruct
|
13 |
+
---
|
14 |
+
```python
|
15 |
+
from peft import PeftModel, PeftConfig
|
16 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
17 |
+
import torch
|
18 |
+
|
19 |
+
MODEL_NAME = "Vikhrmodels/Vikhr-7B-instruct_0.2"
|
20 |
+
DEFAULT_MESSAGE_TEMPLATE = "<s>{role}\n{content}</s>\n"
|
21 |
+
DEFAULT_SYSTEM_PROMPT = "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им."
|
22 |
+
|
23 |
+
class Conversation:
|
24 |
+
def __init__(
|
25 |
+
self,
|
26 |
+
message_template=DEFAULT_MESSAGE_TEMPLATE,
|
27 |
+
system_prompt=DEFAULT_SYSTEM_PROMPT,
|
28 |
+
):
|
29 |
+
self.message_template = message_template
|
30 |
+
self.messages = [{
|
31 |
+
"role": "system",
|
32 |
+
"content": system_prompt
|
33 |
+
}]
|
34 |
+
|
35 |
+
def add_user_message(self, message):
|
36 |
+
self.messages.append({
|
37 |
+
"role": "user",
|
38 |
+
"content": message
|
39 |
+
})
|
40 |
+
|
41 |
+
def get_prompt(self, tokenizer):
|
42 |
+
final_text = ""
|
43 |
+
for message in self.messages:
|
44 |
+
message_text = self.message_template.format(**message)
|
45 |
+
final_text += message_text
|
46 |
+
final_text += 'bot'
|
47 |
+
return final_text.strip()
|
48 |
+
|
49 |
+
|
50 |
+
def generate(model, tokenizer, prompt, generation_config):
|
51 |
+
data = tokenizer(prompt, return_tensors="pt")
|
52 |
+
data = {k: v.to(model.device) for k, v in data.items()}
|
53 |
+
output_ids = model.generate(
|
54 |
+
**data,
|
55 |
+
generation_config=generation_config
|
56 |
+
)[0]
|
57 |
+
output_ids = output_ids[len(data["input_ids"][0]):]
|
58 |
+
output = tokenizer.decode(output_ids, skip_special_tokens=True)
|
59 |
+
return output.strip()
|
60 |
+
|
61 |
+
#config = PeftConfig.from_pretrained(MODEL_NAME)
|
62 |
+
model = AutoModelForCausalLM.from_pretrained(
|
63 |
+
config.base_model_name_or_path,
|
64 |
+
load_in_8bit=True,
|
65 |
+
torch_dtype=torch.float16,
|
66 |
+
device_map="auto"
|
67 |
+
)
|
68 |
+
#model = PeftModel.from_pretrained( model, MODEL_NAME, torch_dtype=torch.float16)
|
69 |
+
model.eval()
|
70 |
+
|
71 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
|
72 |
+
generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
|
73 |
+
print(generation_config)
|
74 |
+
|
75 |
+
inputs = ["Как тебя зовут?", "Кто такой Колмогоров?"]
|
76 |
+
|
77 |
+
for inp in inputs:
|
78 |
+
conversation = Conversation()
|
79 |
+
conversation.add_user_message(inp)
|
80 |
+
prompt = conversation.get_prompt(tokenizer)
|
81 |
+
|
82 |
+
output = generate(model, tokenizer, prompt, generation_config)
|
83 |
+
print(inp)
|
84 |
+
print(output)
|
85 |
+
print('\n')
|
86 |
+
```
|
87 |
+
|
88 |
+
[wandb](https://wandb.ai/karina_romanova/vikhr/runs/up2hw5eh?workspace=user-karina_romanova)
|
model_config.json
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"name": "Config for Chat ID 1708268236270",
|
3 |
+
"load_params": {
|
4 |
+
"n_ctx": 2048,
|
5 |
+
"n_batch": 512,
|
6 |
+
"rope_freq_base": 0,
|
7 |
+
"rope_freq_scale": 0,
|
8 |
+
"n_gpu_layers": 24,
|
9 |
+
"use_mlock": true,
|
10 |
+
"main_gpu": 0,
|
11 |
+
"tensor_split": [
|
12 |
+
0
|
13 |
+
],
|
14 |
+
"seed": -1,
|
15 |
+
"f16_kv": true,
|
16 |
+
"use_mmap": true,
|
17 |
+
"no_kv_offload": false,
|
18 |
+
"num_experts_used": 0
|
19 |
+
},
|
20 |
+
"inference_params": {
|
21 |
+
"n_threads": 12,
|
22 |
+
"n_predict": -1,
|
23 |
+
"top_k": 40,
|
24 |
+
"min_p": 0.05,
|
25 |
+
"top_p": 0.95,
|
26 |
+
"temp": 0.8,
|
27 |
+
"repeat_penalty": 1.1,
|
28 |
+
"input_prefix": "<s>user\n",
|
29 |
+
"input_suffix": "</s>\nbot\n",
|
30 |
+
"antiprompt": [
|
31 |
+
"</s>"
|
32 |
+
],
|
33 |
+
"pre_prompt": "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им.",
|
34 |
+
"pre_prompt_suffix": "</s>\n",
|
35 |
+
"pre_prompt_prefix": "<s>system\n",
|
36 |
+
"seed": -1,
|
37 |
+
"tfs_z": 1,
|
38 |
+
"typical_p": 1,
|
39 |
+
"repeat_last_n": 64,
|
40 |
+
"frequency_penalty": 0,
|
41 |
+
"presence_penalty": 0,
|
42 |
+
"n_keep": 0,
|
43 |
+
"logit_bias": {},
|
44 |
+
"mirostat": 0,
|
45 |
+
"mirostat_tau": 5,
|
46 |
+
"mirostat_eta": 0.1,
|
47 |
+
"memory_f16": true,
|
48 |
+
"multiline_input": false,
|
49 |
+
"penalize_nl": true
|
50 |
+
}
|
51 |
+
}
|
vikhr-7b-instruct_0.2.Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e14bb39b39422f89214ed12c29661d9f4dbe37fc3a043157f070f0cecf907b0
|
3 |
+
size 5719879552
|
vikhr-7b-instruct_0.2.Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7fba35dda2e1717f5e96d4768ad31d0d99845e94e8f5c2c48d227f0feea4b61f
|
3 |
+
size 7407835008
|