Longer context length
#26 opened about 1 hour ago
by
comorado
Qwen 32B Compatibility on PC w/ Ryzen 7 Pro 8840HS w/ 780M Graphics 2x32GB RAM 1TB DDR5 SSD
#25 opened about 21 hours ago
by
arzx
请问我在用llama-factory微调distill-qwen系列模型时,模型名称选哪个?
#24 opened 2 days ago
by
wangda1
Update README.md
#23 opened 2 days ago
by
Rizki-firman
Update README.md
#22 opened 4 days ago
by
payam8499
Tokenizer config's `chat_template` removes everything before `</think>` XML closing tag
#21 opened 5 days ago
by
jamesbraza
Consistency, can Deepseek pass? 一致性,deepseek能及格吗?
#20 opened 7 days ago
by
zwpython
running on local machine
6
#19 opened 7 days ago
by
saidavanam
Poor performance in the leaderboard?
7
#17 opened 11 days ago
by
L29Ah
Add text-generation pipeline tag
#16 opened 12 days ago
by
nielsr
comfyui-deepseek-r1
3
#15 opened 13 days ago
by
zwpython
sharing something maybe beneficial ?
#13 opened 13 days ago
by
9x25dillon
Please convert these models to GGUF format...
5
#12 opened 14 days ago
by
Moodym
Support For Japanese Model
5
#11 opened 14 days ago
by
alfredplpl
Tokenizer config is wrong
8
#10 opened 14 days ago
by
stoshniwal
Garbage characters generated with using 32B
3
#9 opened 15 days ago
by
carlosbdw
Please add a qwen2.5-72b distill
#8 opened 15 days ago
by
warlock-edward
Does this have tooling support?
4
#7 opened 15 days ago
by
xceptor
What temp are these expected to be used at?
1
#6 opened 15 days ago
by
rombodawg
YaRN block required?
3
#5 opened 15 days ago
by
robbiemu
Please add a qwen coder 32b distill.
#4 opened 15 days ago
by
ciprianv
System Prompt
16
#2 opened 15 days ago
by
Wanfq