Where did the BF16 come from?
2
#10 opened about 3 hours ago
by
gshpychka
Inference speed
2
#9 opened about 21 hours ago
by
Iker
Running this model using vLLM Docker
#8 opened 1 day ago
by
moficodes
Got it running after downloading some RAM!
1
#7 opened 1 day ago
by
ubergarm
UD-IQ1_M models for distilled R1 versions?
2
#6 opened 1 day ago
by
SamPurkis
Llama.cpp server chat template
2
#4 opened 4 days ago
by
softwareweaver
Are the Q4 and Q5 models R1 or R1-Zero
18
#2 opened 9 days ago
by
gng2info
What is the VRAM requirement to run this ?
5
#1 opened 9 days ago
by
RageshAntony