
NousResearch/DeepHermes-3-Llama-3-8B-Preview-GGUF
Updated
β’
22.6k
β’
64
Now that the software I'm using updated the llamacpp version, I'm changingbgguf. I don't get what's meant with IQ4_NL does this include IQ4_XS? So IQ4_XS is also supposed to run performant on arm or just Q4_0?
On a side note, since I had good performance with Q4ks in the past I would wish that that would also benefit from these change.