Made directly from https://huggingface.co./Qwen/Qwen1.5-14B-Chat I think official GGUF was made from already compressed AWQ. I converted original model to f32 first instead. Results are subjectively slightly better than official GGUF. But I didn't perform any perplexity test.
- Downloads last month
- 7
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model's library.