DeepSeek-R1-ReDistill Collection Re-distilled DeepSeek R1 models • 3 items • Updated about 2 hours ago • 4
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.1 Text Generation • Updated about 7 hours ago • 6 • 1
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.0 Text Generation • Updated about 7 hours ago • 98 • 35
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.1 Text Generation • Updated about 7 hours ago • 6 • 1
DeepSeek-R1-ReDistill Collection Re-distilled DeepSeek R1 models • 3 items • Updated about 2 hours ago • 4
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.0 Text Generation • Updated about 7 hours ago • 98 • 35
view post Post 2069 Releasing HQQ Llama-3.1-70b 4-bit quantized version! Check it out at mobiuslabsgmbh/Llama-3.1-70b-instruct_4bitgs64_hqq. Achieves 99% of the base model performance across various benchmarks! Details in the model card. 🔥 8 8 + Reply
view post Post 1790 Excited to announce the release of our high-quality Llama-3.1 8B 4-bit HQQ/calibrated quantized model! Achieving an impressive 99.3% relative performance to FP16, it also delivers the fastest inference speed for transformers. mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib 1 reply · 🔥 9 9 + Reply
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-attn-4bit-moe-3bit-metaoffload-HQQ Text Generation • Updated Feb 29, 2024 • 14 • 13
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-attn-4bit-moe-2bitgs8-metaoffload-HQQ Text Generation • Updated Feb 29, 2024 • 15 • 20
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-attn-4bit-moe-2bit-metaoffload-HQQ Text Generation • Updated Feb 29, 2024 • 21 • 15