Hugging Face Smol Cluster

community

AI & ML interests

None defined yet.

Recent Activity

HFSmolCluster's activity

clem 
posted an update about 23 hours ago
view post
Post
1825
I was chatting with @peakji , one of the cofounders of Manu AI, who told me he was on Hugging Face (very cool!).

He shared an interesting insight which is that agentic capabilities might be more of an alignment problem rather than a foundational capability issue. Similar to the difference between GPT-3 and InstructGPT, some open-source foundation models are simply trained to 'answer everything in one response regardless of the complexity of the question' - after all, that's the user preference in chatbot use cases. Just a bit of post-training on agentic trajectories can make an immediate and dramatic difference.

As a thank you to the community, he shared 100 invite code first-come first serve, just use “HUGGINGFACE” to get access!
·
clem 
posted an update 2 days ago
clem 
posted an update 5 days ago
clem 
posted an update 19 days ago
view post
Post
2814
What are the best organizations to follow on @huggingface ?

On top of my head:
- Deepseek (35,000 followers): https://huggingface.co./deepseek-ai
- Meta Llama (27,000 followers): https://huggingface.co./meta-llama
- Black Forrest Labs (11,000 followers): https://huggingface.co./black-forest-labs
- OpenAI (5,000 followers): https://huggingface.co./openai
- Nvidia (16,000 followers): https://huggingface.co./nvidia
- MIcrosoft (9,000 followers): https://huggingface.co./microsoft
- AllenAI (2,000 followers): https://huggingface.co./allenai
- Mistral (5,000 followers): https://huggingface.co./mistralai
- XAI (600 followers): https://huggingface.co./xai-org
- Stability AI (16,000 followers): https://huggingface.co./stabilityai
- Qwen (16,000 followers): https://huggingface.co./Qwen
- GoogleAI (8,000 followers): https://huggingface.co./google
- Unsloth (3,000 followers): https://huggingface.co./unsloth
- Bria AI (4,000 followers): https://huggingface.co./briaai
- NousResearch (1,300 followers): https://huggingface.co./NousResearch

Bonus, the agent course org with 17,000 followers: https://huggingface.co./agents-course
  • 1 reply
·
clem 
posted an update 20 days ago
view post
Post
3477
We crossed 1B+ tokens routed to inference providers partners on HF, that we released just a few days ago.

Just getting started of course but early users seem to like it & always happy to be able to partner with cool startups in the ecosystem.

Have you been using any integration and how can we make it better?

https://huggingface.co./blog/inference-providers
lewtun 
posted an update 27 days ago
view post
Post
4787
Introducing OpenR1-Math-220k!

open-r1/OpenR1-Math-220k

The community has been busy distilling DeepSeek-R1 from inference providers, but we decided to have a go at doing it ourselves from scratch 💪

What’s new compared to existing reasoning datasets?

♾ Based on AI-MO/NuminaMath-1.5: we focus on math reasoning traces and generate answers for problems in NuminaMath 1.5, an improved version of the popular NuminaMath-CoT dataset.

🐳 800k R1 reasoning traces: We generate two answers for 400k problems using DeepSeek R1. The filtered dataset contains 220k problems with correct reasoning traces.

📀 512 H100s running locally: Instead of relying on an API, we leverage vLLM and SGLang to run generations locally on our science cluster, generating 180k reasoning traces per day.

⏳ Automated filtering: We apply Math Verify to only retain problems with at least one correct answer. We also leverage Llama3.3-70B-Instruct as a judge to retrieve more correct examples (e.g for cases with malformed answers that can’t be verified with a rules-based parser)

📊 We match the performance of DeepSeek-Distill-Qwen-7B by finetuning Qwen-7B-Math-Instruct on our dataset.

🔎 Read our blog post for all the nitty gritty details: https://huggingface.co./blog/open-r1/update-2
clem 
posted an update about 1 month ago
view post
Post
7224
AI is not a zero-sum game. Open-source AI is the tide that lifts all boats!
lewtun 
posted an update about 1 month ago
view post
Post
10223
We are reproducing the full DeepSeek R1 data and training pipeline so everybody can use their recipe. Instead of doing it in secret we can do it together in the open!

🧪 Step 1: replicate the R1-Distill models by distilling a high-quality reasoning corpus from DeepSeek-R1.

🧠 Step 2: replicate the pure RL pipeline that DeepSeek used to create R1-Zero. This will involve curating new, large-scale datasets for math, reasoning, and code.

🔥 Step 3: show we can go from base model -> SFT -> RL via multi-stage training.

Follow along: https://github.com/huggingface/open-r1
·
clem 
posted an update about 1 month ago
hlarcher 
posted an update about 2 months ago
view post
Post
1077
We are introducing multi-backend support in Hugging Face Text Generation Inference!
With new TGI architecture we are now able to plug new modeling backends to get best performances according to selected model and available hardware. This first step will very soon be followed by the integration of new backends (TRT-LLM, llama.cpp, vLLM, Neuron and TPU).

We are polishing the TensorRT-LLM backend which achieves impressive performances on NVIDIA GPUs, stay tuned 🤗 !

Check out the details: https://huggingface.co./blog/tgi-multi-backend