Rรฉgis Pierrard

regisss

AI & ML interests

None yet

Recent Activity

Articles

Organizations

Hugging Face's profile picture Habana AI's profile picture Hugging Face Optimum's profile picture group2's profile picture Hugging Face H4's profile picture Hugging Face OSS Metrics's profile picture HuggingFace Doc Builds's profile picture Blog-explorers's profile picture AI Energy Score Project's profile picture Social Post Explorers's profile picture Hugging Face Machine Learning Optimization's profile picture Optimum Internal Testing's profile picture SLLHF's profile picture Privacy Preserving AI Hackathon (Zama, Hugging Face, Entrepreneur First)'s profile picture

regisss's activity

posted an update 7 days ago
New activity in Habana/mamba 19 days ago

Upload 2 files

#2 opened 19 days ago by
zzhang37
New activity in Habana/mamba 23 days ago

Upload 2 files

2
#1 opened 23 days ago by
zzhang37
reacted to onekq's post with ๐Ÿ”ฅ 2 months ago
view post
Post
1848
I'm now working on finetuning of coding models. If you are GPU-hungry like me, you will find quantized models very helpful. But quantization for finetuning and inference are different and incompatible. So I made two collections here.

Inference (GGUF, via Ollama, CPU is enough)
onekq-ai/ollama-ready-coding-models-67118c3cfa1af2cf04a926d6

Finetuning (Bitsandbytes, QLora, GPU is needed)
onekq-ai/qlora-ready-coding-models-67118771ce001b8f4cf946b2

For quantization, the inference models are far more popular on HF than finetuning models. I use https://huggingface.co./QuantFactory to generate inference models (GGUF), and there are a few other choices.

But there hasn't been such a service for finetuning models. DIY isn't too hard though. I made a few myself and you can find the script in the model cards. If the original model is small enough, you can even do it on a free T4 (available via Google Colab).

If you know a (small) coding model worthy of quantization, please let me know and I'd love to add it to the collections.
posted an update 2 months ago
view post
Post
1378
Interested in performing inference with an ONNX model?โšก๏ธ

The Optimum docs about model inference with ONNX Runtime is now much clearer and simpler!

You want to deploy your favorite model on the hub but you don't know how to export it to the ONNX format? You can do it in one line of code as follows:
from optimum.onnxruntime import ORTModelForSequenceClassification

# Load the model from the hub and export it to the ONNX format
model_id = "distilbert-base-uncased-finetuned-sst-2-english"
model = ORTModelForSequenceClassification.from_pretrained(model_id, export=True)

Check out the whole guide ๐Ÿ‘‰ https://huggingface.co./docs/optimum/onnxruntime/usage_guides/models
upvoted an article 2 months ago
view article
Article

Organizing a Privacy-preserving Hackathon

By binoua โ€ข
โ€ข 8
New activity in optimum/llm-perf-leaderboard 3 months ago
New activity in hf-doc-build/doc-build 3 months ago