Dolphin 3.0 Collection Dolphin 3.0 is the next generation of the Dolphin series of instruct-tuned models. Designed to be the ultimate general purpose local model. • 7 items • Updated 8 days ago • 52
Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback Paper • 2406.09279 • Published Jun 13, 2024 • 2
SPaR: Self-Play with Tree-Search Refinement to Improve Instruction-Following in Large Language Models Paper • 2412.11605 • Published 28 days ago • 17
🇮🇹👓 LLaVA-NDiNO Collection HF Collection for the models of the paper "LLaVA-NDiNO: Empowering LLMs with Multimodality for the Italian Language" • 7 items • Updated Oct 20, 2024 • 3
Stronger Models are NOT Stronger Teachers for Instruction Tuning Paper • 2411.07133 • Published Nov 11, 2024 • 35
view article Article SauerkrautLM's Multi-Phase Spectrum Training: A Technical Deep Dive By DavidGF • Nov 9, 2024 • 9
view article Article 🇮🇹🇯🇵🇧🇷 Generating multilingual instruction datasets with Magpie 🐦⬛ By anakin87 • Oct 21, 2024 • 18
view article Article Model2Vec: Distill a Small Fast Model from any Sentence Transformer By Pringled • Oct 14, 2024 • 61
Non Verbis, Sed Rebus: Large Language Models are Weak Solvers of Italian Rebuses Paper • 2408.00584 • Published Aug 1, 2024 • 6
view article Article Selective fine-tuning of Language Models with Spectrum By anakin87 • Sep 3, 2024 • 30
🧩 Verbalized Rebus @ CLiC-it 2024 Collection Materials for the paper "Non Verbis, Sed Rebus: Large Language Models are Weak Solvers of Italian Rebuses" • 13 items • Updated Aug 5, 2024 • 3
view article Article 🔥 Argilla 2.0: the data-centric tool for AI makers 🤗 By dvilasuero • Jul 30, 2024 • 37
view article Article Mixedbread 🤝 deepset: Announcing our New German/English Embedding Model By shadeMe • Jul 19, 2024 • 15