view article Article 🐺🐦⬛ LLM Comparison/Test: 25 SOTA LLMs (including QwQ) through 59 MMLU-Pro CS benchmark runs By wolfram • 21 days ago • 71
view article Article Towards a Fully Arabic Retrieval-Augmented Generation (RAG) Pipeline: By Omartificial-Intelligence-Space • 26 days ago • 6
view article Article To what extent are we responsible for our content and how to create safer Spaces? By davidberenstein1957 • Aug 30 • 3
view article Article Let’s make a generation of amazing image generation models By burtenshaw • 30 days ago • 33
view article Article Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models By mikelabs • Nov 21 • 2
view article Article Introducing Observers: AI Observability with Hugging Face datasets through a lightweight SDK By davidberenstein1957 • Nov 21 • 34
view article Article Releasing the largest multilingual open pretraining dataset By Pclanglais • Nov 13 • 98
Marqo-Ecommerce-Embeddings Collection State-of-the-art embedding models fine-tuned for the ecommerce domain. +67% increase in evaluation metrics vs ViT-B-16-SigLIP. • 10 items • Updated Nov 14 • 17
Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models Paper • 2411.04996 • Published Nov 7 • 49
view article Article ColPali: Efficient Document Retrieval with Vision Language Models 👀 By manu • Jul 5 • 182
view article Article Recipe: Preparing Multilingual Speech Datasets for TTS Training By PHBJT • Nov 4 • 14
AMD-OLMo Collection AMD-OLMo are a series of 1 billion parameter language models trained by AMD on AMD Instinct™ MI250 GPUs based on OLMo. • 4 items • Updated Oct 31 • 17