Triangle104 commited on
Commit
4b52974
·
verified ·
1 Parent(s): d408b96

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md CHANGED
@@ -26,6 +26,40 @@ tags:
26
  This model was converted to GGUF format from [`mistralai/Mistral-Small-24B-Instruct-2501`](https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501) for more details on the model.
28
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  ## Use with llama.cpp
30
  Install llama.cpp through brew (works on Mac and Linux)
31
 
 
26
  This model was converted to GGUF format from [`mistralai/Mistral-Small-24B-Instruct-2501`](https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501) for more details on the model.
28
 
29
+ ---
30
+ Model details:
31
+ -
32
+ Mistral Small 3 ( 2501 ) sets a new benchmark in the "small" Large Language Models category below 70B, boasting 24B parameters and achieving state-of-the-art capabilities comparable to larger models!
33
+ This model is an instruction-fine-tuned version of the base model: Mistral-Small-24B-Base-2501.
34
+
35
+ Mistral Small can be deployed locally and is exceptionally "knowledge-dense", fitting in a single RTX 4090 or a 32GB RAM MacBook once quantized.
36
+
37
+ Perfect for:
38
+
39
+ Fast response conversational agents.
40
+ Low latency function calling.
41
+ Subject matter experts via fine-tuning.
42
+ Local inference for hobbyists and organizations handling sensitive data.
43
+
44
+ For enterprises that need specialized capabilities (increased context, particular modalities, domain specific knowledge, etc.), we will be releasing commercial models beyond what Mistral AI contributes to the community.
45
+
46
+ This release demonstrates our commitment to open source, serving as a strong base model.
47
+
48
+ Learn more about Mistral Small in our blog post.
49
+
50
+ Model developper: Mistral AI Team
51
+
52
+ Key Features
53
+
54
+ Multilingual: Supports dozens of languages, including English, French, German, Spanish, Italian, Chinese, Japanese, Korean, Portuguese, Dutch, and Polish.
55
+ Agent-Centric: Offers best-in-class agentic capabilities with native function calling and JSON outputting.
56
+ Advanced Reasoning: State-of-the-art conversational and reasoning capabilities.
57
+ Apache 2.0 License: Open license allowing usage and modification for both commercial and non-commercial purposes.
58
+ Context Window: A 32k context window.
59
+ System Prompt: Maintains strong adherence and support for system prompts.
60
+ Tokenizer: Utilizes a Tekken tokenizer with a 131k vocabulary size.
61
+
62
+ ---
63
  ## Use with llama.cpp
64
  Install llama.cpp through brew (works on Mac and Linux)
65