YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co./docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

greesychat-turbo - GGUF

Original model description:

base_model: unsloth/llama-3-8b-Instruct-bnb-4bit language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl - sft datasets: - OnlyCheeini/greesychat

GreesyAI

GreesyChat-Turbo AI Model

Overview

GreesyChat-Turbo is an advanced AI model designed for robust text generation using the LLaMA 3 architecture. This model excels in providing high-quality responses for general conversation, mathematical queries, and more. It’s perfect for powering chatbots, virtual assistants, and any application requiring intelligent dialogue capabilities.

Benchmark Results

Metric Value
Perplexity 22.5
Generation Speed 75 ms per token
Accuracy 70%
Response Time 200 ms
Metric GreesyChat-Turbo Mixtral-8x7b GPT-4
Code 79.2 75.6 83.6
MMLU 74.5 79.9 85.1
Gms8k 89.2 (5) 88.7 94.2

Contact

For support or inquiries, please contact: [email protected]

Downloads last month
34
GGUF
Model size
8.03B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .