|
--- |
|
license: mit |
|
language: |
|
- en |
|
base_model: |
|
- microsoft/phi-4 |
|
--- |
|
# Model Card for Microsoft-phi-4-Instruct-AutoRound-GPTQ-4bit |
|
|
|
## Model Overview |
|
|
|
**Model Name**: Microsoft-phi-4-Instruct-AutoRound-GPTQ-4bit |
|
**Model Type**: Instruction-tuned, Quantized GPT-4-based language model |
|
**Quantization**: GPTQ 4-bit |
|
**Author**: Satwik11 |
|
**Hosted on**: Hugging Face |
|
|
|
## Description |
|
|
|
This model is a quantized version of the Microsoft phi-4 Instruct model, designed to deliver high performance while maintaining computational efficiency. By leveraging the GPTQ 4-bit quantization method, it enables deployment in environments with limited resources while retaining a high degree of accuracy. |
|
|
|
The model is fine-tuned for instruction-following tasks, making it ideal for applications in conversational AI, question answering, and general-purpose text generation. |
|
|
|
## Key Features |
|
|
|
- **Instruction-tuned**: Fine-tuned to follow human-like instructions effectively. |
|
- **Quantized for Efficiency**: Uses GPTQ 4-bit quantization to reduce memory requirements and inference latency. |
|
- **Pre-trained Base**: Built on the Microsoft phi-4 framework, ensuring state-of-the-art performance on NLP tasks. |
|
|
|
## Use Cases |
|
|
|
- Chatbots and virtual assistants. |
|
- Summarization and content generation. |
|
- Research and educational applications. |
|
- Semantic search and knowledge retrieval. |
|
|
|
## Model Details |
|
|
|
### Architecture |
|
|
|
- **Base Model**: Microsoft phi-4 |
|
- **Quantization Technique**: GPTQ (4-bit) |
|
- **Language**: English |
|
- **Training Objective**: Instruction-following fine-tuning |