YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co./docs/hub/model-cards#model-card-metadata)
Official AQLM quantization of meta-llama/Llama-2-13b-hf
.
For this quantization, we used 2 codebook of 8 bits.
Selected evaluation results for this and other models:
Model | Quantization | WikiText 2 PPL | Model size, Gb | |
---|---|---|---|---|
Llama-2-13b | - | 4.57 | 26.0 | |
2x8 | 5.63 | 3.8 |
- Downloads last month
- 14
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.