Description

This repo contains fp8 model files for aya-expanse-32b.

Quantization parameter

  • activation_scheme : dynamic
  • quant_method : fp8
Downloads last month
16
Safetensors
Model size
32.3B params
Tensor type
FP16
·
F8_E4M3
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for minyichen/aya-expanse-32b-Dynamic-fp8

Quantized
(20)
this model