Quantized candle weights for the CoEdIT model

Quantized weights of CoEdIT for inference with candle.

Usage

You can run the smaller models directly from the browser using this space.

Clone candle, and run the quantized-t5 example:

$ cargo run --example quantized-t5 --release  -- \
  --model-id "jbochi/candle-coedit-quantized" \
  --prompt "Make this text coherent: Their flight is weak. They run quickly through the tree canopy." \
  --temperature 0
...
 Although their flight is weak, they run quickly through the tree canopy.

By default, it will use CoEdIT-large with q6k quantization (770M params, 643 MB).

To use CoEdIT-xl (3B params, 2.34 GB), or any other provided model, specify the weight-file and config-file:

$ cargo run --example quantized-t5 --release  -- \
  --model-id "jbochi/candle-coedit-quantized" \
  --weight-file "model-xl.gguf" \
  --config-file "config-xl.json" \
  --prompt "Rewrite to make this easier to understand: Note that a storm surge is what forecasters consider a hurricane's most treacherous aspect." \
  --temperature 0
...
 Note that a storm surge is what forecasters consider a hurricane's most dangerous part.

Models available

These are all the available formats. Weight file is named {model}.gguf and the config file is config-{base_model}.json

Model Base model Quantization # Params Size
- small (unofficial) None 77M 308 MB
model-small small 6k 77M 78.2 MB
model-small-q4k small 4k 77M 59.6 MB
model-small-q4_0 small 4_0 77M 43.4 MB
- base (unofficial) None 248M 990 MB
model-base base 6k 248M 194M
model-base-q4k base 4k 248M 133M
model-base-q4_0 base 4_0 248M 133M
- large None 770M 3.13 GB
model large 6k 770M 643 MB
model-q4k large 4k 770M 441 MB
model-q4_0 large 4_0 770M 441 MB
- xl None 3B 11.4 GB
model-xl xl 6k 3B 2.34 GB
model-xl-q4k xl 4k 3B 1.6 GB
model-xl-q4_0 xl 4_0 3B 1.6 GB
- xxl None 11B 44.5 GB
model-xxl xxl 6k 11B 9.14 GB
model-xxl-q4k xxl 4k 11B 6.27 GB
model-xxl-q4_0 xxl 4_0 11B 6.27 GB

Model generation

The weights were quantized using candle:

cargo run --example tensor-tools --release -- quantize \
  --quantization q6k \
  /path/to/coedit-<version>/model.safetensors \
  --out-file model<version>.gguf
Downloads last month
503
Safetensors
Model size
783M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train jbochi/candle-coedit-quantized

Spaces using jbochi/candle-coedit-quantized 3