GGUF
Inference Endpoints
conversational
siacus's picture
Create README.md
3352d85 verified
---
license: mit
datasets:
- siacus/cap_pe_verified
base_model:
- meta-llama/Llama-2-7b-chat-hf
new_version: siacus/llama-2-7b-cap_verified
---
The data used to train the model are on Huggingface under [siacus/cap_pe_verified](https://huggingface.co./datasets/siacus/cap_pe_verified)
F16 version from merged weights created with [llama.cpp](https://github.com/ggerganov/llama.cpp) on a
CUDA GPU and the 4bit quantized version created on a Mac M2 Ultra Metal architecture. If you want
to use the 4bit quantized version on CUDA,
please quantize it directly from the F16 version.
For more information about this model refer the [main repository](https://github.com/siacus/rethinking-scale) for the supplementary material of the manuscript [Rethinking Scale: The Efficacy of Fine-Tuned Open-Source LLMs in Large-Scale Reproducible Social Science Research](https://arxiv.org/abs/2411.00890).