Hercules-phi-2

We fine-tuned phi2 on Locutusque's Hercules-v4.5.

Model Details

Model Description

This model has capabilities in math, coding, function calling, roleplay, and more. We fine-tuned it using all examples of Hercules-v4.5.

  • Developed by: M4-ai
  • Language(s) (NLP): English
  • License: apache-2.0

Uses

General purpose assistant, question answering, chain-of-thought, etc..

Evaluation

Coming soon

Training Details

Training Data

https://huggingface.co./datasets/Locutusque/hercules-v4.5

Training Hyperparameters

  • Training regime: bf16 non-mixed precision

Technical Specifications

Hardware

We used 8 Kaggle TPUs, and we trained at a global batch size of 1152.

Downloads last month
19
Safetensors
Model size
2.78B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train M4-ai/Hercules-phi-2