|
--- |
|
model-index: |
|
- name: Junrulu/Reproduced-tulu2-dpo-13b |
|
results: [] |
|
datasets: |
|
- HuggingFaceH4/ultrafeedback_binarized |
|
- Junrulu/Reproduced-tulu2-test-sets |
|
language: |
|
- en |
|
base_model: allenai/tulu-2-13b |
|
--- |
|
|
|
# Model Card for Reproduced Tulu2 DPO 13B |
|
|
|
This repository provides a reproduction version of Tulu2-DPO-13B finetuned upon [Tulu2-13B](https://huggingface.co./allenai/tulu-2-13b) and [Ultrafeedback](https://huggingface.co./datasets/HuggingFaceH4/ultrafeedback_binarized). Therefore, we obey all licenses mentioned in Tulu2's work. Check our codes for more details: https://github.com/LuJunru/LLM_Finetune/tree/DPO, which is built with [TRL](https://github.com/huggingface/trl/tree/main). |
|
|
|
## Performance |
|
|
|
| Model | Size | Alignment | MT-Bench (score) | AlpacaEval 2.0 (win rate %) | |
|
|-------------|-----|----|---------------|--------------| |
|
| **Tulu-v2-13b** 🐪 | **13B** | **SFT** | **5.79** | **2.61** | |
|
| **Tulu-v2-dpo-13b** 🐪 | **13B** | **DPO** | **6.06** | **6.96** | |
|
| **Reproduced-tulu2-dpo-13b** | **13B** | **DPO** | **6.27** | **6.71** | |
|
|
|
## Input Format |
|
|
|
The model is trained to use the following format (note the newlines): |
|
``` |
|
<|user|> |
|
Your message here! |
|
<|assistant|> |
|
``` |
|
|
|
For best results, format all inputs in this manner. **Make sure to include a newline after `<|assistant|>`, this can affect generation quality quite a bit.** Note: if fine-tuning with this chat template, ensure to evaluate and test with the chat template. Otherwise, fine-tining without the template if you choose to not use template during testing. Any mismatch of the chatting template between training and testing phases can obviously dampen the final performance. |
|
|
|
## Training hyperparameters |
|
|
|
The following hyperparameters were used during DPO training: |
|
- DPO beta: 0.1 |
|
- learning_rate: 1e-6 * sqrt(Num of Nodes) |
|
- total_train_batch_size: 128 * Num of Nodes |
|
- optimizer: AdamW with beta1 0.9, beta2 0.999 and epsilon 1e-8 |
|
- lr_scheduler_type: linear |
|
- lr_scheduler_warmup_ratio: 0.1 |
|
- Weight Decay: 0.0 |
|
- num_epochs: 3.0 |
|
- Specifically add above input format over training samples |