Open Japanese LLM leaderboard
Collection
4 items
•
Updated
model_type
stringclasses 4
values | model
stringlengths 13
62
| AVG
float64 0.04
0.7
| CG
float64 0
0.68
| EL
float64 0
0.61
| FA
float64 0
0.35
| HE
float64 0
0.78
| MC
float64 0
0.92
| MR
float64 0
0.95
| MT
float64 0.3
0.86
| NLI
float64 0
0.82
| QA
float64 0.01
0.74
| RC
float64 0.04
0.93
| SUM
float64 0
0.28
| aio_char_f1
float64 0
0.89
| alt-e-to-j_bert_score_ja_f1
float64 0.26
0.88
| alt-e-to-j_bleu_ja
float64 0.32
16
| alt-e-to-j_comet_wmt22
float64 0.29
0.92
| alt-j-to-e_bert_score_en_f1
float64 0.37
0.96
| alt-j-to-e_bleu_en
float64 0.02
19.8
| alt-j-to-e_comet_wmt22
float64 0.3
0.89
| chabsa_set_f1
float64 0
0.61
| commonsensemoralja_exact_match
float64 0
0.94
| jamp_exact_match
float64 0
0.72
| janli_exact_match
float64 0
0.95
| jcommonsenseqa_exact_match
float64 0
0.96
| jemhopqa_char_f1
float64 0
0.68
| jmmlu_exact_match
float64 0
0.76
| jnli_exact_match
float64 0
0.9
| jsem_exact_match
float64 0
0.81
| jsick_exact_match
float64 0
0.87
| jsquad_char_f1
float64 0.04
0.93
| jsts_pearson
float64 -0.23
0.91
| jsts_spearman
float64 -0.19
0.88
| kuci_exact_match
float64 0
0.86
| mawps_exact_match
float64 0
0.95
| mbpp_code_exec
float64 0
0.68
| mbpp_pylint_check
float64 0
0.99
| mmlu_en_exact_match
float64 0
0.81
| niilc_char_f1
float64 0.01
0.68
| wiki_coreference_set_f1
float64 0
0.13
| wiki_dependency_set_f1
float64 0
0.54
| wiki_ner_set_f1
float64 0
0.17
| wiki_pas_set_f1
float64 0
0.11
| wiki_reading_char_f1
float64 0.02
0.91
| wikicorpus-e-to-j_bert_score_ja_f1
float64 0.15
0.87
| wikicorpus-e-to-j_bleu_ja
float64 0.17
18.2
| wikicorpus-e-to-j_comet_wmt22
float64 0.3
0.87
| wikicorpus-j-to-e_bert_score_en_f1
float64 0.26
0.92
| wikicorpus-j-to-e_bleu_en
float64 0.03
13
| wikicorpus-j-to-e_comet_wmt22
float64 0.29
0.78
| xlsum_ja_bert_score_ja_f1
float64 0
0.79
| xlsum_ja_bleu_ja
float64 0
7.25
| xlsum_ja_rouge1
float64 0.05
52.6
| xlsum_ja_rouge2
float64 0.01
28
| xlsum_ja_rouge2_scaling
float64 0
0.28
| xlsum_ja_rougeLsum
float64 0.05
44.6
| architecture
stringclasses 8
values | precision
stringclasses 2
values | license
stringclasses 11
values | params
float64 0.14
70.6
| likes
int64 0
4.03k
| revision
stringclasses 1
value | num_few_shot
int64 0
4
| add_special_tokens
stringclasses 2
values | llm_jp_eval_version
stringclasses 1
value | vllm_version
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
⭕ : instruction-tuned | llm-jp/llm-jp-3-1.8b-instruct | 0.3923 | 0.002 | 0.4046 | 0.1919 | 0.2842 | 0.3697 | 0.418 | 0.7999 | 0.4786 | 0.4957 | 0.834 | 0.0371 | 0.6003 | 0.8445 | 10.6032 | 0.8842 | 0.9337 | 12.773 | 0.8448 | 0.4046 | 0.5421 | 0.3822 | 0.4847 | 0.3101 | 0.4307 | 0.2895 | 0.5838 | 0.6976 | 0.2446 | 0.834 | 0.4386 | 0.5179 | 0.257 | 0.418 | 0.002 | 0.0141 | 0.2789 | 0.4561 | 0.0288 | 0.172 | 0.0177 | 0.0139 | 0.7272 | 0.7723 | 7.495 | 0.7693 | 0.8829 | 9.2901 | 0.7011 | 0.6449 | 1.1871 | 18.1463 | 3.7324 | 0.0371 | 15.2299 | LlamaForCausalLM | bfloat16 | apache-2.0 | 1.868 | 20 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | llm-jp/llm-jp-3-1.8b-instruct | 0.2147 | 0.002 | 0 | 0.0442 | 0.1461 | 0.2648 | 0 | 0.7888 | 0.2062 | 0.1956 | 0.6768 | 0.0371 | 0.2575 | 0.8361 | 10.3819 | 0.8796 | 0.931 | 12.1086 | 0.8421 | 0 | 0.5321 | 0.1724 | 0.0028 | 0 | 0.0767 | 0.0263 | 0.2424 | 0.0265 | 0.587 | 0.6768 | -0.0876 | -0.0888 | 0.2624 | 0 | 0.002 | 0.0141 | 0.2659 | 0.2526 | 0 | 0 | 0 | 0 | 0.2208 | 0.7485 | 6.3026 | 0.7504 | 0.873 | 7.9213 | 0.683 | 0.6449 | 1.1871 | 18.1463 | 3.7324 | 0.0371 | 15.2299 | LlamaForCausalLM | bfloat16 | apache-2.0 | 1.868 | 20 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | llm-jp/llm-jp-3-13b-instruct | 0.5462 | 0.0843 | 0.4843 | 0.2782 | 0.4739 | 0.8438 | 0.71 | 0.8393 | 0.669 | 0.6097 | 0.9014 | 0.1143 | 0.8415 | 0.8618 | 13.1859 | 0.9068 | 0.9476 | 14.866 | 0.8726 | 0.4843 | 0.8903 | 0.5575 | 0.6403 | 0.8928 | 0.3473 | 0.4651 | 0.7309 | 0.7481 | 0.6682 | 0.9014 | 0.7957 | 0.7949 | 0.7484 | 0.71 | 0.0843 | 0.2731 | 0.4827 | 0.6403 | 0.0396 | 0.4027 | 0.0354 | 0.047 | 0.8665 | 0.82 | 10.6349 | 0.8302 | 0.9031 | 10.9633 | 0.7478 | 0.7015 | 2.9771 | 29.3132 | 11.4455 | 0.1143 | 25.0207 | LlamaForCausalLM | bfloat16 | apache-2.0 | 13.708 | 16 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | llm-jp/llm-jp-3-13b-instruct | 0.412 | 0.0843 | 0.0104 | 0.1583 | 0.2727 | 0.7287 | 0.374 | 0.8316 | 0.6881 | 0.3822 | 0.8868 | 0.1143 | 0.6447 | 0.8561 | 12.9677 | 0.9014 | 0.9483 | 15.0672 | 0.8752 | 0.0104 | 0.7492 | 0.5431 | 0.6014 | 0.8704 | 0.0655 | 0.1519 | 0.7896 | 0.7424 | 0.7642 | 0.8868 | 0.8204 | 0.8177 | 0.5663 | 0.374 | 0.0843 | 0.2731 | 0.3935 | 0.4366 | 0 | 0.0025 | 0 | 0 | 0.7892 | 0.7918 | 8.7021 | 0.8163 | 0.8927 | 9.9921 | 0.7335 | 0.7015 | 2.9771 | 29.3132 | 11.4455 | 0.1143 | 25.0207 | LlamaForCausalLM | bfloat16 | apache-2.0 | 13.708 | 16 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | llm-jp/llm-jp-3-3.7b-instruct | 0.4597 | 0 | 0.3691 | 0.2398 | 0.3628 | 0.5839 | 0.58 | 0.822 | 0.551 | 0.5709 | 0.8488 | 0.1283 | 0.7141 | 0.8504 | 10.634 | 0.8959 | 0.9388 | 13.2399 | 0.8562 | 0.3691 | 0.7791 | 0.4282 | 0.5236 | 0.5407 | 0.4693 | 0.3555 | 0.6845 | 0.7374 | 0.3816 | 0.8488 | 0.7156 | 0.6453 | 0.4321 | 0.58 | 0 | 0 | 0.37 | 0.5293 | 0.0142 | 0.3409 | 0.0088 | 0.0575 | 0.7774 | 0.7961 | 8.5097 | 0.8052 | 0.8954 | 9.6317 | 0.7307 | 0.7119 | 2.9956 | 35.8746 | 12.8065 | 0.1283 | 29.3662 | LlamaForCausalLM | bfloat16 | apache-2.0 | 3.783 | 7 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | llm-jp/llm-jp-3-3.7b-instruct | 0.3016 | 0 | 0 | 0.115 | 0.1869 | 0.4517 | 0.01 | 0.8106 | 0.5867 | 0.2262 | 0.8022 | 0.1283 | 0.3361 | 0.8374 | 10.3631 | 0.8874 | 0.939 | 13.6521 | 0.8578 | 0 | 0.5321 | 0.3649 | 0.6028 | 0.4209 | 0.0591 | 0.0181 | 0.6002 | 0.7115 | 0.6539 | 0.8022 | 0.2802 | 0.3889 | 0.4022 | 0.01 | 0 | 0 | 0.3556 | 0.2835 | 0 | 0 | 0 | 0 | 0.5751 | 0.7703 | 7.1206 | 0.7906 | 0.8781 | 8.625 | 0.7064 | 0.7119 | 2.9956 | 35.8746 | 12.8065 | 0.1283 | 29.3662 | LlamaForCausalLM | bfloat16 | apache-2.0 | 3.783 | 7 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Qwen/Qwen2.5-32B-Instruct | 0.6553 | 0.5281 | 0.5894 | 0.2737 | 0.7757 | 0.8966 | 0.944 | 0.8479 | 0.8106 | 0.541 | 0.9047 | 0.097 | 0.553 | 0.8644 | 13.2738 | 0.9081 | 0.9554 | 17.7737 | 0.8859 | 0.5894 | 0.8975 | 0.6724 | 0.8431 | 0.958 | 0.5672 | 0.7515 | 0.8973 | 0.7835 | 0.8569 | 0.9047 | 0.8895 | 0.877 | 0.8343 | 0.944 | 0.5281 | 0.755 | 0.8 | 0.5029 | 0.0543 | 0.3837 | 0 | 0.1104 | 0.8204 | 0.8291 | 10.9975 | 0.8389 | 0.9045 | 11.1213 | 0.7585 | 0.6926 | 2.7959 | 25.855 | 9.7054 | 0.097 | 22.5323 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 120 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Qwen/Qwen2.5-32B-Instruct | 0.5443 | 0.5281 | 0.107 | 0.1453 | 0.568 | 0.8739 | 0.79 | 0.8386 | 0.7647 | 0.3873 | 0.8871 | 0.097 | 0.4392 | 0.8489 | 11.3776 | 0.9009 | 0.9511 | 15.766 | 0.8797 | 0.107 | 0.9046 | 0.6494 | 0.7944 | 0.9303 | 0.2681 | 0.5561 | 0.82 | 0.798 | 0.7615 | 0.8871 | 0.8951 | 0.8761 | 0.7869 | 0.79 | 0.5281 | 0.755 | 0.58 | 0.4547 | 0.0281 | 0.0071 | 0.0354 | 0.0058 | 0.6499 | 0.8004 | 8.7234 | 0.8268 | 0.8969 | 9.5439 | 0.7471 | 0.6926 | 2.7959 | 25.855 | 9.7054 | 0.097 | 22.5323 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 120 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Qwen/Qwen2.5-7B-Instruct | 0.5304 | 0.0221 | 0.498 | 0.1822 | 0.6385 | 0.8416 | 0.796 | 0.7302 | 0.7371 | 0.4006 | 0.8899 | 0.0983 | 0.3771 | 0.8222 | 11.2802 | 0.826 | 0.9327 | 15.5139 | 0.8347 | 0.498 | 0.859 | 0.6034 | 0.7431 | 0.9151 | 0.4305 | 0.6103 | 0.8295 | 0.6989 | 0.8106 | 0.8899 | 0.8729 | 0.8468 | 0.7507 | 0.796 | 0.0221 | 0.012 | 0.6666 | 0.3941 | 0.0449 | 0.2922 | 0.0354 | 0.0773 | 0.4614 | 0.7288 | 8.1478 | 0.6153 | 0.8592 | 9.0236 | 0.6446 | 0.692 | 2.1112 | 29.1184 | 9.8419 | 0.0983 | 23.7063 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 7.616 | 271 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Qwen/Qwen2.5-7B-Instruct | 0.2896 | 0.0221 | 0.2081 | 0.0822 | 0.4418 | 0.496 | 0.292 | 0.6415 | 0.2336 | 0.2228 | 0.4478 | 0.0983 | 0.2136 | 0.7191 | 7.042 | 0.7345 | 0.831 | 11.0762 | 0.5875 | 0.2081 | 0.0143 | 0.5029 | 0 | 0.8213 | 0.2105 | 0.4936 | 0.3246 | 0 | 0.3408 | 0.4478 | 0.875 | 0.8434 | 0.6524 | 0.292 | 0.0221 | 0.012 | 0.3899 | 0.2442 | 0.01 | 0.0234 | 0 | 0 | 0.3778 | 0.6795 | 6.3014 | 0.6785 | 0.8165 | 7.1661 | 0.5654 | 0.692 | 2.1112 | 29.1184 | 9.8419 | 0.0983 | 23.7063 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 7.616 | 271 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | weblab-GENIAC/Tanuki-8B-dpo-v1.0 | 0.3779 | 0.0783 | 0.398 | 0.1031 | 0.3067 | 0.4096 | 0.3 | 0.8275 | 0.5318 | 0.4266 | 0.6665 | 0.1083 | 0.6303 | 0.8351 | 10.0227 | 0.899 | 0.938 | 13.2353 | 0.8598 | 0.398 | 0.6964 | 0.4368 | 0.5431 | 0.2717 | 0.2847 | 0.3152 | 0.5965 | 0.661 | 0.4216 | 0.6665 | 0.3897 | 0.3894 | 0.2606 | 0.3 | 0.0783 | 0.2088 | 0.2983 | 0.3647 | 0 | 0.0212 | 0 | 0.0096 | 0.4849 | 0.7795 | 7.8083 | 0.8166 | 0.8903 | 8.5388 | 0.7346 | 0.7025 | 2.3051 | 34.8952 | 10.8447 | 0.1083 | 28.3762 | LlamaForCausalLM | bfloat16 | apache-2.0 | 7.512 | 29 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | weblab-GENIAC/Tanuki-8B-dpo-v1.0 | 0.2122 | 0.0783 | 0 | 0.0873 | 0 | 0.1774 | 0 | 0.829 | 0.414 | 0.1057 | 0.534 | 0.1083 | 0.0853 | 0.8373 | 10.8897 | 0.9 | 0.9427 | 12.9964 | 0.8684 | 0 | 0.5321 | 0.3563 | 0.3861 | 0 | 0.098 | 0 | 0.1553 | 0.5997 | 0.5726 | 0.534 | 0 | 0 | 0 | 0 | 0.0783 | 0.2088 | 0 | 0.1337 | 0 | 0 | 0 | 0 | 0.4366 | 0.7729 | 6.9734 | 0.8122 | 0.8905 | 7.4821 | 0.7355 | 0.7025 | 2.3051 | 34.8952 | 10.8447 | 0.1083 | 28.3762 | LlamaForCausalLM | bfloat16 | apache-2.0 | 7.512 | 29 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.1 | 0.62 | 0.0201 | 0.5529 | 0.3371 | 0.7313 | 0.9057 | 0.932 | 0.8565 | 0.7605 | 0.7131 | 0.9252 | 0.0856 | 0.8606 | 0.876 | 14.5217 | 0.9157 | 0.9603 | 19.3015 | 0.8924 | 0.5529 | 0.9296 | 0.6523 | 0.8069 | 0.9589 | 0.619 | 0.7063 | 0.7477 | 0.7923 | 0.8033 | 0.9252 | 0.878 | 0.8477 | 0.8287 | 0.932 | 0.0201 | 0.0422 | 0.7562 | 0.6597 | 0.0906 | 0.4663 | 0.1327 | 0.0903 | 0.9053 | 0.8501 | 13.863 | 0.8529 | 0.9137 | 12.9227 | 0.7651 | 0.6804 | 3.3097 | 20.1389 | 8.5601 | 0.0856 | 18.0061 | LlamaForCausalLM | bfloat16 | llama3.1;gemma | 70.554 | 3 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.1 | 0.507 | 0.0201 | 0.2715 | 0.1917 | 0.6465 | 0.8642 | 0.744 | 0.7167 | 0.7081 | 0.5798 | 0.749 | 0.0856 | 0.7149 | 0.8644 | 13.3248 | 0.9114 | 0.8297 | 17.2139 | 0.5661 | 0.2715 | 0.9216 | 0.6351 | 0.7458 | 0.9267 | 0.4623 | 0.6058 | 0.576 | 0.7753 | 0.8084 | 0.749 | 0.8408 | 0.7986 | 0.7443 | 0.744 | 0.0201 | 0.0422 | 0.6873 | 0.5622 | 0.0163 | 0.0057 | 0.0619 | 0.0075 | 0.867 | 0.8134 | 10.2308 | 0.8336 | 0.8206 | 10.8013 | 0.5556 | 0.6804 | 3.3097 | 20.1389 | 8.5601 | 0.0856 | 18.0061 | LlamaForCausalLM | bfloat16 | llama3.1;gemma | 70.554 | 3 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.2 | 0.565 | 0.0321 | 0.5385 | 0.2951 | 0.5743 | 0.8428 | 0.72 | 0.8484 | 0.7158 | 0.6428 | 0.9167 | 0.0883 | 0.7629 | 0.8709 | 13.8312 | 0.9115 | 0.9537 | 16.8957 | 0.8828 | 0.5385 | 0.8823 | 0.5345 | 0.725 | 0.9276 | 0.559 | 0.5507 | 0.8073 | 0.7601 | 0.752 | 0.9167 | 0.7888 | 0.7604 | 0.7184 | 0.72 | 0.0321 | 0.1365 | 0.5978 | 0.6066 | 0.0352 | 0.395 | 0.0885 | 0.1022 | 0.8549 | 0.835 | 11.7829 | 0.8406 | 0.9092 | 11.5867 | 0.7587 | 0.6827 | 2.8084 | 21.4684 | 8.8428 | 0.0883 | 18.9717 | LlamaForCausalLM | bfloat16 | llama3.1;gemma | 8.03 | 6 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.2 | 0.4292 | 0.0321 | 0.1047 | 0.1583 | 0.4096 | 0.799 | 0.294 | 0.8335 | 0.6602 | 0.4712 | 0.8707 | 0.0883 | 0.6063 | 0.8593 | 12.1636 | 0.905 | 0.9508 | 15.8558 | 0.8781 | 0.1047 | 0.8682 | 0.5172 | 0.6528 | 0.8838 | 0.4527 | 0.3663 | 0.6167 | 0.7601 | 0.7542 | 0.8707 | 0.8008 | 0.7914 | 0.6449 | 0.294 | 0.0321 | 0.1365 | 0.4529 | 0.3547 | 0.0026 | 0.0004 | 0.0088 | 0 | 0.7798 | 0.8002 | 8.6761 | 0.818 | 0.8933 | 10.2604 | 0.7328 | 0.6827 | 2.8084 | 21.4684 | 8.8428 | 0.0883 | 18.9717 | LlamaForCausalLM | bfloat16 | llama3.1;gemma | 8.03 | 6 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | rinna/gemma-2-baku-2b-it | 0.4477 | 0 | 0.4434 | 0.1815 | 0.4444 | 0.7059 | 0.536 | 0.8302 | 0.4111 | 0.3726 | 0.8741 | 0.1256 | 0.4408 | 0.8527 | 10.6366 | 0.8981 | 0.9441 | 13.6152 | 0.8697 | 0.4434 | 0.7004 | 0.4023 | 0.5639 | 0.8534 | 0.2712 | 0.4084 | 0.4129 | 0.2986 | 0.3779 | 0.8741 | 0.6961 | 0.7189 | 0.5639 | 0.536 | 0 | 0 | 0.4805 | 0.4057 | 0.0348 | 0.0877 | 0.0708 | 0.0355 | 0.6786 | 0.7845 | 7.0376 | 0.8119 | 0.8944 | 8.8013 | 0.741 | 0.7177 | 2.2358 | 37.6981 | 12.5686 | 0.1256 | 30.0896 | Gemma2ForCausalLM | float16 | gemma | 2.614 | 14 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | rinna/gemma-2-baku-2b-it | 0.2449 | 0 | 0.0093 | 0.1214 | 0.0013 | 0.2852 | 0.008 | 0.823 | 0.3259 | 0.2184 | 0.7761 | 0.1256 | 0.3312 | 0.8416 | 9.5574 | 0.8942 | 0.9392 | 11.6762 | 0.8618 | 0.0093 | 0.8021 | 0.3649 | 0.5014 | 0 | 0.0083 | 0.0025 | 0.2991 | 0.1629 | 0.3012 | 0.7761 | 0.2881 | 0.2782 | 0.0533 | 0.008 | 0 | 0 | 0 | 0.3157 | 0 | 0 | 0 | 0 | 0.6068 | 0.7784 | 6.6772 | 0.7992 | 0.8924 | 7.8981 | 0.7369 | 0.7177 | 2.2358 | 37.6981 | 12.5686 | 0.1256 | 30.0896 | Gemma2ForCausalLM | float16 | gemma | 2.614 | 14 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/calm3-22b-chat | 0.5612 | 0 | 0.5404 | 0.247 | 0.5592 | 0.8527 | 0.684 | 0.8464 | 0.7519 | 0.6677 | 0.9118 | 0.1121 | 0.8433 | 0.8705 | 13.6161 | 0.9117 | 0.9553 | 16.6257 | 0.8857 | 0.5404 | 0.8677 | 0.5805 | 0.7875 | 0.933 | 0.5109 | 0.5408 | 0.8583 | 0.7715 | 0.7617 | 0.9118 | 0.8637 | 0.86 | 0.7575 | 0.684 | 0 | 0 | 0.5776 | 0.6489 | 0.001 | 0.2251 | 0.1239 | 0.0316 | 0.8532 | 0.8177 | 9.4163 | 0.8375 | 0.9039 | 10.3951 | 0.7506 | 0.7019 | 2.5206 | 31.9179 | 11.2189 | 0.1121 | 26.8509 | LlamaForCausalLM | bfloat16 | apache-2.0 | 22.543 | 67 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/calm3-22b-chat | 0.4792 | 0 | 0.2776 | 0.1668 | 0.4546 | 0.7844 | 0.578 | 0.8388 | 0.6843 | 0.4854 | 0.8887 | 0.1121 | 0.6197 | 0.8517 | 12.6994 | 0.907 | 0.9535 | 16.1239 | 0.8829 | 0.2776 | 0.8297 | 0.5517 | 0.6986 | 0.8954 | 0.3577 | 0.4818 | 0.7683 | 0.7689 | 0.6341 | 0.8887 | 0.8853 | 0.8589 | 0.6281 | 0.578 | 0 | 0 | 0.4275 | 0.4788 | 0.0019 | 0.0016 | 0.0022 | 0 | 0.8285 | 0.7865 | 7.7148 | 0.8174 | 0.9002 | 9.9171 | 0.748 | 0.7019 | 2.5206 | 31.9179 | 11.2189 | 0.1121 | 26.8509 | LlamaForCausalLM | bfloat16 | apache-2.0 | 22.543 | 67 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | elyza/Llama-3-ELYZA-JP-8B | 0.556 | 0.2972 | 0.5105 | 0.27 | 0.4999 | 0.8048 | 0.712 | 0.8285 | 0.6305 | 0.5457 | 0.9112 | 0.1058 | 0.6179 | 0.8476 | 11.1585 | 0.8843 | 0.9453 | 14.1463 | 0.8677 | 0.5105 | 0.8166 | 0.4856 | 0.6069 | 0.8981 | 0.5168 | 0.4761 | 0.7555 | 0.6654 | 0.6389 | 0.9112 | 0.8215 | 0.7794 | 0.6995 | 0.712 | 0.2972 | 0.7269 | 0.5236 | 0.5023 | 0.0209 | 0.3524 | 0.0885 | 0.0848 | 0.8035 | 0.8125 | 9.6534 | 0.8214 | 0.8976 | 10.308 | 0.7407 | 0.6986 | 2.4239 | 28.7904 | 10.5753 | 0.1058 | 24.7182 | LlamaForCausalLM | bfloat16 | llama3 | 8.03 | 76 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | elyza/Llama-3-ELYZA-JP-8B | 0.4106 | 0.2972 | 0.0401 | 0.1057 | 0.3654 | 0.5815 | 0.446 | 0.8151 | 0.549 | 0.3639 | 0.8475 | 0.1058 | 0.4164 | 0.8396 | 9.0926 | 0.886 | 0.94 | 12.3152 | 0.8587 | 0.0401 | 0.7778 | 0.3678 | 0.4972 | 0.563 | 0.3615 | 0.3211 | 0.5415 | 0.7096 | 0.629 | 0.8475 | 0.4597 | 0.4683 | 0.4036 | 0.446 | 0.2972 | 0.7269 | 0.4098 | 0.3137 | 0 | 0.0021 | 0 | 0.0006 | 0.5257 | 0.7819 | 7.1661 | 0.7972 | 0.89 | 9.2311 | 0.7182 | 0.6986 | 2.4239 | 28.7904 | 10.5753 | 0.1058 | 24.7182 | LlamaForCausalLM | bfloat16 | llama3 | 8.03 | 76 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Rakuten/RakutenAI-7B-chat | 0.4627 | 0.0863 | 0.3989 | 0.1521 | 0.4763 | 0.8122 | 0.41 | 0.8169 | 0.5524 | 0.418 | 0.8645 | 0.1017 | 0.5386 | 0.8624 | 13.0202 | 0.8952 | 0.9509 | 16.8546 | 0.8778 | 0.3989 | 0.7708 | 0.4713 | 0.6458 | 0.9053 | 0.26 | 0.4377 | 0.7449 | 0.4444 | 0.4554 | 0.8645 | 0.7956 | 0.7642 | 0.7605 | 0.41 | 0.0863 | 0.3072 | 0.515 | 0.4555 | 0.0108 | 0.1118 | 0.0442 | 0.0445 | 0.5493 | 0.7868 | 9.6886 | 0.7865 | 0.8938 | 9.7413 | 0.708 | 0.6961 | 2.2839 | 28.4861 | 10.1816 | 0.1017 | 23.911 | MistralForCausalLM | bfloat16 | apache-2.0 | 7.373 | 59 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Rakuten/RakutenAI-7B-chat | 0.3167 | 0.0863 | 0.0067 | 0.0985 | 0.0041 | 0.7895 | 0.02 | 0.8058 | 0.6148 | 0.2592 | 0.6967 | 0.1017 | 0.3986 | 0.8555 | 13.0705 | 0.8908 | 0.9425 | 14.4667 | 0.8668 | 0.0067 | 0.7455 | 0.477 | 0.6583 | 0.9133 | 0.0218 | 0.0003 | 0.8509 | 0.5941 | 0.4936 | 0.6967 | 0.7406 | 0.7445 | 0.7097 | 0.02 | 0.0863 | 0.3072 | 0.0078 | 0.3572 | 0.0033 | 0 | 0 | 0 | 0.4891 | 0.7883 | 8.5147 | 0.7951 | 0.8782 | 9.0212 | 0.6703 | 0.6961 | 2.2839 | 28.4861 | 10.1816 | 0.1017 | 23.911 | MistralForCausalLM | bfloat16 | apache-2.0 | 7.373 | 59 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/calm2-7b-chat | 0.3474 | 0 | 0.3317 | 0.1411 | 0.254 | 0.3255 | 0.108 | 0.7897 | 0.5213 | 0.5648 | 0.7719 | 0.0136 | 0.7041 | 0.8445 | 11.3228 | 0.8869 | 0.9368 | 13.4747 | 0.8479 | 0.3317 | 0.4679 | 0.4052 | 0.5028 | 0.2395 | 0.4973 | 0.2426 | 0.6253 | 0.4646 | 0.6085 | 0.7719 | -0.0728 | -0.0108 | 0.2691 | 0.108 | 0 | 0 | 0.2654 | 0.4929 | 0.0025 | 0.0851 | 0.0177 | 0.0324 | 0.5676 | 0.7587 | 6.7623 | 0.7498 | 0.8779 | 8.7567 | 0.6743 | 0.5871 | 0.4037 | 8.8424 | 1.354 | 0.0136 | 7.5443 | LlamaForCausalLM | bfloat16 | apache-2.0 | 7.009 | 76 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/calm2-7b-chat | 0.1367 | 0 | 0 | 0.0496 | 0.0416 | 0.1877 | 0.004 | 0.4887 | 0.0005 | 0.2584 | 0.4593 | 0.0136 | 0.4671 | 0.6331 | 2.2109 | 0.4951 | 0.8192 | 4.0251 | 0.5615 | 0 | 0.4679 | 0 | 0 | 0.0831 | 0.0573 | 0.0387 | 0 | 0.0025 | 0 | 0.4593 | 0.0689 | 0.0684 | 0.0121 | 0.004 | 0 | 0 | 0.0446 | 0.2508 | 0 | 0 | 0 | 0 | 0.2481 | 0.5869 | 1.81 | 0.4237 | 0.7916 | 2.6109 | 0.4745 | 0.5871 | 0.4037 | 8.8424 | 1.354 | 0.0136 | 7.5443 | LlamaForCausalLM | bfloat16 | apache-2.0 | 7.009 | 76 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🟢 : pretrained | tokyotech-llm/Llama-3.1-Swallow-8B-v0.2 | 0.5414 | 0.012 | 0.4659 | 0.2799 | 0.5427 | 0.8066 | 0.732 | 0.8445 | 0.6617 | 0.6405 | 0.8886 | 0.0813 | 0.8015 | 0.8747 | 13.8553 | 0.9099 | 0.9546 | 16.517 | 0.883 | 0.4659 | 0.8555 | 0.5259 | 0.7125 | 0.9071 | 0.4925 | 0.5244 | 0.7773 | 0.7595 | 0.5334 | 0.8886 | 0.7765 | 0.7511 | 0.6572 | 0.732 | 0.012 | 0.0562 | 0.561 | 0.6276 | 0.007 | 0.3933 | 0.0531 | 0.0803 | 0.8658 | 0.8368 | 13.1609 | 0.8317 | 0.9093 | 12.1377 | 0.7534 | 0.6801 | 1.9566 | 20.1115 | 8.1373 | 0.0813 | 17.7927 | LlamaForCausalLM | bfloat16 | llama3.1 | 8.03 | 2 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🟢 : pretrained | tokyotech-llm/Llama-3.1-Swallow-8B-v0.2 | 0.3813 | 0.012 | 0.024 | 0.1587 | 0.3409 | 0.623 | 0.382 | 0.8264 | 0.4968 | 0.4812 | 0.7682 | 0.0813 | 0.6752 | 0.8586 | 12.1659 | 0.8998 | 0.9498 | 14.0954 | 0.8754 | 0.024 | 0.8034 | 0.4684 | 0.5 | 0.6604 | 0.2578 | 0.2906 | 0.3878 | 0.7045 | 0.423 | 0.7682 | 0.0806 | 0.0855 | 0.4052 | 0.382 | 0.012 | 0.0562 | 0.3911 | 0.5105 | 0.002 | 0.0016 | 0 | 0 | 0.79 | 0.7912 | 8.2683 | 0.8 | 0.8949 | 9.4323 | 0.7302 | 0.6801 | 1.9566 | 20.1115 | 8.1373 | 0.0813 | 17.7927 | LlamaForCausalLM | bfloat16 | llama3.1 | 8.03 | 2 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🟢 : pretrained | llm-jp/llm-jp-3-13b | 0.4861 | 0 | 0.4856 | 0.256 | 0.4043 | 0.6456 | 0.598 | 0.8373 | 0.5444 | 0.6721 | 0.8821 | 0.0218 | 0.8577 | 0.8682 | 13.5301 | 0.9064 | 0.95 | 16.6943 | 0.8747 | 0.4856 | 0.8507 | 0.3822 | 0.5028 | 0.6506 | 0.5161 | 0.3979 | 0.6011 | 0.7311 | 0.505 | 0.8821 | 0.2931 | 0.2998 | 0.4356 | 0.598 | 0 | 0.9578 | 0.4108 | 0.6424 | 0.0068 | 0.3231 | 0.0442 | 0.0366 | 0.8693 | 0.828 | 11.1755 | 0.8298 | 0.9036 | 11.0157 | 0.7381 | 0.6126 | 0.6253 | 13.6586 | 2.1977 | 0.0218 | 11.752 | LlamaForCausalLM | bfloat16 | apache-2.0 | 13.708 | 0 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🟢 : pretrained | llm-jp/llm-jp-3-13b | 0.1505 | 0 | 0.0021 | 0.1179 | 0.0097 | 0.0003 | 0.006 | 0.6423 | 0.0166 | 0.308 | 0.5313 | 0.0218 | 0.4529 | 0.758 | 7.0049 | 0.7205 | 0.8881 | 10.7045 | 0.7285 | 0.0021 | 0.0008 | 0 | 0.0806 | 0 | 0.1155 | 0.0175 | 0.0008 | 0 | 0.0016 | 0.5313 | 0.2138 | 0.2025 | 0 | 0.006 | 0 | 0.9578 | 0.0019 | 0.3557 | 0 | 0 | 0 | 0 | 0.5897 | 0.6568 | 3.6264 | 0.5514 | 0.8298 | 5.779 | 0.5687 | 0.6126 | 0.6253 | 13.6586 | 2.1977 | 0.0218 | 11.752 | LlamaForCausalLM | bfloat16 | apache-2.0 | 13.708 | 0 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.1-8B-Instruct | 0.5319 | 0.3213 | 0.4598 | 0.2316 | 0.5808 | 0.7657 | 0.736 | 0.7522 | 0.6298 | 0.4205 | 0.8881 | 0.0648 | 0.419 | 0.8496 | 11.4306 | 0.8909 | 0.9171 | 15.0559 | 0.7837 | 0.4598 | 0.8011 | 0.4971 | 0.6722 | 0.8838 | 0.4502 | 0.5171 | 0.7173 | 0.6553 | 0.6069 | 0.8881 | 0.7479 | 0.7498 | 0.6121 | 0.736 | 0.3213 | 0.6205 | 0.6446 | 0.3923 | 0.0187 | 0.3105 | 0.1239 | 0.0315 | 0.6735 | 0.7956 | 11.0723 | 0.7762 | 0.8332 | 9.7048 | 0.5581 | 0.663 | 3.1983 | 15.7749 | 6.481 | 0.0648 | 14.169 | LlamaForCausalLM | bfloat16 | llama3.1 | 8.03 | 3,059 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.1-8B-Instruct | 0.2303 | 0.3213 | 0.0154 | 0.0483 | 0.0038 | 0.3839 | 0.076 | 0.4675 | 0.2672 | 0.2386 | 0.6467 | 0.0648 | 0.2288 | 0.6335 | 7.9862 | 0.5261 | 0.7807 | 12.5089 | 0.4486 | 0.0154 | 0.0008 | 0.4425 | 0.3583 | 0.7131 | 0.2612 | 0.0028 | 0.2219 | 0.0063 | 0.3071 | 0.6467 | 0.6254 | 0.5865 | 0.4377 | 0.076 | 0.3213 | 0.6205 | 0.0047 | 0.2259 | 0 | 0.008 | 0 | 0 | 0.2334 | 0.6074 | 7.3187 | 0.4883 | 0.7688 | 7.4026 | 0.4071 | 0.663 | 3.1983 | 15.7749 | 6.481 | 0.0648 | 14.169 | LlamaForCausalLM | bfloat16 | llama3.1 | 8.03 | 3,059 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-23-8B | 0.484 | 0 | 0.3829 | 0.1502 | 0.4958 | 0.7439 | 0.578 | 0.8245 | 0.6414 | 0.4883 | 0.8982 | 0.1213 | 0.5051 | 0.8583 | 12.3567 | 0.9043 | 0.9507 | 16.5267 | 0.8763 | 0.3829 | 0.748 | 0.5345 | 0.6611 | 0.8731 | 0.4958 | 0.4671 | 0.6187 | 0.7153 | 0.6773 | 0.8982 | 0.7539 | 0.7297 | 0.6105 | 0.578 | 0 | 0 | 0.5244 | 0.4639 | 0.0198 | 0.1152 | 0.0354 | 0.0154 | 0.5652 | 0.7963 | 8.9616 | 0.79 | 0.8934 | 9.7535 | 0.7274 | 0.7072 | 2.6432 | 30.8244 | 12.1131 | 0.1213 | 26.0328 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 8.028 | 392 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-23-8B | 0.31 | 0 | 0.0009 | 0.0738 | 0.0287 | 0.5859 | 0.176 | 0.8075 | 0.4615 | 0.3263 | 0.8285 | 0.1213 | 0.4499 | 0.8423 | 10.334 | 0.893 | 0.9461 | 14.353 | 0.8684 | 0.0009 | 0.8041 | 0.3534 | 0.5222 | 0.5371 | 0.1938 | 0.05 | 0.2091 | 0.6742 | 0.5486 | 0.8285 | 0.504 | 0.4839 | 0.4165 | 0.176 | 0 | 0 | 0.0075 | 0.3353 | 0.0054 | 0 | 0 | 0 | 0.3636 | 0.7675 | 7.5791 | 0.7666 | 0.8807 | 8.2645 | 0.7021 | 0.7072 | 2.6432 | 30.8244 | 12.1131 | 0.1213 | 26.0328 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 8.028 | 392 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-23-35B | 0.5742 | 0.1888 | 0.5045 | 0.2595 | 0.6038 | 0.8407 | 0.692 | 0.8443 | 0.7218 | 0.5991 | 0.9175 | 0.1443 | 0.7088 | 0.8668 | 13.519 | 0.9097 | 0.9564 | 17.9446 | 0.8862 | 0.5045 | 0.8236 | 0.569 | 0.7833 | 0.9446 | 0.5383 | 0.5688 | 0.7091 | 0.7538 | 0.7936 | 0.9175 | 0.8689 | 0.8381 | 0.754 | 0.692 | 0.1888 | 0.3936 | 0.6389 | 0.5501 | 0.022 | 0.3005 | 0.115 | 0.043 | 0.8169 | 0.8337 | 11.9736 | 0.8333 | 0.9054 | 11.0157 | 0.7482 | 0.7212 | 4.1534 | 34.0784 | 14.4071 | 0.1443 | 29.011 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 34.981 | 264 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-23-35B | 0.4229 | 0.1888 | 0.0323 | 0.1458 | 0.3212 | 0.7576 | 0.346 | 0.8366 | 0.6044 | 0.5036 | 0.7708 | 0.1443 | 0.6728 | 0.8567 | 12.0892 | 0.9046 | 0.9545 | 16.4732 | 0.8841 | 0.0323 | 0.8582 | 0.4971 | 0.5847 | 0.8418 | 0.3595 | 0.1573 | 0.4626 | 0.7128 | 0.7646 | 0.7708 | 0.7269 | 0.7625 | 0.5728 | 0.346 | 0.1888 | 0.3936 | 0.4852 | 0.4784 | 0 | 0.0017 | 0 | 0 | 0.7275 | 0.8127 | 9.6579 | 0.8224 | 0.8987 | 10.0728 | 0.7353 | 0.7212 | 4.1534 | 34.0784 | 14.4071 | 0.1443 | 29.011 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 34.981 | 264 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-expanse-8b | 0.5163 | 0 | 0.5127 | 0.2133 | 0.5521 | 0.8281 | 0.64 | 0.8427 | 0.5808 | 0.5008 | 0.9005 | 0.1088 | 0.5245 | 0.8658 | 13.3065 | 0.9096 | 0.9526 | 16.1479 | 0.8817 | 0.5127 | 0.864 | 0.5086 | 0.7 | 0.9223 | 0.4943 | 0.5137 | 0.4906 | 0.7393 | 0.4658 | 0.9005 | 0.8382 | 0.8038 | 0.6981 | 0.64 | 0 | 0 | 0.5905 | 0.4835 | 0.0103 | 0.2809 | 0.0708 | 0.032 | 0.6724 | 0.8152 | 9.3487 | 0.8304 | 0.8988 | 9.9183 | 0.7488 | 0.6887 | 2.3612 | 30.7157 | 10.8872 | 0.1088 | 24.9318 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 8.028 | 276 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-expanse-8b | 0.371 | 0 | 0.0296 | 0.133 | 0.227 | 0.6603 | 0.398 | 0.8352 | 0.6017 | 0.3012 | 0.7864 | 0.1088 | 0.331 | 0.8544 | 12.701 | 0.9078 | 0.953 | 16.887 | 0.8831 | 0.0296 | 0.8464 | 0.4626 | 0.6333 | 0.6962 | 0.3262 | 0.0387 | 0.5066 | 0.7513 | 0.6548 | 0.7864 | 0.8373 | 0.8031 | 0.4384 | 0.398 | 0 | 0 | 0.4153 | 0.2465 | 0.0103 | 0 | 0 | 0.0017 | 0.6531 | 0.7814 | 7.1072 | 0.8112 | 0.8903 | 9.1866 | 0.7388 | 0.6887 | 2.3612 | 30.7157 | 10.8872 | 0.1088 | 24.9318 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 8.028 | 276 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-expanse-32b | 0.5082 | 0.1426 | 0.535 | 0.2935 | 0.4225 | 0.8468 | 0.206 | 0.8553 | 0.5893 | 0.6495 | 0.9112 | 0.1388 | 0.7489 | 0.8755 | 14.7691 | 0.9167 | 0.9593 | 19.4268 | 0.8912 | 0.535 | 0.774 | 0.5891 | 0.1556 | 0.9625 | 0.5815 | 0.1367 | 0.8311 | 0.7822 | 0.5884 | 0.9112 | 0.8863 | 0.8648 | 0.8038 | 0.206 | 0.1426 | 0.255 | 0.7082 | 0.6179 | 0.0228 | 0.4033 | 0.0855 | 0.0853 | 0.8705 | 0.8462 | 13.016 | 0.8511 | 0.9099 | 11.557 | 0.7621 | 0.7167 | 3.6554 | 35.73 | 13.8915 | 0.1388 | 30.2516 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 32.296 | 168 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | CohereForAI/aya-expanse-32b | 0.4795 | 0.1426 | 0.2165 | 0.1746 | 0.5658 | 0.84 | 0.514 | 0.8516 | 0.3936 | 0.5438 | 0.893 | 0.1388 | 0.6921 | 0.8627 | 13.3137 | 0.9142 | 0.9581 | 17.7546 | 0.8913 | 0.2165 | 0.883 | 0.5747 | 0.0306 | 0.9205 | 0.4042 | 0.5916 | 0.5645 | 0.3371 | 0.4613 | 0.893 | 0.8629 | 0.8428 | 0.7166 | 0.514 | 0.1426 | 0.255 | 0.5399 | 0.535 | 0.0115 | 0 | 0.0177 | 0.0043 | 0.8395 | 0.8184 | 9.7295 | 0.8422 | 0.9034 | 10.1636 | 0.7585 | 0.7167 | 3.6554 | 35.73 | 13.8915 | 0.1388 | 30.2516 | CohereForCausalLM | float16 | cc-by-nc-4.0 | 32.296 | 168 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | google/gemma-2-2b-it | 0.4051 | 0 | 0.3826 | 0.094 | 0.4379 | 0.5978 | 0.492 | 0.7515 | 0.5586 | 0.2522 | 0.8379 | 0.0513 | 0.2125 | 0.8197 | 8.5773 | 0.8372 | 0.9333 | 13.6777 | 0.8375 | 0.3826 | 0.5754 | 0.4253 | 0.5208 | 0.7819 | 0.3277 | 0.3838 | 0.5711 | 0.7014 | 0.5742 | 0.8379 | 0.5681 | 0.5442 | 0.4359 | 0.492 | 0 | 0 | 0.4921 | 0.2165 | 0.0123 | 0.1042 | 0.0973 | 0.0127 | 0.2434 | 0.7388 | 6.3148 | 0.6881 | 0.8677 | 7.5156 | 0.6434 | 0.6573 | 1.0802 | 16.8494 | 5.1204 | 0.0513 | 13.5661 | Gemma2ForCausalLM | bfloat16 | gemma | 2.614 | 678 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | google/gemma-2-2b-it | 0.1798 | 0 | 0 | 0.0487 | 0.0075 | 0.2901 | 0.044 | 0.576 | 0.2194 | 0.1283 | 0.6129 | 0.0513 | 0.0651 | 0.7304 | 5.5835 | 0.6514 | 0.792 | 3.1738 | 0.6637 | 0 | 0.4123 | 0.3276 | 0.0056 | 0.2055 | 0.2203 | 0.0031 | 0.145 | 0.3851 | 0.2336 | 0.6129 | 0.25 | 0.2473 | 0.2524 | 0.044 | 0 | 0 | 0.0118 | 0.0994 | 0 | 0 | 0 | 0 | 0.2434 | 0.6395 | 3.479 | 0.5003 | 0.7665 | 3.0008 | 0.4887 | 0.6573 | 1.0802 | 16.8494 | 5.1204 | 0.0513 | 13.5661 | Gemma2ForCausalLM | bfloat16 | gemma | 2.614 | 678 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | google/gemma-2-9b-it | 0.5206 | 0 | 0.45 | 0.2212 | 0.6149 | 0.801 | 0.75 | 0.8332 | 0.6079 | 0.4357 | 0.8883 | 0.1244 | 0.4964 | 0.8542 | 11.4808 | 0.8905 | 0.9446 | 15.7498 | 0.8611 | 0.45 | 0.8249 | 0.5 | 0.6819 | 0.8695 | 0.4161 | 0.5586 | 0.5592 | 0.6755 | 0.6231 | 0.8883 | 0.8474 | 0.815 | 0.7085 | 0.75 | 0 | 0 | 0.6713 | 0.3946 | 0.0114 | 0.294 | 0.0973 | 0.0792 | 0.624 | 0.8201 | 9.9977 | 0.8297 | 0.9044 | 10.9491 | 0.7513 | 0.7114 | 2.5273 | 36.1583 | 12.4421 | 0.1244 | 29.7442 | Gemma2ForCausalLM | bfloat16 | gemma | 9.242 | 551 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | google/gemma-2-9b-it | 0.3631 | 0 | 0 | 0.053 | 0.3421 | 0.6453 | 0.512 | 0.7368 | 0.6036 | 0.2294 | 0.7473 | 0.1244 | 0.196 | 0.8117 | 9.1697 | 0.8305 | 0.929 | 12.8368 | 0.8198 | 0 | 0.7219 | 0.5115 | 0.5764 | 0.7605 | 0.2548 | 0.4177 | 0.493 | 0.7197 | 0.7175 | 0.7473 | 0.7495 | 0.7103 | 0.4536 | 0.512 | 0 | 0 | 0.2665 | 0.2373 | 0.005 | 0.0021 | 0 | 0 | 0.2578 | 0.7155 | 6.0728 | 0.6682 | 0.8632 | 8.6791 | 0.6287 | 0.7114 | 2.5273 | 36.1583 | 12.4421 | 0.1244 | 29.7442 | Gemma2ForCausalLM | bfloat16 | gemma | 9.242 | 551 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | google/gemma-2-27b-it | 0.5809 | 0 | 0.4993 | 0.2738 | 0.676 | 0.8434 | 0.9 | 0.8539 | 0.7242 | 0.5932 | 0.8934 | 0.1322 | 0.6927 | 0.8709 | 13.2355 | 0.9126 | 0.9551 | 18.5224 | 0.8844 | 0.4993 | 0.8707 | 0.6552 | 0.8014 | 0.9267 | 0.5262 | 0.6278 | 0.6935 | 0.7727 | 0.6982 | 0.8934 | 0.9006 | 0.8776 | 0.7329 | 0.9 | 0 | 0 | 0.7243 | 0.5607 | 0.0127 | 0.3508 | 0.1062 | 0.0692 | 0.8303 | 0.8457 | 12.4513 | 0.8504 | 0.9127 | 12.1106 | 0.7681 | 0.7162 | 3.0568 | 35.8752 | 13.2254 | 0.1322 | 29.9498 | Gemma2ForCausalLM | bfloat16 | gemma | 27.227 | 443 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | google/gemma-2-27b-it | 0.4948 | 0 | 0.4455 | 0.1595 | 0.6023 | 0.721 | 0.798 | 0.8253 | 0.6287 | 0.449 | 0.6812 | 0.1322 | 0.5564 | 0.8321 | 11.1934 | 0.8824 | 0.9501 | 15.8838 | 0.8764 | 0.4455 | 0.7823 | 0.5259 | 0.6722 | 0.8365 | 0.3194 | 0.5479 | 0.521 | 0.7519 | 0.6726 | 0.6812 | 0.8502 | 0.8294 | 0.5442 | 0.798 | 0 | 0 | 0.6567 | 0.4713 | 0.0075 | 0.0012 | 0.0177 | 0 | 0.7712 | 0.7655 | 8.7339 | 0.7965 | 0.899 | 10.1975 | 0.7459 | 0.7162 | 3.0568 | 35.8752 | 13.2254 | 0.1322 | 29.9498 | Gemma2ForCausalLM | bfloat16 | gemma | 27.227 | 443 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.1-70B-Instruct | 0.6638 | 0.6004 | 0.5611 | 0.2644 | 0.7678 | 0.8854 | 0.93 | 0.8467 | 0.7821 | 0.6456 | 0.9215 | 0.097 | 0.7264 | 0.8675 | 13.7097 | 0.9095 | 0.9576 | 18.4691 | 0.8881 | 0.5611 | 0.9031 | 0.658 | 0.8556 | 0.95 | 0.6331 | 0.7258 | 0.7823 | 0.7992 | 0.8155 | 0.9215 | 0.8729 | 0.8407 | 0.8032 | 0.93 | 0.6004 | 0.9719 | 0.8099 | 0.5773 | 0.0808 | 0.2779 | 0.0442 | 0.0576 | 0.8616 | 0.8444 | 16.6273 | 0.8355 | 0.9073 | 12.3916 | 0.7537 | 0.6883 | 3.4286 | 23.6076 | 9.7127 | 0.097 | 20.9281 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 685 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.1-70B-Instruct | 0.4997 | 0.6004 | 0.2763 | 0.153 | 0.6753 | 0.8254 | 0.09 | 0.8243 | 0.7071 | 0.3729 | 0.8744 | 0.097 | 0.4822 | 0.8627 | 12.6354 | 0.9047 | 0.8888 | 11.0722 | 0.8239 | 0.2763 | 0.893 | 0.6063 | 0.7278 | 0.8579 | 0.2659 | 0.6543 | 0.5793 | 0.7841 | 0.8382 | 0.8744 | 0.8862 | 0.8528 | 0.7254 | 0.09 | 0.6004 | 0.9719 | 0.6963 | 0.3706 | 0.002 | 0.0257 | 0.0265 | 0.0008 | 0.7099 | 0.8326 | 12.7335 | 0.8355 | 0.8921 | 10.8621 | 0.7332 | 0.6883 | 3.4286 | 23.6076 | 9.7127 | 0.097 | 20.9281 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 685 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | nvidia/Llama-3.1-Nemotron-70B-Instruct-HF | 0.6624 | 0.5582 | 0.5734 | 0.2715 | 0.7673 | 0.8851 | 0.928 | 0.8495 | 0.7889 | 0.6453 | 0.9202 | 0.0984 | 0.727 | 0.8668 | 13.9208 | 0.9099 | 0.9577 | 18.3148 | 0.8882 | 0.5734 | 0.9051 | 0.6609 | 0.8694 | 0.9464 | 0.6205 | 0.7224 | 0.8028 | 0.7891 | 0.8224 | 0.9202 | 0.881 | 0.8533 | 0.8037 | 0.928 | 0.5582 | 0.8896 | 0.8123 | 0.5883 | 0.0984 | 0.3077 | 0.0265 | 0.0668 | 0.8582 | 0.8485 | 16.7533 | 0.8436 | 0.9059 | 12.0753 | 0.7565 | 0.6918 | 3.4088 | 23.8265 | 9.8542 | 0.0984 | 21.032 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 1,680 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | nvidia/Llama-3.1-Nemotron-70B-Instruct-HF | 0.5134 | 0.5582 | 0.2532 | 0.1623 | 0.6794 | 0.8242 | 0.232 | 0.8424 | 0.7365 | 0.3953 | 0.8654 | 0.0984 | 0.4611 | 0.8644 | 12.7358 | 0.9064 | 0.9481 | 16.8129 | 0.8795 | 0.2532 | 0.893 | 0.6408 | 0.7583 | 0.8642 | 0.3665 | 0.6614 | 0.6537 | 0.7633 | 0.8662 | 0.8654 | 0.8959 | 0.8638 | 0.7154 | 0.232 | 0.5582 | 0.8896 | 0.6975 | 0.3583 | 0.0024 | 0.0281 | 0.0177 | 0 | 0.7631 | 0.8328 | 12.6953 | 0.8383 | 0.9004 | 11.1724 | 0.7455 | 0.6918 | 3.4088 | 23.8265 | 9.8542 | 0.0984 | 21.032 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 1,680 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/Llama-3.1-70B-Japanese-Instruct-2407 | 0.6705 | 0.5823 | 0.5464 | 0.2862 | 0.7522 | 0.9108 | 0.924 | 0.8518 | 0.7869 | 0.7169 | 0.9188 | 0.0989 | 0.8472 | 0.8747 | 15.2977 | 0.9123 | 0.9586 | 18.0168 | 0.8888 | 0.5464 | 0.9311 | 0.658 | 0.85 | 0.9589 | 0.6432 | 0.7159 | 0.7634 | 0.8106 | 0.8526 | 0.9188 | 0.878 | 0.8431 | 0.8424 | 0.924 | 0.5823 | 0.9699 | 0.7885 | 0.6602 | 0.0513 | 0.416 | 0.0177 | 0.0609 | 0.8852 | 0.8463 | 14.9733 | 0.8474 | 0.9099 | 12.2264 | 0.7588 | 0.6921 | 3.1649 | 23.7422 | 9.8925 | 0.0989 | 20.9876 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 64 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/Llama-3.1-70B-Japanese-Instruct-2407 | 0.4359 | 0.5823 | 0.1812 | 0.1673 | 0.1655 | 0.8665 | 0.246 | 0.8368 | 0.752 | 0.1899 | 0.7083 | 0.0989 | 0.2384 | 0.8693 | 14.3716 | 0.907 | 0.9579 | 18.4078 | 0.888 | 0.1812 | 0.9181 | 0.6839 | 0.7514 | 0.9106 | 0.196 | 0.0085 | 0.6878 | 0.7879 | 0.8492 | 0.7083 | 0.8787 | 0.862 | 0.7708 | 0.246 | 0.5823 | 0.9699 | 0.3225 | 0.1352 | 0.0024 | 0.0102 | 0 | 0.0009 | 0.823 | 0.7949 | 10.767 | 0.8057 | 0.9029 | 11.2189 | 0.7465 | 0.6921 | 3.1649 | 23.7422 | 9.8925 | 0.0989 | 20.9876 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 64 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | SakanaAI/EvoLLM-JP-A-v1-7B | 0.3075 | 0.0783 | 0.0602 | 0.092 | 0.1096 | 0.4092 | 0.18 | 0.7626 | 0.5566 | 0.2941 | 0.7324 | 0.1075 | 0.2852 | 0.7996 | 7.8617 | 0.8364 | 0.9257 | 11.172 | 0.8301 | 0.0602 | 0.5496 | 0.4741 | 0.3583 | 0.3655 | 0.3017 | 0.1214 | 0.5329 | 0.6301 | 0.7875 | 0.7324 | 0.814 | 0.7803 | 0.3126 | 0.18 | 0.0783 | 0.2851 | 0.0977 | 0.2955 | 0.0064 | 0 | 0 | 0 | 0.4537 | 0.7262 | 5.4783 | 0.7133 | 0.8689 | 7.0128 | 0.6707 | 0.7015 | 2.2417 | 31.8526 | 10.7599 | 0.1075 | 26.5604 | MistralForCausalLM | bfloat16 | apache-2.0 | 7.242 | 11 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | SakanaAI/EvoLLM-JP-A-v1-7B | 0.4981 | 0.0783 | 0.4291 | 0.2025 | 0.4799 | 0.7665 | 0.662 | 0.8102 | 0.6447 | 0.411 | 0.8874 | 0.1075 | 0.4051 | 0.8354 | 9.5638 | 0.8798 | 0.9423 | 14.1638 | 0.8646 | 0.4291 | 0.8469 | 0.5718 | 0.6458 | 0.8686 | 0.4383 | 0.4262 | 0.6828 | 0.6806 | 0.6424 | 0.8874 | 0.8239 | 0.7917 | 0.5839 | 0.662 | 0.0783 | 0.2851 | 0.5337 | 0.3898 | 0.0079 | 0.3115 | 0.0354 | 0.0561 | 0.6015 | 0.7786 | 8.2927 | 0.7788 | 0.88 | 8.3322 | 0.7178 | 0.7015 | 2.2417 | 31.8526 | 10.7599 | 0.1075 | 26.5604 | MistralForCausalLM | bfloat16 | apache-2.0 | 7.242 | 11 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.2-3B-Instruct | 0.2042 | 0.0201 | 0 | 0.0345 | 0.1693 | 0.1582 | 0.008 | 0.6778 | 0.3061 | 0.1221 | 0.6936 | 0.0562 | 0.0882 | 0.794 | 7.0688 | 0.8148 | 0.906 | 9.086 | 0.7803 | 0 | 0.01 | 0.3851 | 0.5 | 0.2118 | 0.1375 | 0.0508 | 0.2342 | 0.1307 | 0.2807 | 0.6936 | 0.0555 | 0.0543 | 0.2526 | 0.008 | 0.0201 | 0.0843 | 0.2879 | 0.1406 | 0 | 0.0009 | 0 | 0 | 0.1716 | 0.6874 | 3.9743 | 0.6419 | 0.8229 | 5.2307 | 0.4743 | 0.6484 | 1.9187 | 16.1297 | 5.6101 | 0.0562 | 14.0826 | LlamaForCausalLM | bfloat16 | llama3.2 | 3.213 | 638 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.2-3B-Instruct | 0.4111 | 0.0201 | 0.403 | 0.1037 | 0.4761 | 0.5994 | 0.61 | 0.7681 | 0.3565 | 0.2786 | 0.8506 | 0.0562 | 0.2529 | 0.8285 | 9.1433 | 0.86 | 0.9366 | 13.7136 | 0.8465 | 0.403 | 0.6528 | 0.3477 | 0.4958 | 0.7551 | 0.3175 | 0.3973 | 0.3414 | 0.4059 | 0.1916 | 0.8506 | 0.1652 | 0.1046 | 0.3903 | 0.61 | 0.0201 | 0.0843 | 0.5548 | 0.2655 | 0.0153 | 0.1642 | 0.0265 | 0.022 | 0.2904 | 0.7271 | 7.1521 | 0.6776 | 0.8772 | 8.3252 | 0.6881 | 0.6484 | 1.9187 | 16.1297 | 5.6101 | 0.0562 | 14.0826 | LlamaForCausalLM | bfloat16 | llama3.2 | 3.213 | 638 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | rinna/llama-3-youko-8b-instruct | 0.5444 | 0.2811 | 0.4998 | 0.2651 | 0.4879 | 0.8302 | 0.526 | 0.8517 | 0.6812 | 0.5522 | 0.9004 | 0.113 | 0.6523 | 0.8605 | 11.8261 | 0.9056 | 0.9515 | 15.745 | 0.8797 | 0.4998 | 0.876 | 0.5661 | 0.6806 | 0.9169 | 0.485 | 0.4383 | 0.8069 | 0.7285 | 0.6237 | 0.9004 | 0.8224 | 0.7804 | 0.6976 | 0.526 | 0.2811 | 0.739 | 0.5375 | 0.5192 | 0.034 | 0.3175 | 0.0619 | 0.0749 | 0.8374 | 0.847 | 13.2105 | 0.8486 | 0.9147 | 12.3221 | 0.773 | 0.7051 | 3.0857 | 28.1083 | 11.3046 | 0.113 | 24.4835 | LlamaForCausalLM | bfloat16 | llama3 | 8.03 | 11 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | rinna/llama-3-youko-8b-instruct | 0.3585 | 0.2811 | 0.1149 | 0.1395 | 0.0168 | 0.6881 | 0.006 | 0.8366 | 0.5024 | 0.3991 | 0.8462 | 0.113 | 0.5068 | 0.8499 | 10.1041 | 0.8963 | 0.9469 | 13.8825 | 0.8722 | 0.1149 | 0.7733 | 0.4626 | 0.5861 | 0.7936 | 0.2486 | 0 | 0.4022 | 0.6824 | 0.3785 | 0.8462 | 0.4037 | 0.4642 | 0.4975 | 0.006 | 0.2811 | 0.739 | 0.0337 | 0.442 | 0 | 0.0008 | 0 | 0.0059 | 0.6909 | 0.8191 | 9.923 | 0.8214 | 0.9051 | 10.3137 | 0.7563 | 0.7051 | 3.0857 | 28.1083 | 11.3046 | 0.113 | 24.4835 | LlamaForCausalLM | bfloat16 | llama3 | 8.03 | 11 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | rinna/llama-3-youko-70b-instruct | 0.5358 | 0.4498 | 0.3448 | 0.1644 | 0.446 | 0.7631 | 0.808 | 0.8454 | 0.636 | 0.4595 | 0.8891 | 0.0871 | 0.4512 | 0.8558 | 12.2981 | 0.904 | 0.9512 | 17.0644 | 0.8793 | 0.3448 | 0.7911 | 0.5086 | 0.675 | 0.8445 | 0.4467 | 0.4581 | 0.4573 | 0.7677 | 0.7713 | 0.8891 | 0.86 | 0.8183 | 0.6538 | 0.808 | 0.4498 | 0.7028 | 0.434 | 0.4806 | 0 | 0.0224 | 0.0177 | 0.005 | 0.777 | 0.8197 | 10.0643 | 0.8368 | 0.9068 | 10.9401 | 0.7617 | 0.671 | 2.8969 | 21.0754 | 8.7164 | 0.0871 | 18.5691 | LlamaForCausalLM | bfloat16 | llama3 | 70.554 | 0 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | rinna/llama-3-youko-70b-instruct | 0.6505 | 0.4498 | 0.5876 | 0.2869 | 0.7458 | 0.8863 | 0.934 | 0.8575 | 0.7479 | 0.6521 | 0.9203 | 0.0871 | 0.7374 | 0.8677 | 13.7925 | 0.9099 | 0.9584 | 18.5638 | 0.8905 | 0.5876 | 0.9018 | 0.6724 | 0.7764 | 0.9517 | 0.6348 | 0.7151 | 0.7609 | 0.7942 | 0.7355 | 0.9203 | 0.8946 | 0.8666 | 0.8054 | 0.934 | 0.4498 | 0.7028 | 0.7765 | 0.584 | 0.0542 | 0.3116 | 0.1416 | 0.0677 | 0.8596 | 0.8471 | 13.5325 | 0.8505 | 0.9163 | 12.5548 | 0.7793 | 0.671 | 2.8969 | 21.0754 | 8.7164 | 0.0871 | 18.5691 | LlamaForCausalLM | bfloat16 | llama3 | 70.554 | 0 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🤝 : base merges and moerges | ryota39/Tora-12B | 0.5036 | 0 | 0.5251 | 0.1908 | 0.4971 | 0.7748 | 0.684 | 0.8132 | 0.6189 | 0.4522 | 0.8868 | 0.0972 | 0.4565 | 0.8568 | 11.6997 | 0.9014 | 0.9445 | 14.3901 | 0.8681 | 0.5251 | 0.8434 | 0.5259 | 0.7097 | 0.8794 | 0.4327 | 0.4685 | 0.5362 | 0.75 | 0.5728 | 0.8868 | 0.8288 | 0.8246 | 0.6016 | 0.684 | 0 | 0 | 0.5256 | 0.4674 | 0.0163 | 0.2399 | 0.0133 | 0.0305 | 0.6541 | 0.7675 | 6.7435 | 0.7813 | 0.8824 | 8.4109 | 0.702 | 0.6902 | 2.606 | 27.0004 | 9.7306 | 0.0972 | 22.9946 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 1 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🤝 : base merges and moerges | ryota39/Tora-12B | 0.3728 | 0 | 0.0171 | 0.1148 | 0.2861 | 0.6261 | 0.56 | 0.798 | 0.5746 | 0.204 | 0.8228 | 0.0972 | 0.1097 | 0.8398 | 10.3864 | 0.8934 | 0.9431 | 13.1423 | 0.8644 | 0.0171 | 0.763 | 0.5201 | 0.7056 | 0.7417 | 0.2828 | 0.2824 | 0.4552 | 0.7544 | 0.4378 | 0.8228 | 0.8389 | 0.8398 | 0.3734 | 0.56 | 0 | 0 | 0.2898 | 0.2197 | 0.0025 | 0.0008 | 0.023 | 0 | 0.5474 | 0.734 | 5.0103 | 0.7377 | 0.8804 | 7.659 | 0.6964 | 0.6902 | 2.606 | 27.0004 | 9.7306 | 0.0972 | 22.9946 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 1 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | Qwen/Qwen2.5-Coder-32B-Instruct | 0.6196 | 0.6847 | 0.5416 | 0.2769 | 0.6754 | 0.8225 | 0.896 | 0.8349 | 0.7292 | 0.4215 | 0.858 | 0.0744 | 0.3762 | 0.8533 | 11.6603 | 0.8952 | 0.9501 | 15.695 | 0.8762 | 0.5416 | 0.8437 | 0.6523 | 0.8708 | 0.908 | 0.5112 | 0.6315 | 0.7502 | 0.6818 | 0.6911 | 0.858 | 0.8716 | 0.845 | 0.716 | 0.896 | 0.6847 | 0.755 | 0.7193 | 0.3772 | 0.0529 | 0.3638 | 0.1239 | 0.0749 | 0.769 | 0.8104 | 9.7551 | 0.8238 | 0.8991 | 10.3091 | 0.7445 | 0.6725 | 1.8374 | 21.7121 | 7.4326 | 0.0744 | 18.911 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 958 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | Qwen/Qwen2.5-Coder-32B-Instruct | 0.4649 | 0.6847 | 0.2414 | 0.1542 | 0.5666 | 0.7859 | 0.332 | 0.8067 | 0.4242 | 0.2567 | 0.7872 | 0.0744 | 0.2702 | 0.825 | 8.635 | 0.8742 | 0.9437 | 13.3534 | 0.8633 | 0.2414 | 0.8236 | 0.4598 | 0.7194 | 0.8758 | 0.2579 | 0.5055 | 0.3977 | 0.1957 | 0.3485 | 0.7872 | 0.8391 | 0.8255 | 0.6582 | 0.332 | 0.6847 | 0.755 | 0.6278 | 0.2419 | 0.0231 | 0.0118 | 0.0708 | 0.0048 | 0.6604 | 0.7637 | 7.3008 | 0.7796 | 0.8868 | 8.9065 | 0.7098 | 0.6725 | 1.8374 | 21.7121 | 7.4326 | 0.0744 | 18.911 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 958 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/EZO-Common-9B-gemma-2-it | 0.3855 | 0.006 | 0.0125 | 0.0537 | 0.3964 | 0.6565 | 0.56 | 0.781 | 0.5944 | 0.284 | 0.7735 | 0.1228 | 0.2447 | 0.8243 | 9.2281 | 0.8642 | 0.9336 | 12.3934 | 0.8402 | 0.0125 | 0.7725 | 0.5259 | 0.5625 | 0.7498 | 0.313 | 0.4253 | 0.479 | 0.714 | 0.6905 | 0.7735 | 0.7556 | 0.7149 | 0.4472 | 0.56 | 0.006 | 0.0141 | 0.3675 | 0.2943 | 0.0064 | 0.0003 | 0 | 0 | 0.2616 | 0.7432 | 6.4405 | 0.7328 | 0.8779 | 8.6208 | 0.6867 | 0.714 | 2.8145 | 35.9536 | 12.2713 | 0.1228 | 29.3246 | Gemma2ForCausalLM | bfloat16 | gemma | 9.242 | 31 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/EZO-Common-9B-gemma-2-it | 0.5275 | 0.006 | 0.4616 | 0.2047 | 0.609 | 0.7961 | 0.764 | 0.8381 | 0.6519 | 0.4585 | 0.89 | 0.1228 | 0.4946 | 0.86 | 11.9103 | 0.8978 | 0.9452 | 16.075 | 0.8646 | 0.4616 | 0.8304 | 0.4943 | 0.6778 | 0.8633 | 0.4892 | 0.5498 | 0.7005 | 0.6711 | 0.7161 | 0.89 | 0.8535 | 0.8079 | 0.6945 | 0.764 | 0.006 | 0.0141 | 0.6682 | 0.3916 | 0.0131 | 0.2754 | 0.0796 | 0.0708 | 0.5846 | 0.8216 | 9.7229 | 0.8338 | 0.904 | 10.8852 | 0.7563 | 0.714 | 2.8145 | 35.9536 | 12.2713 | 0.1228 | 29.3246 | Gemma2ForCausalLM | bfloat16 | gemma | 9.242 | 31 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/EZO-Humanities-9B-gemma-2-it | 0.5294 | 0.0301 | 0.4571 | 0.1934 | 0.6078 | 0.8 | 0.758 | 0.8364 | 0.6586 | 0.4751 | 0.8835 | 0.1236 | 0.5006 | 0.8569 | 11.5508 | 0.8935 | 0.9446 | 15.2615 | 0.8634 | 0.4571 | 0.8447 | 0.5029 | 0.7222 | 0.8624 | 0.4724 | 0.5484 | 0.6927 | 0.6597 | 0.7154 | 0.8835 | 0.8472 | 0.8046 | 0.6929 | 0.758 | 0.0301 | 0.0562 | 0.6671 | 0.4522 | 0.017 | 0.2816 | 0.0354 | 0.0762 | 0.5568 | 0.8215 | 9.5432 | 0.8344 | 0.9036 | 10.7545 | 0.7545 | 0.7143 | 2.6119 | 36.4581 | 12.3595 | 0.1236 | 29.6341 | Gemma2ForCausalLM | bfloat16 | gemma | 9.242 | 21 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/EZO-Humanities-9B-gemma-2-it | 0.3874 | 0.0301 | 0.0083 | 0.0518 | 0.4085 | 0.6664 | 0.572 | 0.7624 | 0.6096 | 0.2605 | 0.7687 | 0.1236 | 0.2273 | 0.8157 | 9.1465 | 0.8474 | 0.9309 | 12.2476 | 0.8339 | 0.0083 | 0.7861 | 0.5259 | 0.5611 | 0.7516 | 0.2606 | 0.4211 | 0.537 | 0.7172 | 0.7069 | 0.7687 | 0.7466 | 0.7017 | 0.4616 | 0.572 | 0.0301 | 0.0562 | 0.396 | 0.2936 | 0.0014 | 0 | 0 | 0 | 0.2577 | 0.7292 | 6.0949 | 0.7 | 0.8718 | 8.3038 | 0.6684 | 0.7143 | 2.6119 | 36.4581 | 12.3595 | 0.1236 | 29.6341 | Gemma2ForCausalLM | bfloat16 | gemma | 9.242 | 21 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/EZO-Qwen2.5-32B-Instruct | 0.6217 | 0.2189 | 0.5779 | 0.2486 | 0.7633 | 0.8982 | 0.928 | 0.8501 | 0.8095 | 0.5476 | 0.9029 | 0.0942 | 0.5622 | 0.8649 | 13.1459 | 0.9085 | 0.9545 | 17.3022 | 0.8858 | 0.5779 | 0.9123 | 0.6925 | 0.8542 | 0.9625 | 0.5803 | 0.7343 | 0.8673 | 0.7784 | 0.8551 | 0.9029 | 0.9014 | 0.8823 | 0.8197 | 0.928 | 0.2189 | 0.2691 | 0.7923 | 0.5004 | 0.0025 | 0.3567 | 0 | 0.0806 | 0.8032 | 0.8248 | 10.4495 | 0.8444 | 0.9024 | 10.5557 | 0.7616 | 0.6932 | 2.5841 | 26.0188 | 9.4165 | 0.0942 | 22.2414 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 6 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/EZO-Qwen2.5-32B-Instruct | 0.5255 | 0.2189 | 0.164 | 0.1459 | 0.6883 | 0.8671 | 0.736 | 0.8388 | 0.7692 | 0.3722 | 0.886 | 0.0942 | 0.3589 | 0.8446 | 10.8444 | 0.9003 | 0.9508 | 15.2871 | 0.8811 | 0.164 | 0.8938 | 0.6839 | 0.7778 | 0.9357 | 0.3282 | 0.6408 | 0.8073 | 0.7614 | 0.8155 | 0.886 | 0.8954 | 0.8736 | 0.7718 | 0.736 | 0.2189 | 0.2691 | 0.7358 | 0.4294 | 0.0083 | 0.0044 | 0 | 0.005 | 0.7118 | 0.7931 | 8.3381 | 0.8245 | 0.8967 | 9.4364 | 0.7493 | 0.6932 | 2.5841 | 26.0188 | 9.4165 | 0.0942 | 22.2414 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 6 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | karakuri-ai/karakuri-lm-8x7b-chat-v0.1 | 0.579 | 0.4398 | 0.5106 | 0.2509 | 0.553 | 0.7859 | 0.738 | 0.8295 | 0.6669 | 0.5578 | 0.906 | 0.1308 | 0.6857 | 0.859 | 12.5191 | 0.9048 | 0.948 | 15.3313 | 0.8715 | 0.5106 | 0.8835 | 0.5086 | 0.6569 | 0.8391 | 0.4662 | 0.5165 | 0.8003 | 0.7633 | 0.6052 | 0.906 | 0.8325 | 0.7804 | 0.6349 | 0.738 | 0.4398 | 0.9157 | 0.5895 | 0.5214 | 0.0142 | 0.2769 | 0.115 | 0.0787 | 0.7696 | 0.8112 | 9.7984 | 0.8191 | 0.8947 | 9.9615 | 0.7225 | 0.714 | 3.0687 | 34.8302 | 13.0565 | 0.1308 | 29.23 | MixtralForCausalLM | bfloat16 | apache-2.0 | 46.703 | 20 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | karakuri-ai/karakuri-lm-8x7b-chat-v0.1 | 0.4643 | 0.4398 | 0.1461 | 0.1553 | 0.4483 | 0.6098 | 0.586 | 0.8132 | 0.5708 | 0.4013 | 0.8056 | 0.1308 | 0.5219 | 0.8413 | 9.5173 | 0.8921 | 0.9451 | 12.9912 | 0.8695 | 0.1461 | 0.7282 | 0.4684 | 0.6806 | 0.7596 | 0.2338 | 0.3988 | 0.3652 | 0.6818 | 0.6582 | 0.8056 | 0.8181 | 0.7938 | 0.3416 | 0.586 | 0.4398 | 0.9157 | 0.4979 | 0.4483 | 0.0208 | 0.0054 | 0.0354 | 0 | 0.7151 | 0.7783 | 7.1414 | 0.7944 | 0.8831 | 8.0624 | 0.6967 | 0.714 | 3.0687 | 34.8302 | 13.0565 | 0.1308 | 29.23 | MixtralForCausalLM | bfloat16 | apache-2.0 | 46.703 | 20 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.2-1B-Instruct | 0.3059 | 0.006 | 0.3175 | 0.0299 | 0.3193 | 0.4104 | 0.448 | 0.6088 | 0.3826 | 0.2129 | 0.5948 | 0.0344 | 0.1483 | 0.7505 | 6.5622 | 0.7029 | 0.913 | 10.3897 | 0.7881 | 0.3175 | 0.4765 | 0.3391 | 0.5056 | 0.4683 | 0.3055 | 0.2801 | 0.5308 | 0.3163 | 0.2212 | 0.5948 | 0 | 0 | 0.2865 | 0.448 | 0.006 | 0.0321 | 0.3584 | 0.1849 | 0 | 0.0521 | 0 | 0.0123 | 0.0849 | 0.5608 | 1.4338 | 0.3883 | 0.8311 | 5.383 | 0.5558 | 0.6012 | 1.3537 | 12.4088 | 3.448 | 0.0344 | 9.9124 | LlamaForCausalLM | bfloat16 | llama3.2 | 1.236 | 573 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | meta-llama/Llama-3.2-1B-Instruct | 0.1052 | 0.006 | 0 | 0.0318 | 0.0014 | 0.189 | 0 | 0.4682 | 0 | 0.1129 | 0.3138 | 0.0344 | 0.0727 | 0.6638 | 1.5895 | 0.5399 | 0.7425 | 0.1903 | 0.4989 | 0 | 0.4051 | 0 | 0 | 0.0179 | 0.1733 | 0 | 0 | 0 | 0 | 0.3138 | -0.0837 | -0.1146 | 0.1441 | 0 | 0.006 | 0.0321 | 0.0028 | 0.0929 | 0 | 0 | 0 | 0 | 0.1589 | 0.5919 | 1.3013 | 0.4303 | 0.7328 | 0.3692 | 0.4039 | 0.6012 | 1.3537 | 12.4088 | 3.448 | 0.0344 | 9.9124 | LlamaForCausalLM | bfloat16 | llama3.2 | 1.236 | 573 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/Llama-3.1-70B-EZO-1.1-it | 0.4493 | 0.4759 | 0.1461 | 0.167 | 0.1811 | 0.87 | 0.374 | 0.8432 | 0.7253 | 0.3434 | 0.7049 | 0.1112 | 0.4223 | 0.8652 | 13.0595 | 0.9109 | 0.9559 | 17.4464 | 0.8853 | 0.1461 | 0.9236 | 0.6351 | 0.7486 | 0.9178 | 0.3892 | 0.0873 | 0.6265 | 0.7822 | 0.834 | 0.7049 | 0.8848 | 0.8564 | 0.7687 | 0.374 | 0.4759 | 0.7671 | 0.275 | 0.2188 | 0 | 0 | 0.0088 | 0 | 0.8263 | 0.8017 | 9.9795 | 0.8247 | 0.9003 | 10.2301 | 0.7519 | 0.7018 | 3.1569 | 26.8732 | 11.1146 | 0.1112 | 23.5595 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 11 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | AXCXEPT/Llama-3.1-70B-EZO-1.1-it | 0.6537 | 0.4759 | 0.5341 | 0.2846 | 0.7542 | 0.9071 | 0.902 | 0.8441 | 0.7667 | 0.7038 | 0.9073 | 0.1112 | 0.8429 | 0.8606 | 13.1933 | 0.8997 | 0.9486 | 17.0482 | 0.8675 | 0.5341 | 0.9354 | 0.6178 | 0.7847 | 0.9607 | 0.6168 | 0.7201 | 0.7896 | 0.8056 | 0.836 | 0.9073 | 0.8743 | 0.8479 | 0.8253 | 0.902 | 0.4759 | 0.7671 | 0.7883 | 0.6516 | 0.0263 | 0.3609 | 0.1239 | 0.0555 | 0.8567 | 0.8353 | 13.154 | 0.8452 | 0.9071 | 10.9842 | 0.7638 | 0.7018 | 3.1569 | 26.8732 | 11.1146 | 0.1112 | 23.5595 | LlamaForCausalLM | bfloat16 | llama3.1 | 70.554 | 11 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | karakuri-ai/karakuri-lm-8x7b-instruct-v0.1 | 0.5485 | 0.0502 | 0.5043 | 0.2695 | 0.565 | 0.7869 | 0.72 | 0.8283 | 0.6865 | 0.596 | 0.9122 | 0.115 | 0.6905 | 0.8604 | 11.8326 | 0.9048 | 0.9486 | 15.6642 | 0.8721 | 0.5043 | 0.8442 | 0.5172 | 0.7292 | 0.8919 | 0.5456 | 0.5253 | 0.7576 | 0.7664 | 0.6621 | 0.9122 | 0.8363 | 0.7498 | 0.6245 | 0.72 | 0.0502 | 0.1245 | 0.6048 | 0.552 | 0.0222 | 0.3319 | 0.1239 | 0.0908 | 0.7789 | 0.8137 | 10.0974 | 0.8189 | 0.8917 | 9.9485 | 0.7173 | 0.6983 | 2.8331 | 30.452 | 11.4983 | 0.115 | 25.7002 | MixtralForCausalLM | bfloat16 | apache-2.0 | 46.704 | 19 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | karakuri-ai/karakuri-lm-8x7b-instruct-v0.1 | 0.4249 | 0.0502 | 0.2769 | 0.1339 | 0.4183 | 0.6429 | 0.5 | 0.8189 | 0.5858 | 0.3979 | 0.7338 | 0.115 | 0.4678 | 0.8427 | 10.4168 | 0.8933 | 0.9465 | 12.8796 | 0.8725 | 0.2769 | 0.8818 | 0.5144 | 0.6944 | 0.7364 | 0.2846 | 0.3812 | 0.3726 | 0.6723 | 0.6753 | 0.7338 | 0.8244 | 0.8011 | 0.3106 | 0.5 | 0.0502 | 0.1245 | 0.4554 | 0.4413 | 0.015 | 0.0011 | 0.0088 | 0 | 0.6445 | 0.7794 | 7.5151 | 0.7969 | 0.8851 | 8.337 | 0.7127 | 0.6983 | 2.8331 | 30.452 | 11.4983 | 0.115 | 25.7002 | MixtralForCausalLM | bfloat16 | apache-2.0 | 46.704 | 19 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | karakuri-ai/karakuri-lm-70b-chat-v0.1 | 0.52 | 0 | 0.4993 | 0.2285 | 0.516 | 0.8123 | 0.636 | 0.8387 | 0.5423 | 0.6236 | 0.9103 | 0.1135 | 0.7876 | 0.8589 | 12.8743 | 0.9023 | 0.9513 | 15.9804 | 0.8781 | 0.4993 | 0.8702 | 0.4138 | 0.7139 | 0.8856 | 0.5177 | 0.4679 | 0.4491 | 0.5871 | 0.5476 | 0.9103 | 0.6075 | 0.5192 | 0.6811 | 0.636 | 0 | 0.002 | 0.564 | 0.5656 | 0.0013 | 0.274 | 0.0354 | 0.0384 | 0.7933 | 0.8237 | 11.0849 | 0.8292 | 0.9051 | 11.4481 | 0.745 | 0.7048 | 2.0013 | 32.715 | 11.3497 | 0.1135 | 27.593 | LlamaForCausalLM | bfloat16 | other | 69.196 | 23 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | karakuri-ai/karakuri-lm-70b-chat-v0.1 | 0.3458 | 0 | 0.0708 | 0.1421 | 0.3197 | 0.4886 | 0.032 | 0.8226 | 0.5578 | 0.4086 | 0.848 | 0.1135 | 0.5177 | 0.8451 | 11.1751 | 0.8928 | 0.9485 | 14.225 | 0.8742 | 0.0708 | 0.7525 | 0.3649 | 0.5069 | 0.3342 | 0.2332 | 0.2231 | 0.5616 | 0.7102 | 0.645 | 0.848 | 0.202 | 0.4942 | 0.3792 | 0.032 | 0 | 0.002 | 0.4164 | 0.475 | 0.0025 | 0.0023 | 0 | 0 | 0.7056 | 0.7811 | 7.3019 | 0.7962 | 0.8954 | 9.1576 | 0.727 | 0.7048 | 2.0013 | 32.715 | 11.3497 | 0.1135 | 27.593 | LlamaForCausalLM | bfloat16 | other | 69.196 | 23 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | anthracite-org/magnum-v2.5-12b-kto | 0.5803 | 0.4659 | 0.5056 | 0.2575 | 0.5765 | 0.8432 | 0.742 | 0.8433 | 0.6803 | 0.4789 | 0.8929 | 0.0973 | 0.5346 | 0.8523 | 11.0049 | 0.9043 | 0.9507 | 15.6039 | 0.8798 | 0.5056 | 0.892 | 0.5546 | 0.75 | 0.9187 | 0.4296 | 0.5247 | 0.751 | 0.7557 | 0.5904 | 0.8929 | 0.8691 | 0.8385 | 0.7188 | 0.742 | 0.4659 | 0.8835 | 0.6283 | 0.4727 | 0.0461 | 0.3907 | 0.0265 | 0.0565 | 0.7675 | 0.8196 | 10.223 | 0.8337 | 0.901 | 9.7323 | 0.7555 | 0.6934 | 2.7577 | 28.2469 | 9.7173 | 0.0973 | 24.3763 | MistralForCausalLM | float16 | apache-2.0 | 12.248 | 40 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | anthracite-org/magnum-v2.5-12b-kto | 0.431 | 0.4659 | 0.2191 | 0.1406 | 0.3713 | 0.6504 | 0.256 | 0.7663 | 0.6381 | 0.3328 | 0.8032 | 0.0973 | 0.3706 | 0.8273 | 9.1971 | 0.8933 | 0.8894 | 13.3192 | 0.7395 | 0.2191 | 0.7866 | 0.5115 | 0.7389 | 0.6711 | 0.26 | 0.316 | 0.3977 | 0.7462 | 0.7964 | 0.8032 | 0.7715 | 0.7927 | 0.4936 | 0.256 | 0.4659 | 0.8835 | 0.4266 | 0.3679 | 0.0233 | 0.0011 | 0 | 0 | 0.6786 | 0.7848 | 7.6412 | 0.8147 | 0.8558 | 8.2921 | 0.6179 | 0.6934 | 2.7577 | 28.2469 | 9.7173 | 0.0973 | 24.3763 | MistralForCausalLM | float16 | apache-2.0 | 12.248 | 40 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | mistralai/Mistral-Nemo-Instruct-2407 | 0.5006 | 0.4438 | 0.3349 | 0.1314 | 0.505 | 0.6974 | 0.664 | 0.8173 | 0.6304 | 0.3476 | 0.8248 | 0.1102 | 0.4349 | 0.8356 | 9.7971 | 0.8909 | 0.9465 | 13.9763 | 0.8721 | 0.3349 | 0.7142 | 0.4339 | 0.7028 | 0.8284 | 0.2038 | 0.447 | 0.5255 | 0.7247 | 0.765 | 0.8248 | 0.8271 | 0.7887 | 0.5497 | 0.664 | 0.4438 | 0.8594 | 0.563 | 0.4041 | 0.0058 | 0 | 0.0265 | 0 | 0.6245 | 0.7781 | 8.1055 | 0.7868 | 0.8889 | 8.9971 | 0.7192 | 0.702 | 2.6477 | 30.1236 | 11.0167 | 0.1102 | 25.8189 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 1,242 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | mistralai/Mistral-Nemo-Instruct-2407 | 0.582 | 0.4438 | 0.5297 | 0.251 | 0.5899 | 0.8386 | 0.754 | 0.8382 | 0.6393 | 0.5034 | 0.904 | 0.1102 | 0.5181 | 0.8563 | 11.6408 | 0.902 | 0.9503 | 15.776 | 0.8793 | 0.5297 | 0.8848 | 0.408 | 0.7458 | 0.9205 | 0.5032 | 0.5397 | 0.689 | 0.6907 | 0.6629 | 0.904 | 0.8645 | 0.821 | 0.7105 | 0.754 | 0.4438 | 0.8594 | 0.6401 | 0.489 | 0.0358 | 0.3292 | 0.0442 | 0.0898 | 0.7562 | 0.8181 | 9.9231 | 0.8269 | 0.8978 | 9.7514 | 0.7447 | 0.702 | 2.6477 | 30.1236 | 11.0167 | 0.1102 | 25.8189 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 1,242 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/Mistral-Nemo-Japanese-Instruct-2408 | 0.5436 | 0.0181 | 0.5247 | 0.2218 | 0.5626 | 0.8079 | 0.768 | 0.8355 | 0.7091 | 0.52 | 0.9142 | 0.0977 | 0.5393 | 0.8645 | 12.7846 | 0.9087 | 0.9497 | 15.6384 | 0.8786 | 0.5247 | 0.8196 | 0.592 | 0.8375 | 0.9133 | 0.5091 | 0.5346 | 0.7194 | 0.7696 | 0.6272 | 0.9142 | 0.8762 | 0.8532 | 0.6908 | 0.768 | 0.0181 | 0.0422 | 0.5905 | 0.5116 | 0.0209 | 0.2932 | 0.0442 | 0.0512 | 0.6994 | 0.7948 | 7.9257 | 0.8172 | 0.893 | 9.2525 | 0.7373 | 0.6934 | 2.6755 | 26.798 | 9.7702 | 0.0977 | 23.1335 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 23 | main | 4 | True | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | cyberagent/Mistral-Nemo-Japanese-Instruct-2408 | 0.4725 | 0.0181 | 0.2617 | 0.1413 | 0.4206 | 0.7856 | 0.684 | 0.8286 | 0.7287 | 0.3388 | 0.892 | 0.0977 | 0.2819 | 0.85 | 12.1913 | 0.9051 | 0.9484 | 14.8813 | 0.8773 | 0.2617 | 0.8312 | 0.6063 | 0.7847 | 0.8937 | 0.3612 | 0.5095 | 0.804 | 0.7696 | 0.6787 | 0.892 | 0.8913 | 0.8646 | 0.632 | 0.684 | 0.0181 | 0.0422 | 0.3316 | 0.3733 | 0.0095 | 0.0039 | 0.0265 | 0.0092 | 0.6575 | 0.7695 | 6.2908 | 0.8015 | 0.8889 | 8.8337 | 0.7306 | 0.6934 | 2.6755 | 26.798 | 9.7702 | 0.0977 | 23.1335 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 23 | main | 0 | True | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | Elizezen/Himeyuri-v0.1-12B | 0.5838 | 0.4719 | 0.5117 | 0.2548 | 0.5801 | 0.8502 | 0.742 | 0.844 | 0.6753 | 0.4922 | 0.8967 | 0.1033 | 0.5394 | 0.8537 | 11.1719 | 0.9055 | 0.9511 | 16.1927 | 0.88 | 0.5117 | 0.8945 | 0.5345 | 0.7472 | 0.9267 | 0.4574 | 0.5301 | 0.7564 | 0.7519 | 0.5864 | 0.8967 | 0.8673 | 0.8357 | 0.7295 | 0.742 | 0.4719 | 0.8735 | 0.6301 | 0.4798 | 0.0395 | 0.3845 | 0.0177 | 0.0627 | 0.7698 | 0.8222 | 10.426 | 0.8357 | 0.9015 | 9.9963 | 0.7547 | 0.6968 | 2.7793 | 28.6041 | 10.3203 | 0.1033 | 24.6055 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 10 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | Elizezen/Himeyuri-v0.1-12B | 0.4116 | 0.4719 | 0.2167 | 0.1393 | 0.2734 | 0.5405 | 0.268 | 0.7667 | 0.6341 | 0.3081 | 0.8052 | 0.1033 | 0.3542 | 0.8028 | 9.2799 | 0.8537 | 0.9143 | 13.4908 | 0.777 | 0.2167 | 0.5168 | 0.5086 | 0.7264 | 0.6345 | 0.2323 | 0.2866 | 0.3895 | 0.7487 | 0.797 | 0.8052 | 0.8328 | 0.7993 | 0.4701 | 0.268 | 0.4719 | 0.8735 | 0.2601 | 0.3379 | 0.0106 | 0.0019 | 0 | 0.0014 | 0.6827 | 0.7734 | 7.6856 | 0.7975 | 0.8653 | 8.3046 | 0.6386 | 0.6968 | 2.7793 | 28.6041 | 10.3203 | 0.1033 | 24.6055 | MistralForCausalLM | bfloat16 | apache-2.0 | 12.248 | 10 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | spow12/ChatWaifu_v2.0_22B | 0.5755 | 0.4498 | 0.489 | 0.2754 | 0.5818 | 0.7434 | 0.774 | 0.8296 | 0.7435 | 0.425 | 0.9072 | 0.1113 | 0.399 | 0.849 | 11.2044 | 0.8882 | 0.9473 | 15.5331 | 0.872 | 0.489 | 0.8069 | 0.6236 | 0.8236 | 0.7757 | 0.492 | 0.5027 | 0.7576 | 0.7399 | 0.7727 | 0.9072 | 0.8752 | 0.829 | 0.6476 | 0.774 | 0.4498 | 0.9137 | 0.6609 | 0.3841 | 0.002 | 0.5037 | 0.1062 | 0.079 | 0.6864 | 0.8089 | 9.9046 | 0.8128 | 0.8999 | 10.5372 | 0.7454 | 0.7051 | 2.6042 | 32.0946 | 11.1286 | 0.1113 | 27.1554 | MistralForCausalLM | bfloat16 | cc-by-nc-4.0 | 22.247 | 6 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | spow12/ChatWaifu_v2.0_22B | 0.4908 | 0.4498 | 0.0584 | 0.1221 | 0.5491 | 0.6809 | 0.744 | 0.8135 | 0.6798 | 0.3484 | 0.842 | 0.1113 | 0.3116 | 0.826 | 9.8798 | 0.875 | 0.9448 | 14.384 | 0.8682 | 0.0584 | 0.7179 | 0.546 | 0.7611 | 0.7811 | 0.4214 | 0.4708 | 0.6742 | 0.7096 | 0.7079 | 0.842 | 0.8597 | 0.838 | 0.5437 | 0.744 | 0.4498 | 0.9137 | 0.6274 | 0.3123 | 0.002 | 0.004 | 0.0265 | 0.0103 | 0.5676 | 0.7695 | 6.8754 | 0.7822 | 0.8916 | 9.4839 | 0.7285 | 0.7051 | 2.6042 | 32.0946 | 11.1286 | 0.1113 | 27.1554 | MistralForCausalLM | bfloat16 | cc-by-nc-4.0 | 22.247 | 6 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | spow12/ChatWaifu_12B_v2.0 | 0.4009 | 0.4759 | 0.2443 | 0.1112 | 0.2387 | 0.4261 | 0.468 | 0.5958 | 0.6377 | 0.2964 | 0.825 | 0.0904 | 0.3011 | 0.7236 | 10.3014 | 0.6 | 0.8686 | 14.3307 | 0.521 | 0.2443 | 0.0098 | 0.4569 | 0.7431 | 0.7256 | 0.2627 | 0.3335 | 0.4528 | 0.7569 | 0.7788 | 0.825 | 0.8504 | 0.8101 | 0.5429 | 0.468 | 0.4759 | 0.8876 | 0.1439 | 0.3254 | 0.0025 | 0.001 | 0 | 0.0027 | 0.55 | 0.7188 | 8.0496 | 0.6563 | 0.8593 | 8.5589 | 0.6056 | 0.677 | 2.6975 | 24.4803 | 9.0565 | 0.0904 | 19.7651 | MistralForCausalLM | bfloat16 | cc-by-nc-4.0 | 12.248 | 11 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | spow12/ChatWaifu_12B_v2.0 | 0.5865 | 0.4759 | 0.5181 | 0.2591 | 0.5898 | 0.8507 | 0.752 | 0.8421 | 0.6773 | 0.4988 | 0.8976 | 0.0904 | 0.5378 | 0.8578 | 11.8316 | 0.9061 | 0.9515 | 15.9457 | 0.8809 | 0.5181 | 0.893 | 0.4885 | 0.75 | 0.9303 | 0.4831 | 0.5414 | 0.756 | 0.7626 | 0.6292 | 0.8976 | 0.8722 | 0.8432 | 0.7289 | 0.752 | 0.4759 | 0.8876 | 0.6382 | 0.4755 | 0.0492 | 0.358 | 0.0442 | 0.0699 | 0.7743 | 0.8223 | 10.3095 | 0.832 | 0.9009 | 9.937 | 0.7493 | 0.677 | 2.6975 | 24.4803 | 9.0565 | 0.0904 | 19.7651 | MistralForCausalLM | bfloat16 | cc-by-nc-4.0 | 12.248 | 11 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | augmxnt/shisa-gamma-7b-v1 | 0.4659 | 0.2369 | 0.0387 | 0.1173 | 0.3918 | 0.7352 | 0.476 | 0.8017 | 0.8178 | 0.4707 | 0.9121 | 0.1262 | 0.6101 | 0.8455 | 10.7929 | 0.8931 | 0.9422 | 13.3796 | 0.8623 | 0.0387 | 0.7751 | 0.6207 | 0.9486 | 0.8606 | 0.3184 | 0.3668 | 0.878 | 0.7854 | 0.8563 | 0.9121 | 0.9069 | 0.873 | 0.5701 | 0.476 | 0.2369 | 0.8574 | 0.4168 | 0.4836 | 0 | 0 | 0 | 0 | 0.5867 | 0.7639 | 7.1315 | 0.7741 | 0.8767 | 9.7167 | 0.6773 | 0.7158 | 2.6506 | 34.2569 | 12.6511 | 0.1262 | 28.614 | MistralForCausalLM | bfloat16 | apache-2.0 | 7.242 | 15 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🔶 : fine-tuned | augmxnt/shisa-gamma-7b-v1 | 0.547 | 0.2369 | 0.4704 | 0.2215 | 0.4648 | 0.8258 | 0.622 | 0.8253 | 0.7628 | 0.5497 | 0.9113 | 0.1262 | 0.666 | 0.8601 | 12.2649 | 0.9022 | 0.9458 | 15.6194 | 0.8687 | 0.4704 | 0.8765 | 0.6034 | 0.8722 | 0.9178 | 0.4577 | 0.4541 | 0.8118 | 0.7797 | 0.7469 | 0.9113 | 0.8235 | 0.7898 | 0.683 | 0.622 | 0.2369 | 0.8574 | 0.4756 | 0.5253 | 0 | 0.2735 | 0.0265 | 0.0612 | 0.7464 | 0.7979 | 8.7329 | 0.8034 | 0.8925 | 10.2192 | 0.7268 | 0.7158 | 2.6506 | 34.2569 | 12.6511 | 0.1262 | 28.614 | MistralForCausalLM | bfloat16 | apache-2.0 | 7.242 | 15 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
🟢 : pretrained | Qwen/QwQ-32B-Preview | 0.5785 | 0.6606 | 0.2402 | 0.1527 | 0.7286 | 0.8531 | 0.82 | 0.8396 | 0.7285 | 0.3296 | 0.8978 | 0.1134 | 0.1838 | 0.8478 | 11.7965 | 0.9037 | 0.9536 | 16.5924 | 0.8832 | 0.2402 | 0.895 | 0.6063 | 0.7139 | 0.9178 | 0.4363 | 0.6944 | 0.8012 | 0.7891 | 0.7319 | 0.8978 | 0.8994 | 0.8717 | 0.7464 | 0.82 | 0.6606 | 0.9478 | 0.7627 | 0.3686 | 0 | 0.0008 | 0.0531 | 0.0022 | 0.7073 | 0.7991 | 9.0753 | 0.8271 | 0.8955 | 9.2158 | 0.7442 | 0.7031 | 3.1354 | 29.4104 | 11.3342 | 0.1134 | 25.3233 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 594 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
🟢 : pretrained | Qwen/QwQ-32B-Preview | 0.6715 | 0.6606 | 0.5794 | 0.2813 | 0.7756 | 0.8955 | 0.938 | 0.8487 | 0.7912 | 0.5775 | 0.9254 | 0.1134 | 0.5875 | 0.8633 | 12.6447 | 0.9093 | 0.9555 | 16.8456 | 0.8861 | 0.5794 | 0.9126 | 0.6609 | 0.8139 | 0.9517 | 0.6171 | 0.7557 | 0.8965 | 0.8049 | 0.7798 | 0.9254 | 0.8768 | 0.872 | 0.8223 | 0.938 | 0.6606 | 0.9478 | 0.7955 | 0.5279 | 0.0203 | 0.3842 | 0.1327 | 0.0816 | 0.7875 | 0.8335 | 12.1622 | 0.8381 | 0.9058 | 11.4057 | 0.7612 | 0.7031 | 3.1354 | 29.4104 | 11.3342 | 0.1134 | 25.3233 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 32.764 | 594 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Qwen/Qwen2.5-14B-Instruct | 0.5476 | 0.4438 | 0.2239 | 0.1282 | 0.6851 | 0.8539 | 0.794 | 0.8041 | 0.7474 | 0.3898 | 0.8551 | 0.0982 | 0.4017 | 0.8419 | 9.2171 | 0.8957 | 0.9195 | 15.4526 | 0.8051 | 0.2239 | 0.8945 | 0.6351 | 0.8306 | 0.9285 | 0.3404 | 0.6631 | 0.7703 | 0.6894 | 0.8114 | 0.8551 | 0.884 | 0.8556 | 0.7388 | 0.794 | 0.4438 | 0.7068 | 0.7072 | 0.4274 | 0.0209 | 0.0088 | 0.0088 | 0.0008 | 0.6014 | 0.792 | 7.5747 | 0.811 | 0.875 | 8.7641 | 0.7048 | 0.6941 | 2.8388 | 25.5114 | 9.8292 | 0.0982 | 22.4451 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 14.77 | 126 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | Qwen/Qwen2.5-14B-Instruct | 0.6285 | 0.4438 | 0.5656 | 0.2613 | 0.7345 | 0.8802 | 0.892 | 0.8414 | 0.7568 | 0.5383 | 0.9019 | 0.0982 | 0.5091 | 0.8628 | 12.6532 | 0.9084 | 0.9458 | 16.5642 | 0.8663 | 0.5656 | 0.8988 | 0.6149 | 0.8278 | 0.9517 | 0.5883 | 0.7043 | 0.8225 | 0.762 | 0.7569 | 0.9019 | 0.8874 | 0.8586 | 0.7901 | 0.892 | 0.4438 | 0.7068 | 0.7648 | 0.5175 | 0.0969 | 0.3345 | 0.0147 | 0.0848 | 0.7754 | 0.8212 | 9.9418 | 0.8341 | 0.9023 | 10.5742 | 0.7567 | 0.6941 | 2.8388 | 25.5114 | 9.8292 | 0.0982 | 22.4451 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 14.77 | 126 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | AIDC-AI/Marco-o1 | 0.2171 | 0 | 0.1841 | 0.0632 | 0.1052 | 0.4735 | 0.042 | 0.6498 | 0.23 | 0.1423 | 0.4048 | 0.0936 | 0.1335 | 0.695 | 7.3933 | 0.7214 | 0.8533 | 11.4251 | 0.6421 | 0.1841 | 0 | 0.4828 | 0 | 0.7998 | 0.1625 | 0.203 | 0.3246 | 0 | 0.3428 | 0.4048 | 0.8694 | 0.835 | 0.6205 | 0.042 | 0 | 0 | 0.0073 | 0.1308 | 0 | 0.0071 | 0.0114 | 0.0028 | 0.2947 | 0.665 | 6.8745 | 0.6654 | 0.8227 | 7.7347 | 0.5703 | 0.6912 | 2.3785 | 26.7848 | 9.367 | 0.0936 | 22.6888 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 7.616 | 542 | main | 0 | False | v1.4.1 | v0.6.3.post1 |
⭕ : instruction-tuned | AIDC-AI/Marco-o1 | 0.5419 | 0 | 0.4799 | 0.2396 | 0.6505 | 0.8392 | 0.782 | 0.8328 | 0.7167 | 0.4246 | 0.9021 | 0.0936 | 0.3833 | 0.8513 | 11.848 | 0.895 | 0.949 | 15.5601 | 0.8771 | 0.4799 | 0.8615 | 0.5661 | 0.6889 | 0.9214 | 0.4758 | 0.6125 | 0.8221 | 0.7115 | 0.795 | 0.9021 | 0.8697 | 0.847 | 0.7347 | 0.782 | 0 | 0 | 0.6884 | 0.4146 | 0.036 | 0.3458 | 0.0708 | 0.0545 | 0.6909 | 0.8011 | 8.5787 | 0.8175 | 0.894 | 9.6318 | 0.7415 | 0.6912 | 2.3785 | 26.7848 | 9.367 | 0.0936 | 22.6888 | Qwen2ForCausalLM | bfloat16 | apache-2.0 | 7.616 | 542 | main | 4 | False | v1.4.1 | v0.6.3.post1 |
No dataset card yet