{ "results": { "pubmedqa": { "acc,none": 0.338, "acc_stderr,none": 0.021175665695209397, "alias": "pubmedqa" }, "mmlu_professional_medicine": { "alias": "professional_medicine", "acc,none": 0.18382352941176472, "acc_stderr,none": 0.02352924218519311 }, "mmlu_medical_genetics": { "alias": "medical_genetics", "acc,none": 0.3, "acc_stderr,none": 0.046056618647183814 }, "mmlu_college_medicine": { "alias": "college_medicine", "acc,none": 0.20809248554913296, "acc_stderr,none": 0.030952890217749884 }, "mmlu_college_biology": { "alias": "college_biology", "acc,none": 0.2569444444444444, "acc_stderr,none": 0.03653946969442099 }, "mmlu_clinical_knowledge": { "alias": "clinical_knowledge", "acc,none": 0.21509433962264152, "acc_stderr,none": 0.025288394502891377 }, "mmlu_anatomy": { "alias": "anatomy", "acc,none": 0.18518518518518517, "acc_stderr,none": 0.03355677216313142 }, "medqa_4options": { "acc,none": 0.27729772191673213, "acc_stderr,none": 0.0125518952732286, "acc_norm,none": 0.27729772191673213, "acc_norm_stderr,none": 0.0125518952732286, "alias": "medqa_4options" }, "medmcqa": { "acc,none": 0.3222567535261774, "acc_stderr,none": 0.007226726167378288, "acc_norm,none": 0.3222567535261774, "acc_norm_stderr,none": 0.007226726167378288, "alias": "medmcqa" } }, "config": { "model": "hf", "model_args": "pretrained=HiTZ/Medical-mT5-large,revision=main,dtype=float32", "batch_size": "auto", "batch_sizes": [ 4 ], "device": "cuda:0", "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null, "model_dtype": "float32", "model_name": "HiTZ/Medical-mT5-large", "model_sha": "main" } }