OLMoE
Collection
Artifacts for open mixture-of-experts language models.
•
13 items
•
Updated
•
27
Error code: DatasetGenerationError Exception: TypeError Message: Couldn't cast array of type struct<human_eval: double, human_eval_cpp: double, human_eval_js: double, human_eval_return_simple: double, human_eval_return_complex: double, human_eval_25: double, human_eval_50: double, human_eval_75: double> to {'mmlu_zeroshot': Value(dtype='float64', id=None), 'hellaswag_zeroshot': Value(dtype='float64', id=None), 'jeopardy': Value(dtype='float64', id=None), 'triviaqa_sm_sub': Value(dtype='float64', id=None), 'gsm8k_cot': Value(dtype='float64', id=None), 'agi_eval_sat_math_cot': Value(dtype='float64', id=None), 'aqua_cot': Value(dtype='float64', id=None), 'svamp_cot': Value(dtype='float64', id=None), 'bigbench_qa_wikidata': Value(dtype='float64', id=None), 'arc_easy': Value(dtype='float64', id=None), 'arc_challenge': Value(dtype='float64', id=None), 'mmlu_fewshot': Value(dtype='float64', id=None), 'bigbench_misconceptions': Value(dtype='float64', id=None), 'copa': Value(dtype='float64', id=None), 'siqa': Value(dtype='float64', id=None), 'commonsense_qa': Value(dtype='float64', id=None), 'piqa': Value(dtype='float64', id=None), 'openbook_qa': Value(dtype='float64', id=None), 'bigbench_novel_concepts': Value(dtype='float64', id=None), 'bigbench_strange_stories': Value(dtype='float64', id=None), 'bigbench_strategy_qa': Value(dtype='float64', id=None), 'lambada_openai': Value(dtype='float64', id=None), 'hellaswag': Value(dtype='float64', id=None), 'winograd': Value(dtype='float64', id=None), 'winogrande': Value(dtype='float64', id=None), 'bigbench_conlang_translation': Value(dtype='float64', id=None), 'bigbench_language_identification': Value(dtype='float64', id=None), 'bigbench_conceptual_combinations': Value(dtype='float64', id=None), 'bigbench_elementary_math_qa': Value(dtype='float64', id=None), 'bigbench_dyck_languages': Value(dtype='float64', id=None), 'agi_eval_lsat_ar': Value(dtype='float64', id=None), 'bigbench_cs_algorithms': Value(dtype='float64', id=None), 'bigbench_logical_deduction': Value(dtype='float64', id=None), 'bigbench_operators': Value(dtype='float64', id=None), 'bigbench_repeat_copy_logic': Value(dtype='float64', id=None), 'simple_arithmetic_nospaces': Value(dtype='float64', id=None), 'simple_arithmetic_withspaces': Value(dtype='float64', id=None), 'math_qa': Value(dtype='float64', id=None), 'logi_qa': Value(dtype='float64', id=None), 'pubmed_qa_labeled': Value(dtype='float64', id=None), 'squad': Value(dtype='float64', id=None), 'agi_eval_lsat_rc': Value(dtype='float64', id=None), 'agi_eval_lsat_lr': Value(dtype='float64', id=None), 'coqa': Value(dtype='float64', id=None), 'bigbench_understanding_fables': Value(dtype='float64', id=None), 'boolq': Value(dtype='float64', id=None), 'agi_eval_sat_en': Value(dtype='float64', id=None), 'winogender_mc_female': Value(dtype='float64', id=None), 'winogender_mc_male': Value(dtype='float64', id=None), 'enterprise_pii_classification': Value(dtype='float64', id=None), 'bbq': Value(dtype='float64', id=None), 'gpqa_main': Value(dtype='float64', id=None), 'gpqa_diamond': Value(dtype='float64', id=None)} Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2013, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 585, in write_table pa_table = table_cast(pa_table, self._schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2302, in table_cast return cast_table_to_schema(table, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2261, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2261, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1802, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1802, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2020, in cast_array_to_feature arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()] File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2020, in <listcomp> arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()] File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper return func(array, *args, **kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2122, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{_short_str(array.type)}\nto\n{_short_str(feature)}") TypeError: Couldn't cast array of type struct<human_eval: double, human_eval_cpp: double, human_eval_js: double, human_eval_return_simple: double, human_eval_return_complex: double, human_eval_25: double, human_eval_50: double, human_eval_75: double> to {'mmlu_zeroshot': Value(dtype='float64', id=None), 'hellaswag_zeroshot': Value(dtype='float64', id=None), 'jeopardy': Value(dtype='float64', id=None), 'triviaqa_sm_sub': Value(dtype='float64', id=None), 'gsm8k_cot': Value(dtype='float64', id=None), 'agi_eval_sat_math_cot': Value(dtype='float64', id=None), 'aqua_cot': Value(dtype='float64', id=None), 'svamp_cot': Value(dtype='float64', id=None), 'bigbench_qa_wikidata': Value(dtype='float64', id=None), 'arc_easy': Value(dtype='float64', id=None), 'arc_challenge': Value(dtype='float64', id=None), 'mmlu_fewshot': Value(dtype='float64', id=None), 'bigbench_misconceptions': Value(dtype='float64', id=None), 'copa': Value(dtype='float64', id=None), 'siqa': Value(dtype='float64', id=None), 'commonsense_qa': Value(dtype='float64', id=None), 'piqa': Value(dtype='float64', id=None), 'openbook_qa': Value(dtype='float64', id=None), 'bigbench_novel_concepts': Value(dtype='float64', id=None), 'bigbench_strange_stories': Value(dtype='float64', id=None), 'bigbench_strategy_qa': Value(dtype='float64', id=None), 'lambada_openai': Value(dtype='float64', id=None), 'hellaswag': Value(dtype='float64', id=None), 'winograd': Value(dtype='float64', id=None), 'winogrande': Value(dtype='float64', id=None), 'bigbench_conlang_translation': Value(dtype='float64', id=None), 'bigbench_language_identification': Value(dtype='float64', id=None), 'bigbench_conceptual_combinations': Value(dtype='float64', id=None), 'bigbench_elementary_math_qa': Value(dtype='float64', id=None), 'bigbench_dyck_languages': Value(dtype='float64', id=None), 'agi_eval_lsat_ar': Value(dtype='float64', id=None), 'bigbench_cs_algorithms': Value(dtype='float64', id=None), 'bigbench_logical_deduction': Value(dtype='float64', id=None), 'bigbench_operators': Value(dtype='float64', id=None), 'bigbench_repeat_copy_logic': Value(dtype='float64', id=None), 'simple_arithmetic_nospaces': Value(dtype='float64', id=None), 'simple_arithmetic_withspaces': Value(dtype='float64', id=None), 'math_qa': Value(dtype='float64', id=None), 'logi_qa': Value(dtype='float64', id=None), 'pubmed_qa_labeled': Value(dtype='float64', id=None), 'squad': Value(dtype='float64', id=None), 'agi_eval_lsat_rc': Value(dtype='float64', id=None), 'agi_eval_lsat_lr': Value(dtype='float64', id=None), 'coqa': Value(dtype='float64', id=None), 'bigbench_understanding_fables': Value(dtype='float64', id=None), 'boolq': Value(dtype='float64', id=None), 'agi_eval_sat_en': Value(dtype='float64', id=None), 'winogender_mc_female': Value(dtype='float64', id=None), 'winogender_mc_male': Value(dtype='float64', id=None), 'enterprise_pii_classification': Value(dtype='float64', id=None), 'bbq': Value(dtype='float64', id=None), 'gpqa_main': Value(dtype='float64', id=None), 'gpqa_diamond': Value(dtype='float64', id=None)} The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1396, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1045, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1029, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1124, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1884, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2040, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
name
string | uuid
string | model
string | creation_date
string | eval_metrics
dict | missing tasks
string | aggregated_task_categories_centered
dict | aggregated_centered_results
float64 | aggregated_results
float64 | rw_small
float64 | rw_small_centered
float64 | 95%_CI_above
float64 | 95%_CI_above_centered
float64 | 99%_CI_above
float64 | 99%_CI_above_centered
float64 | low_variance_datasets
float64 | low_variance_datasets_centered
float64 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
heavy | 83bb37f6-ddd0-43d8-894e-446d345ff7c1 | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/main | 2024_08_06-17_10_22 | {
"icl": {
"mmlu_zeroshot": 0.4163723833728255,
"hellaswag_zeroshot": 0.7599083781242371,
"jeopardy": 0.4884525716304779,
"triviaqa_sm_sub": 0.5419999957084656,
"gsm8k_cot": 0.06444276124238968,
"agi_eval_sat_math_cot": 0.05454545468091965,
"aqua_cot": 0.02448979578912258,
"svamp_cot": 0.30000001192092896,
"bigbench_qa_wikidata": 0.6824467182159424,
"arc_easy": 0.7714646458625793,
"arc_challenge": 0.5332764387130737,
"mmlu_fewshot": 0.5218371202548345,
"bigbench_misconceptions": 0.5981734991073608,
"copa": 0.8100000023841858,
"siqa": 0.7154554724693298,
"commonsense_qa": 0.6723996996879578,
"piqa": 0.8133841156959534,
"openbook_qa": 0.414000004529953,
"bigbench_novel_concepts": 0.625,
"bigbench_strange_stories": 0.6666666865348816,
"bigbench_strategy_qa": 0.5622542500495911,
"lambada_openai": 0.7269551753997803,
"hellaswag": 0.776239812374115,
"winograd": 0.8278388381004333,
"winogrande": 0.6803472638130188,
"bigbench_conlang_translation": 0.03658536449074745,
"bigbench_language_identification": 0.3192000091075897,
"bigbench_conceptual_combinations": 0.5145630836486816,
"bigbench_elementary_math_qa": 0.26218554377555847,
"bigbench_dyck_languages": 0.19300000369548798,
"agi_eval_lsat_ar": 0.24347825348377228,
"bigbench_cs_algorithms": 0.47121211886405945,
"bigbench_logical_deduction": 0.26600000262260437,
"bigbench_operators": 0.3619047701358795,
"bigbench_repeat_copy_logic": 0.15625,
"simple_arithmetic_nospaces": 0.17599999904632568,
"simple_arithmetic_withspaces": 0.19499999284744263,
"math_qa": 0.26449882984161377,
"logi_qa": 0.3486943244934082,
"pubmed_qa_labeled": 0.5609999895095825,
"squad": 0.5288552641868591,
"agi_eval_lsat_rc": 0.4738805890083313,
"agi_eval_lsat_lr": 0.4019607901573181,
"coqa": 0.43692848086357117,
"bigbench_understanding_fables": 0.47089946269989014,
"boolq": 0.7333333492279053,
"agi_eval_sat_en": 0.553398072719574,
"winogender_mc_female": 0.5,
"winogender_mc_male": 0.550000011920929,
"enterprise_pii_classification": 0.523122251033783,
"bbq": 0.4975668625398116,
"gpqa_main": 0.2477678507566452,
"gpqa_diamond": 0.2222222238779068
}
} | [] | {
"commonsense reasoning": 0.4523028646368181,
"language understanding": 0.4706549446387432,
"reading comprehension": 0.39537327579761805,
"safety": 0.03534456274726175,
"symbolic problem solving": 0.15998990932590068,
"world knowledge": 0.3526500500886761
} | 0.313479 | 0.463348 | 0.714722 | 0.505301 | 0.56121 | 0.415861 | 0.572426 | 0.457942 | 0.563676 | 0.462567 |
heavy | 70f3f719-28b5-46f5-b58e-bd89765d1e40 | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1220000-tokens5117B | 2024_08_06-17_05_21 | {
"icl": {
"mmlu_zeroshot": 0.4273025756865217,
"hellaswag_zeroshot": 0.7597092390060425,
"jeopardy": 0.4871806979179382,
"triviaqa_sm_sub": 0.5303333401679993,
"gsm8k_cot": 0.07354056090116501,
"agi_eval_sat_math_cot": 0.040909089148044586,
"aqua_cot": 0.02857142873108387,
"svamp_cot": 0.2800000011920929,
"bigbench_qa_wikidata": 0.6884503960609436,
"arc_easy": 0.7845118045806885,
"arc_challenge": 0.5341296792030334,
"mmlu_fewshot": 0.5194499257363772,
"bigbench_misconceptions": 0.5525113940238953,
"copa": 0.800000011920929,
"siqa": 0.707267165184021,
"commonsense_qa": 0.6699426770210266,
"piqa": 0.8117519021034241,
"openbook_qa": 0.4399999976158142,
"bigbench_novel_concepts": 0.625,
"bigbench_strange_stories": 0.6839080452919006,
"bigbench_strategy_qa": 0.5810397267341614,
"lambada_openai": 0.7219095826148987,
"hellaswag": 0.7750447988510132,
"winograd": 0.831501841545105,
"winogrande": 0.6850828528404236,
"bigbench_conlang_translation": 0.060975611209869385,
"bigbench_language_identification": 0.34049999713897705,
"bigbench_conceptual_combinations": 0.5048543810844421,
"bigbench_elementary_math_qa": 0.27025681734085083,
"bigbench_dyck_languages": 0.1589999943971634,
"agi_eval_lsat_ar": 0.2652173936367035,
"bigbench_cs_algorithms": 0.5015151500701904,
"bigbench_logical_deduction": 0.25333333015441895,
"bigbench_operators": 0.34285715222358704,
"bigbench_repeat_copy_logic": 0.15625,
"simple_arithmetic_nospaces": 0.1809999942779541,
"simple_arithmetic_withspaces": 0.20600000023841858,
"math_qa": 0.2708682417869568,
"logi_qa": 0.3425499200820923,
"pubmed_qa_labeled": 0.4659999907016754,
"squad": 0.5244086980819702,
"agi_eval_lsat_rc": 0.43656715750694275,
"agi_eval_lsat_lr": 0.386274516582489,
"coqa": 0.4436928331851959,
"bigbench_understanding_fables": 0.4444444477558136,
"boolq": 0.7281345725059509,
"agi_eval_sat_en": 0.5485436916351318,
"winogender_mc_female": 0.46666666865348816,
"winogender_mc_male": 0.5833333134651184,
"enterprise_pii_classification": 0.5372606515884399,
"bbq": 0.48323566534302453,
"gpqa_main": 0.2232142835855484,
"gpqa_diamond": 0.21212121844291687
}
} | [] | {
"commonsense reasoning": 0.46081640452671535,
"language understanding": 0.4762512398893946,
"reading comprehension": 0.36885401178478144,
"safety": 0.03524814952503552,
"symbolic problem solving": 0.15957477013304083,
"world knowledge": 0.34119598718414534
} | 0.30932 | 0.459965 | 0.715261 | 0.504311 | 0.558937 | 0.414268 | 0.570154 | 0.456544 | 0.565945 | 0.465203 |
heavy | c260eedc-dbb5-4ff4-afa9-d163e8d7585b | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1223842-tokens5100B | 2024_08_06-16_39_52 | {
"icl": {
"mmlu_zeroshot": 0.43332746128241223,
"hellaswag_zeroshot": 0.7701653242111206,
"jeopardy": 0.5032301664352417,
"triviaqa_sm_sub": 0.5586666464805603,
"gsm8k_cot": 0.07429870963096619,
"agi_eval_sat_math_cot": 0.06363636255264282,
"aqua_cot": 0.020408162847161293,
"svamp_cot": 0.33000001311302185,
"bigbench_qa_wikidata": 0.6918458938598633,
"arc_easy": 0.7765151262283325,
"arc_challenge": 0.5375426411628723,
"mmlu_fewshot": 0.5333352494135237,
"bigbench_misconceptions": 0.6255707740783691,
"copa": 0.7799999713897705,
"siqa": 0.6934493184089661,
"commonsense_qa": 0.6928746700286865,
"piqa": 0.8204570412635803,
"openbook_qa": 0.4480000138282776,
"bigbench_novel_concepts": 0.65625,
"bigbench_strange_stories": 0.6954023241996765,
"bigbench_strategy_qa": 0.5696811079978943,
"lambada_openai": 0.7327769994735718,
"hellaswag": 0.7857996225357056,
"winograd": 0.8461538553237915,
"winogrande": 0.6898184418678284,
"bigbench_conlang_translation": 0.0731707289814949,
"bigbench_language_identification": 0.31049999594688416,
"bigbench_conceptual_combinations": 0.5631067752838135,
"bigbench_elementary_math_qa": 0.26944443583488464,
"bigbench_dyck_languages": 0.2150000035762787,
"agi_eval_lsat_ar": 0.2869565188884735,
"bigbench_cs_algorithms": 0.47196969389915466,
"bigbench_logical_deduction": 0.2460000067949295,
"bigbench_operators": 0.3380952477455139,
"bigbench_repeat_copy_logic": 0.1875,
"simple_arithmetic_nospaces": 0.20100000500679016,
"simple_arithmetic_withspaces": 0.22100000083446503,
"math_qa": 0.27522629499435425,
"logi_qa": 0.34562212228775024,
"pubmed_qa_labeled": 0.5789999961853027,
"squad": 0.5235572457313538,
"agi_eval_lsat_rc": 0.46641790866851807,
"agi_eval_lsat_lr": 0.37254902720451355,
"coqa": 0.4366779327392578,
"bigbench_understanding_fables": 0.4761904776096344,
"boolq": 0.7318042516708374,
"agi_eval_sat_en": 0.5291262269020081,
"winogender_mc_female": 0.5,
"winogender_mc_male": 0.6000000238418579,
"enterprise_pii_classification": 0.5216494798660278,
"bbq": 0.5055712298913435,
"gpqa_main": 0.2254464328289032,
"gpqa_diamond": 0.19696970283985138
}
} | [] | {
"commonsense reasoning": 0.46523631517512387,
"language understanding": 0.4930994285627118,
"reading comprehension": 0.3874464948710642,
"safety": 0.06361036679961463,
"symbolic problem solving": 0.1715689478790787,
"world knowledge": 0.36090664066069306
} | 0.324591 | 0.471675 | 0.714308 | 0.502137 | 0.571701 | 0.429074 | 0.581191 | 0.468356 | 0.571693 | 0.472335 |
heavy | 72557004-cff2-408c-b3e2-6a004215f09d | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/jetmoe-8b/main | 2024_08_06-16_55_05 | {
"icl": {
"mmlu_zeroshot": 0.4326617845840621,
"hellaswag_zeroshot": 0.7868950366973877,
"jeopardy": 0.45192375779151917,
"triviaqa_sm_sub": 0.4986666738986969,
"gsm8k_cot": 0.3025018870830536,
"agi_eval_sat_math_cot": 0.11363636702299118,
"aqua_cot": 0.05306122452020645,
"svamp_cot": 0.4466666579246521,
"bigbench_qa_wikidata": 0.7317061424255371,
"arc_easy": 0.7356902360916138,
"arc_challenge": 0.4530716836452484,
"mmlu_fewshot": 0.4966564126182021,
"bigbench_misconceptions": 0.5707762837409973,
"copa": 0.8399999737739563,
"siqa": 0.8162742853164673,
"commonsense_qa": 0.7911548018455505,
"piqa": 0.8084874749183655,
"openbook_qa": 0.3959999978542328,
"bigbench_novel_concepts": 0.53125,
"bigbench_strange_stories": 0.6781609058380127,
"bigbench_strategy_qa": 0.5801659822463989,
"lambada_openai": 0.6976518630981445,
"hellaswag": 0.793367862701416,
"winograd": 0.8388278484344482,
"winogrande": 0.6835043430328369,
"bigbench_conlang_translation": 0.09146341681480408,
"bigbench_language_identification": 0.336899995803833,
"bigbench_conceptual_combinations": 0.553398072719574,
"bigbench_elementary_math_qa": 0.2795073390007019,
"bigbench_dyck_languages": 0.3190000057220459,
"agi_eval_lsat_ar": 0.2869565188884735,
"bigbench_cs_algorithms": 0.5098484754562378,
"bigbench_logical_deduction": 0.3033333420753479,
"bigbench_operators": 0.538095235824585,
"bigbench_repeat_copy_logic": 0.21875,
"simple_arithmetic_nospaces": 0.3070000112056732,
"simple_arithmetic_withspaces": 0.3100000023841858,
"math_qa": 0.28226616978645325,
"logi_qa": 0.33794161677360535,
"pubmed_qa_labeled": 0.6430000066757202,
"squad": 0.2401135265827179,
"agi_eval_lsat_rc": 0.46268656849861145,
"agi_eval_lsat_lr": 0.3450980484485626,
"coqa": 0.43893274664878845,
"bigbench_understanding_fables": 0.41798943281173706,
"boolq": 0.8223241567611694,
"agi_eval_sat_en": 0.6213592290878296,
"winogender_mc_female": 0.5,
"winogender_mc_male": 0.5833333134651184,
"enterprise_pii_classification": 0.6017673015594482,
"bbq": 0.5354551727121527,
"gpqa_main": 0.2566964328289032,
"gpqa_diamond": 0.2222222238779068
}
} | [] | {
"commonsense reasoning": 0.4857313116330063,
"language understanding": 0.4936472507617851,
"reading comprehension": 0.3914801840458,
"safety": 0.11027789386835968,
"symbolic problem solving": 0.24344731914942666,
"world knowledge": 0.3286514173125663
} | 0.346155 | 0.48857 | 0.720248 | 0.535608 | 0.577331 | 0.443234 | 0.592481 | 0.489897 | 0.578146 | 0.485983 |
/net/nfs.cirrascale/allennlp/davidw/proj/OLMoE/script/humaneval | dc03f2af-e3c1-479d-98bd-6efc73e684e4 | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/main | 2024_08_06-02_09_56 | {
"icl": {
"human_eval": 0.024390242993831635,
"human_eval_cpp": 0.1304347813129425,
"human_eval_js": 0,
"human_eval_return_simple": 0.6486486196517944,
"human_eval_return_complex": 0.14960630238056183,
"human_eval_25": 0.03658536449074745,
"human_eval_50": 0.08536585420370102,
"human_eval_75": 0.16463415324687958
}
} | ['mmlu_zeroshot', 'hellaswag_zeroshot', 'jeopardy', 'triviaqa_sm_sub', 'gsm8k_cot', 'agi_eval_sat_math_cot', 'aqua_cot', 'svamp_cot', 'bigbench_qa_wikidata', 'arc_easy', 'arc_challenge', 'mmlu_fewshot', 'bigbench_misconceptions', 'copa', 'siqa', 'commonsense_qa', 'piqa', 'openbook_qa', 'bigbench_novel_concepts', 'bigbench_strange_stories', 'bigbench_strategy_qa', 'lambada_openai', 'hellaswag', 'winograd', 'winogrande', 'bigbench_conlang_translation', 'bigbench_language_identification', 'bigbench_conceptual_combinations', 'bigbench_elementary_math_qa', 'bigbench_dyck_languages', 'agi_eval_lsat_ar', 'bigbench_cs_algorithms', 'bigbench_logical_deduction', 'bigbench_operators', 'bigbench_repeat_copy_logic', 'simple_arithmetic_nospaces', 'simple_arithmetic_withspaces', 'math_qa', 'logi_qa', 'pubmed_qa_labeled', 'squad', 'agi_eval_lsat_rc', 'agi_eval_lsat_lr', 'coqa', 'bigbench_understanding_fables', 'boolq', 'agi_eval_sat_en', 'winogender_mc_female', 'winogender_mc_male', 'enterprise_pii_classification', 'bbq', 'gpqa_main', 'gpqa_diamond'] | {
"commonsense reasoning": null,
"language understanding": null,
"reading comprehension": null,
"safety": null,
"symbolic problem solving": null,
"world knowledge": null
} | null | null | null | null | null | null | null | null | null | null |
/net/nfs.cirrascale/allennlp/davidw/proj/OLMoE/script/humaneval | d31d6b3f-15ee-4d96-ac2e-b2d780bf206e | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1220000-tokens5117B | 2024_08_06-02_14_00 | {
"icl": {
"human_eval": 0.006097560748457909,
"human_eval_cpp": 0.12422360479831696,
"human_eval_js": 0,
"human_eval_return_simple": 0.7567567825317383,
"human_eval_return_complex": 0.14173229038715363,
"human_eval_25": 0.060975611209869385,
"human_eval_50": 0.1036585345864296,
"human_eval_75": 0.12195122241973877
}
} | ['mmlu_zeroshot', 'hellaswag_zeroshot', 'jeopardy', 'triviaqa_sm_sub', 'gsm8k_cot', 'agi_eval_sat_math_cot', 'aqua_cot', 'svamp_cot', 'bigbench_qa_wikidata', 'arc_easy', 'arc_challenge', 'mmlu_fewshot', 'bigbench_misconceptions', 'copa', 'siqa', 'commonsense_qa', 'piqa', 'openbook_qa', 'bigbench_novel_concepts', 'bigbench_strange_stories', 'bigbench_strategy_qa', 'lambada_openai', 'hellaswag', 'winograd', 'winogrande', 'bigbench_conlang_translation', 'bigbench_language_identification', 'bigbench_conceptual_combinations', 'bigbench_elementary_math_qa', 'bigbench_dyck_languages', 'agi_eval_lsat_ar', 'bigbench_cs_algorithms', 'bigbench_logical_deduction', 'bigbench_operators', 'bigbench_repeat_copy_logic', 'simple_arithmetic_nospaces', 'simple_arithmetic_withspaces', 'math_qa', 'logi_qa', 'pubmed_qa_labeled', 'squad', 'agi_eval_lsat_rc', 'agi_eval_lsat_lr', 'coqa', 'bigbench_understanding_fables', 'boolq', 'agi_eval_sat_en', 'winogender_mc_female', 'winogender_mc_male', 'enterprise_pii_classification', 'bbq', 'gpqa_main', 'gpqa_diamond'] | {
"commonsense reasoning": null,
"language understanding": null,
"reading comprehension": null,
"safety": null,
"symbolic problem solving": null,
"world knowledge": null
} | null | null | null | null | null | null | null | null | null | null |
/net/nfs.cirrascale/allennlp/davidw/proj/OLMoE/script/humaneval | 3a11e770-7075-42f3-81f6-c14a6db4953d | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1223842-tokens5100B | 2024_08_06-02_21_03 | {
"icl": {
"human_eval": 0.012195121496915817,
"human_eval_cpp": 0.1304347813129425,
"human_eval_js": 0,
"human_eval_return_simple": 0.8648648858070374,
"human_eval_return_complex": 0.14960630238056183,
"human_eval_25": 0.04268292710185051,
"human_eval_50": 0.09146341681480408,
"human_eval_75": 0.16463415324687958
}
} | ['mmlu_zeroshot', 'hellaswag_zeroshot', 'jeopardy', 'triviaqa_sm_sub', 'gsm8k_cot', 'agi_eval_sat_math_cot', 'aqua_cot', 'svamp_cot', 'bigbench_qa_wikidata', 'arc_easy', 'arc_challenge', 'mmlu_fewshot', 'bigbench_misconceptions', 'copa', 'siqa', 'commonsense_qa', 'piqa', 'openbook_qa', 'bigbench_novel_concepts', 'bigbench_strange_stories', 'bigbench_strategy_qa', 'lambada_openai', 'hellaswag', 'winograd', 'winogrande', 'bigbench_conlang_translation', 'bigbench_language_identification', 'bigbench_conceptual_combinations', 'bigbench_elementary_math_qa', 'bigbench_dyck_languages', 'agi_eval_lsat_ar', 'bigbench_cs_algorithms', 'bigbench_logical_deduction', 'bigbench_operators', 'bigbench_repeat_copy_logic', 'simple_arithmetic_nospaces', 'simple_arithmetic_withspaces', 'math_qa', 'logi_qa', 'pubmed_qa_labeled', 'squad', 'agi_eval_lsat_rc', 'agi_eval_lsat_lr', 'coqa', 'bigbench_understanding_fables', 'boolq', 'agi_eval_sat_en', 'winogender_mc_female', 'winogender_mc_male', 'enterprise_pii_classification', 'bbq', 'gpqa_main', 'gpqa_diamond'] | {
"commonsense reasoning": null,
"language understanding": null,
"reading comprehension": null,
"safety": null,
"symbolic problem solving": null,
"world knowledge": null
} | null | null | null | null | null | null | null | null | null | null |
/net/nfs.cirrascale/allennlp/davidw/proj/OLMoE/script/humaneval | 9a26c983-2b2b-46bc-bc2f-9e1d9c153710 | /net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/jetmoe-8b/main | 2024_08_05-22_40_40 | {
"icl": {
"human_eval": 0.32926830649375916,
"human_eval_cpp": 0.2732919156551361,
"human_eval_js": 0,
"human_eval_return_simple": 0.7837837934494019,
"human_eval_return_complex": 0.4488188922405243,
"human_eval_25": 0.4390243887901306,
"human_eval_50": 0.5243902206420898,
"human_eval_75": 0.6463414430618286
}
} | ['mmlu_zeroshot', 'hellaswag_zeroshot', 'jeopardy', 'triviaqa_sm_sub', 'gsm8k_cot', 'agi_eval_sat_math_cot', 'aqua_cot', 'svamp_cot', 'bigbench_qa_wikidata', 'arc_easy', 'arc_challenge', 'mmlu_fewshot', 'bigbench_misconceptions', 'copa', 'siqa', 'commonsense_qa', 'piqa', 'openbook_qa', 'bigbench_novel_concepts', 'bigbench_strange_stories', 'bigbench_strategy_qa', 'lambada_openai', 'hellaswag', 'winograd', 'winogrande', 'bigbench_conlang_translation', 'bigbench_language_identification', 'bigbench_conceptual_combinations', 'bigbench_elementary_math_qa', 'bigbench_dyck_languages', 'agi_eval_lsat_ar', 'bigbench_cs_algorithms', 'bigbench_logical_deduction', 'bigbench_operators', 'bigbench_repeat_copy_logic', 'simple_arithmetic_nospaces', 'simple_arithmetic_withspaces', 'math_qa', 'logi_qa', 'pubmed_qa_labeled', 'squad', 'agi_eval_lsat_rc', 'agi_eval_lsat_lr', 'coqa', 'bigbench_understanding_fables', 'boolq', 'agi_eval_sat_en', 'winogender_mc_female', 'winogender_mc_male', 'enterprise_pii_classification', 'bbq', 'gpqa_main', 'gpqa_diamond'] | {
"commonsense reasoning": null,
"language understanding": null,
"reading comprehension": null,
"safety": null,
"symbolic problem solving": null,
"world knowledge": null
} | null | null | null | null | null | null | null | null | null | null |
No dataset card yet