AugTriever-QExt-PLM / eval_results.txt
memray's picture
Upload 130 files
2fbbe0b
eval/beir-arguana_ndcg@10 = 0.39138
eval/beir-arguana_recall@100 = 0.93599
eval/beir-avg_ndcg@10 = 0.349554
eval/beir-avg_recall@10 = 0.412942
eval/beir-avg_recall@100 = 0.636488
eval/beir-avg_recall@20 = 0.4818140000000001
eval/beir-climate-fever_ndcg@10 = 0.17682
eval/beir-climate-fever_recall@10 = 0.21882
eval/beir-climate-fever_recall@100 = 0.46955
eval/beir-climate-fever_recall@20 = 0.28607
eval/beir-cqadupstack_ndcg@10 = 0.27104833333333334
eval/beir-cqadupstack_recall@100 = 0.5892308333333333
eval/beir-dbpedia-entity_ndcg@10 = 0.28985
eval/beir-dbpedia-entity_recall@10 = 0.19138
eval/beir-dbpedia-entity_recall@100 = 0.44355
eval/beir-dbpedia-entity_recall@20 = 0.25266
eval/beir-fever_ndcg@10 = 0.59681
eval/beir-fever_recall@10 = 0.78033
eval/beir-fever_recall@100 = 0.91255
eval/beir-fever_recall@20 = 0.84084
eval/beir-fiqa_ndcg@10 = 0.22343
eval/beir-fiqa_recall@100 = 0.53699
eval/beir-hotpotqa_ndcg@10 = 0.47864
eval/beir-hotpotqa_recall@10 = 0.51803
eval/beir-hotpotqa_recall@100 = 0.68575
eval/beir-hotpotqa_recall@20 = 0.57259
eval/beir-msmarco_ndcg@10 = 0.20565
eval/beir-msmarco_recall@10 = 0.35615
eval/beir-msmarco_recall@100 = 0.67104
eval/beir-msmarco_recall@20 = 0.45691
eval/beir-nfcorpus_ndcg@10 = 0.30319
eval/beir-nfcorpus_recall@100 = 0.27586
eval/beir-nq_ndcg@10 = 0.27168
eval/beir-nq_recall@100 = 0.79179
eval/beir-quora_ndcg@10 = 0.82685
eval/beir-quora_recall@100 = 0.98627
eval/beir-scidocs_ndcg@10 = 0.14703
eval/beir-scidocs_recall@100 = 0.35427
eval/beir-scifact_ndcg@10 = 0.64459
eval/beir-scifact_recall@100 = 0.93422
eval/beir-trec-covid_ndcg@10 = 0.53589
eval/beir-trec-covid_recall@100 = 0.09788
eval/beir-webis-touche2020_ndcg@10 = 0.18996
eval/beir-webis-touche2020_recall@100 = 0.44399
eval/qa-curatedtrec-test-acc@100 = 0.930835734870317
eval/qa-curatedtrec-test-acc@20 = 0.840057636887608
eval/qa-curatedtrec-test-acc@5 = 0.6556195965417867
eval/qa-entityqs-macro-acc@100 = 0.7631452687244913
eval/qa-entityqs-macro-acc@20 = 0.6435620051965317
eval/qa-entityqs-macro-acc@5 = 0.5091246348392541
eval/qa-nq-test-acc@100 = 0.7994459833795013
eval/qa-nq-test-acc@20 = 0.6518005540166205
eval/qa-nq-test-acc@5 = 0.4523545706371191
eval/qa-squad1-test-acc@100 = 0.7874172185430464
eval/qa-squad1-test-acc@20 = 0.6373699148533586
eval/qa-squad1-test-acc@5 = 0.4470198675496689
eval/qa-trivia-test-acc@100 = 0.8328471669760452
eval/qa-trivia-test-acc@20 = 0.7477238575090603
eval/qa-trivia-test-acc@5 = 0.6126580040661186
eval/qa-webq-test-acc@100 = 0.8134842519685039
eval/qa-webq-test-acc@20 = 0.6806102362204725
eval/qa-webq-test-acc@5 = 0.45718503937007876
eval/senteval-CR = 88.2
eval/senteval-MPQA = 89.61
eval/senteval-MR = 82.31
eval/senteval-MRPC = 71.1
eval/senteval-SICKRelatedness = 0.6699463146611633
eval/senteval-SST2 = 85.55
eval/senteval-STS12 = 0.5800298176889247
eval/senteval-STS13 = 0.7277593282424332
eval/senteval-STS14 = 0.6560249452737311
eval/senteval-STS15 = 0.781469498622039
eval/senteval-STS16 = 0.7815478011237279
eval/senteval-STSBenchmark = 0.731134165691079
eval/senteval-SUBJ = 96.19
eval/senteval-TREC = 84.13
eval/senteval-avg_sts_7 = 0.7039874101861568
eval/senteval-avg_transfer = 85.29857142857144
train/global_step = 100000