eval/beir-arguana_ndcg@10 = 0.39138 eval/beir-arguana_recall@100 = 0.93599 eval/beir-avg_ndcg@10 = 0.349554 eval/beir-avg_recall@10 = 0.412942 eval/beir-avg_recall@100 = 0.636488 eval/beir-avg_recall@20 = 0.4818140000000001 eval/beir-climate-fever_ndcg@10 = 0.17682 eval/beir-climate-fever_recall@10 = 0.21882 eval/beir-climate-fever_recall@100 = 0.46955 eval/beir-climate-fever_recall@20 = 0.28607 eval/beir-cqadupstack_ndcg@10 = 0.27104833333333334 eval/beir-cqadupstack_recall@100 = 0.5892308333333333 eval/beir-dbpedia-entity_ndcg@10 = 0.28985 eval/beir-dbpedia-entity_recall@10 = 0.19138 eval/beir-dbpedia-entity_recall@100 = 0.44355 eval/beir-dbpedia-entity_recall@20 = 0.25266 eval/beir-fever_ndcg@10 = 0.59681 eval/beir-fever_recall@10 = 0.78033 eval/beir-fever_recall@100 = 0.91255 eval/beir-fever_recall@20 = 0.84084 eval/beir-fiqa_ndcg@10 = 0.22343 eval/beir-fiqa_recall@100 = 0.53699 eval/beir-hotpotqa_ndcg@10 = 0.47864 eval/beir-hotpotqa_recall@10 = 0.51803 eval/beir-hotpotqa_recall@100 = 0.68575 eval/beir-hotpotqa_recall@20 = 0.57259 eval/beir-msmarco_ndcg@10 = 0.20565 eval/beir-msmarco_recall@10 = 0.35615 eval/beir-msmarco_recall@100 = 0.67104 eval/beir-msmarco_recall@20 = 0.45691 eval/beir-nfcorpus_ndcg@10 = 0.30319 eval/beir-nfcorpus_recall@100 = 0.27586 eval/beir-nq_ndcg@10 = 0.27168 eval/beir-nq_recall@100 = 0.79179 eval/beir-quora_ndcg@10 = 0.82685 eval/beir-quora_recall@100 = 0.98627 eval/beir-scidocs_ndcg@10 = 0.14703 eval/beir-scidocs_recall@100 = 0.35427 eval/beir-scifact_ndcg@10 = 0.64459 eval/beir-scifact_recall@100 = 0.93422 eval/beir-trec-covid_ndcg@10 = 0.53589 eval/beir-trec-covid_recall@100 = 0.09788 eval/beir-webis-touche2020_ndcg@10 = 0.18996 eval/beir-webis-touche2020_recall@100 = 0.44399 eval/qa-curatedtrec-test-acc@100 = 0.930835734870317 eval/qa-curatedtrec-test-acc@20 = 0.840057636887608 eval/qa-curatedtrec-test-acc@5 = 0.6556195965417867 eval/qa-entityqs-macro-acc@100 = 0.7631452687244913 eval/qa-entityqs-macro-acc@20 = 0.6435620051965317 eval/qa-entityqs-macro-acc@5 = 0.5091246348392541 eval/qa-nq-test-acc@100 = 0.7994459833795013 eval/qa-nq-test-acc@20 = 0.6518005540166205 eval/qa-nq-test-acc@5 = 0.4523545706371191 eval/qa-squad1-test-acc@100 = 0.7874172185430464 eval/qa-squad1-test-acc@20 = 0.6373699148533586 eval/qa-squad1-test-acc@5 = 0.4470198675496689 eval/qa-trivia-test-acc@100 = 0.8328471669760452 eval/qa-trivia-test-acc@20 = 0.7477238575090603 eval/qa-trivia-test-acc@5 = 0.6126580040661186 eval/qa-webq-test-acc@100 = 0.8134842519685039 eval/qa-webq-test-acc@20 = 0.6806102362204725 eval/qa-webq-test-acc@5 = 0.45718503937007876 eval/senteval-CR = 88.2 eval/senteval-MPQA = 89.61 eval/senteval-MR = 82.31 eval/senteval-MRPC = 71.1 eval/senteval-SICKRelatedness = 0.6699463146611633 eval/senteval-SST2 = 85.55 eval/senteval-STS12 = 0.5800298176889247 eval/senteval-STS13 = 0.7277593282424332 eval/senteval-STS14 = 0.6560249452737311 eval/senteval-STS15 = 0.781469498622039 eval/senteval-STS16 = 0.7815478011237279 eval/senteval-STSBenchmark = 0.731134165691079 eval/senteval-SUBJ = 96.19 eval/senteval-TREC = 84.13 eval/senteval-avg_sts_7 = 0.7039874101861568 eval/senteval-avg_transfer = 85.29857142857144 train/global_step = 100000