eval/beir-arguana_ndcg@10 = 0.39579 eval/beir-arguana_recall@100 = 0.94168 eval/beir-avg_ndcg@10 = 0.342446 eval/beir-avg_recall@10 = 0.40421200000000007 eval/beir-avg_recall@100 = 0.624188 eval/beir-avg_recall@20 = 0.473074 eval/beir-climate-fever_ndcg@10 = 0.1562 eval/beir-climate-fever_recall@10 = 0.19275 eval/beir-climate-fever_recall@100 = 0.42836 eval/beir-climate-fever_recall@20 = 0.25077 eval/beir-cqadupstack_ndcg@10 = 0.2689125 eval/beir-cqadupstack_recall@100 = 0.6019383333333334 eval/beir-dbpedia-entity_ndcg@10 = 0.29269 eval/beir-dbpedia-entity_recall@10 = 0.19543 eval/beir-dbpedia-entity_recall@100 = 0.43193 eval/beir-dbpedia-entity_recall@20 = 0.26224 eval/beir-fever_ndcg@10 = 0.55206 eval/beir-fever_recall@10 = 0.73955 eval/beir-fever_recall@100 = 0.89197 eval/beir-fever_recall@20 = 0.79964 eval/beir-fiqa_ndcg@10 = 0.25367 eval/beir-fiqa_recall@100 = 0.59399 eval/beir-hotpotqa_ndcg@10 = 0.49581 eval/beir-hotpotqa_recall@10 = 0.52519 eval/beir-hotpotqa_recall@100 = 0.68339 eval/beir-hotpotqa_recall@20 = 0.57731 eval/beir-msmarco_ndcg@10 = 0.21547 eval/beir-msmarco_recall@10 = 0.36814 eval/beir-msmarco_recall@100 = 0.68529 eval/beir-msmarco_recall@20 = 0.47541 eval/beir-nfcorpus_ndcg@10 = 0.29655 eval/beir-nfcorpus_recall@100 = 0.28415 eval/beir-nq_ndcg@10 = 0.2742 eval/beir-nq_recall@100 = 0.79582 eval/beir-quora_ndcg@10 = 0.77792 eval/beir-quora_recall@100 = 0.97682 eval/beir-scidocs_ndcg@10 = 0.15413 eval/beir-scidocs_recall@100 = 0.37255 eval/beir-scifact_ndcg@10 = 0.65609 eval/beir-scifact_recall@100 = 0.91156 eval/beir-trec-covid_ndcg@10 = 0.56947 eval/beir-trec-covid_recall@100 = 0.11015 eval/beir-webis-touche2020_ndcg@10 = 0.18628 eval/beir-webis-touche2020_recall@100 = 0.4094 eval/qa-curatedtrec-test-acc@100 = 0.9236311239193083 eval/qa-curatedtrec-test-acc@20 = 0.8386167146974063 eval/qa-curatedtrec-test-acc@5 = 0.6455331412103746 eval/qa-entityqs-macro-acc@100 = 0.7569173874462057 eval/qa-entityqs-macro-acc@20 = 0.635020152364218 eval/qa-entityqs-macro-acc@5 = 0.4987727872796979 eval/qa-nq-test-acc@100 = 0.7894736842105263 eval/qa-nq-test-acc@20 = 0.6429362880886427 eval/qa-nq-test-acc@5 = 0.4440443213296399 eval/qa-squad1-test-acc@100 = 0.7674550614947966 eval/qa-squad1-test-acc@20 = 0.6085146641438032 eval/qa-squad1-test-acc@5 = 0.4126773888363292 eval/qa-trivia-test-acc@100 = 0.8330239547423318 eval/qa-trivia-test-acc@20 = 0.7475470697427739 eval/qa-trivia-test-acc@5 = 0.601962344205781 eval/qa-webq-test-acc@100 = 0.8080708661417323 eval/qa-webq-test-acc@20 = 0.6712598425196851 eval/qa-webq-test-acc@5 = 0.46013779527559057 eval/senteval-CR = 88.1 eval/senteval-MPQA = 89.04 eval/senteval-MR = 80.83 eval/senteval-MRPC = 70.68 eval/senteval-SICKRelatedness = 0.6952695805138401 eval/senteval-SST2 = 84.06 eval/senteval-STS12 = 0.6163833612367604 eval/senteval-STS13 = 0.7419430794508367 eval/senteval-STS14 = 0.6677462039090557 eval/senteval-STS15 = 0.7899640319998164 eval/senteval-STS16 = 0.7846388532549995 eval/senteval-STSBenchmark = 0.7564014032837215 eval/senteval-SUBJ = 95.32 eval/senteval-TREC = 81.07 eval/senteval-avg_sts_7 = 0.7217637876641472 eval/senteval-avg_transfer = 84.15714285714287 train/global_step = 100000