AugTriever-TQGen-Title / eval_results.txt
memray's picture
Upload 136 files
ea846ca
eval/beir-arguana_ndcg@10 = 0.39579
eval/beir-arguana_recall@100 = 0.94168
eval/beir-avg_ndcg@10 = 0.342446
eval/beir-avg_recall@10 = 0.40421200000000007
eval/beir-avg_recall@100 = 0.624188
eval/beir-avg_recall@20 = 0.473074
eval/beir-climate-fever_ndcg@10 = 0.1562
eval/beir-climate-fever_recall@10 = 0.19275
eval/beir-climate-fever_recall@100 = 0.42836
eval/beir-climate-fever_recall@20 = 0.25077
eval/beir-cqadupstack_ndcg@10 = 0.2689125
eval/beir-cqadupstack_recall@100 = 0.6019383333333334
eval/beir-dbpedia-entity_ndcg@10 = 0.29269
eval/beir-dbpedia-entity_recall@10 = 0.19543
eval/beir-dbpedia-entity_recall@100 = 0.43193
eval/beir-dbpedia-entity_recall@20 = 0.26224
eval/beir-fever_ndcg@10 = 0.55206
eval/beir-fever_recall@10 = 0.73955
eval/beir-fever_recall@100 = 0.89197
eval/beir-fever_recall@20 = 0.79964
eval/beir-fiqa_ndcg@10 = 0.25367
eval/beir-fiqa_recall@100 = 0.59399
eval/beir-hotpotqa_ndcg@10 = 0.49581
eval/beir-hotpotqa_recall@10 = 0.52519
eval/beir-hotpotqa_recall@100 = 0.68339
eval/beir-hotpotqa_recall@20 = 0.57731
eval/beir-msmarco_ndcg@10 = 0.21547
eval/beir-msmarco_recall@10 = 0.36814
eval/beir-msmarco_recall@100 = 0.68529
eval/beir-msmarco_recall@20 = 0.47541
eval/beir-nfcorpus_ndcg@10 = 0.29655
eval/beir-nfcorpus_recall@100 = 0.28415
eval/beir-nq_ndcg@10 = 0.2742
eval/beir-nq_recall@100 = 0.79582
eval/beir-quora_ndcg@10 = 0.77792
eval/beir-quora_recall@100 = 0.97682
eval/beir-scidocs_ndcg@10 = 0.15413
eval/beir-scidocs_recall@100 = 0.37255
eval/beir-scifact_ndcg@10 = 0.65609
eval/beir-scifact_recall@100 = 0.91156
eval/beir-trec-covid_ndcg@10 = 0.56947
eval/beir-trec-covid_recall@100 = 0.11015
eval/beir-webis-touche2020_ndcg@10 = 0.18628
eval/beir-webis-touche2020_recall@100 = 0.4094
eval/qa-curatedtrec-test-acc@100 = 0.9236311239193083
eval/qa-curatedtrec-test-acc@20 = 0.8386167146974063
eval/qa-curatedtrec-test-acc@5 = 0.6455331412103746
eval/qa-entityqs-macro-acc@100 = 0.7569173874462057
eval/qa-entityqs-macro-acc@20 = 0.635020152364218
eval/qa-entityqs-macro-acc@5 = 0.4987727872796979
eval/qa-nq-test-acc@100 = 0.7894736842105263
eval/qa-nq-test-acc@20 = 0.6429362880886427
eval/qa-nq-test-acc@5 = 0.4440443213296399
eval/qa-squad1-test-acc@100 = 0.7674550614947966
eval/qa-squad1-test-acc@20 = 0.6085146641438032
eval/qa-squad1-test-acc@5 = 0.4126773888363292
eval/qa-trivia-test-acc@100 = 0.8330239547423318
eval/qa-trivia-test-acc@20 = 0.7475470697427739
eval/qa-trivia-test-acc@5 = 0.601962344205781
eval/qa-webq-test-acc@100 = 0.8080708661417323
eval/qa-webq-test-acc@20 = 0.6712598425196851
eval/qa-webq-test-acc@5 = 0.46013779527559057
eval/senteval-CR = 88.1
eval/senteval-MPQA = 89.04
eval/senteval-MR = 80.83
eval/senteval-MRPC = 70.68
eval/senteval-SICKRelatedness = 0.6952695805138401
eval/senteval-SST2 = 84.06
eval/senteval-STS12 = 0.6163833612367604
eval/senteval-STS13 = 0.7419430794508367
eval/senteval-STS14 = 0.6677462039090557
eval/senteval-STS15 = 0.7899640319998164
eval/senteval-STS16 = 0.7846388532549995
eval/senteval-STSBenchmark = 0.7564014032837215
eval/senteval-SUBJ = 95.32
eval/senteval-TREC = 81.07
eval/senteval-avg_sts_7 = 0.7217637876641472
eval/senteval-avg_transfer = 84.15714285714287
train/global_step = 100000