{ "dataset_revision": "70970daeab8776df92f5ea462b6173c0b46fd2d1", "mteb_dataset_name": "TwitterSemEval2015", "mteb_version": "1.1.0", "test": { "cos_sim": { "accuracy": 0.8402574953805806, "accuracy_threshold": 0.7548823356628418, "ap": 0.6782973844761395, "f1": 0.6368255504476167, "f1_threshold": 0.6944434642791748, "precision": 0.5880250223413762, "recall": 0.6944591029023747 }, "dot": { "accuracy": 0.7901889491565834, "accuracy_threshold": 30.103286743164062, "ap": 0.482856775311342, "f1": 0.5083665338645418, "f1_threshold": 20.793228149414062, "precision": 0.40832, "recall": 0.6733509234828496 }, "euclidean": { "accuracy": 0.8328068188591524, "accuracy_threshold": 3.928983211517334, "ap": 0.6539558962841057, "f1": 0.6220163876024225, "f1_threshold": 4.456747055053711, "precision": 0.5655366011660549, "recall": 0.6910290237467018 }, "evaluation_time": 15.28, "manhattan": { "accuracy": 0.8323895809739524, "accuracy_threshold": 60.54985046386719, "ap": 0.652800674337047, "f1": 0.6223784701175901, "f1_threshold": 68.62921142578125, "precision": 0.5756896165059431, "recall": 0.6773087071240106 }, "max": { "accuracy": 0.8402574953805806, "ap": 0.6782973844761395, "f1": 0.6368255504476167 } } }