MMFMChallenge / submission_info /a90f08a6-10ec-4e42-aa31-f07264bc6059.json
helloai0's picture
Upload submission_info/a90f08a6-10ec-4e42-aa31-f07264bc6059.json with huggingface_hub
62c9f14 verified
raw
history blame
2.08 kB
{
"id": "a90f08a6-10ec-4e42-aa31-f07264bc6059",
"submissions": [
{
"datetime": "2024-06-05 01:59:37",
"submission_id": "e1750cff-a0c2-433e-a97f-571dbae2df89",
"submission_comment": "",
"submission_repo": "",
"space_id": "",
"submitted_by": "6627884d4cf53462114ba667",
"status": 3,
"selected": false,
"public_score": {
"iconqa_fill_acc": 0.965,
"funsd_acc": 0.865,
"iconqa_choose_acc": 0.865,
"wildreceipt_acc": 0.9,
"textbookqa_acc": 0.675,
"tabfact_acc": 0.705,
"docvqa_acc": 0.785,
"infographicvqa_acc": 0.42,
"websrc_acc": 0.995,
"wtq_acc": 0.455,
"phase1_overall_acc": 0.763,
"mydoc_acc": 0.7325,
"mychart_acc": 0.105,
"myinfographic_acc": 0.61916,
"phase2_overall_acc": 0.56323
},
"private_score": {
"iconqa_fill_acc": 0,
"funsd_acc": 0,
"iconqa_choose_acc": 0,
"wildreceipt_acc": 0,
"textbookqa_acc": 0,
"tabfact_acc": 0,
"docvqa_acc": 0,
"infographicvqa_acc": 0,
"websrc_acc": 0,
"wtq_acc": 0,
"phase1_overall_acc": 0,
"mydoc_acc": 0,
"mychart_acc": 0,
"myinfographic_acc": 0,
"phase2_overall_acc": 0
}
},
{
"datetime": "2024-06-05 02:02:13",
"submission_id": "3cfb5231-ee9d-4a99-ba89-2371ef078066",
"submission_comment": "",
"submission_repo": "",
"space_id": "",
"submitted_by": "6627884d4cf53462114ba667",
"status": 2,
"selected": false,
"public_score": {},
"private_score": {}
}
]
}