mt5-small-dequad-qg / eval /metric.first.answer.sentence_answer.question.asahi417_qg_dequad.default.json
asahi417's picture
model update
f90d0a0
raw
history blame
535 Bytes
{"validation": {"Bleu_1": 0.10472521828453493, "Bleu_2": 0.043987077697093833, "Bleu_3": 0.018588564837444123, "Bleu_4": 8.406999858284263e-07, "METEOR": 0.10527025066094915, "ROUGE_L": 0.10249577005998024, "BERTScore": 0.7915066623157685, "MoverScore": 0.5423271226113823}, "test": {"Bleu_1": 0.0959904009598992, "Bleu_2": 0.0409551476691875, "Bleu_3": 0.01740277535760309, "Bleu_4": 0.00629917594816096, "METEOR": 0.10686879681599899, "ROUGE_L": 0.09762826749303678, "BERTScore": 0.785173720678487, "MoverScore": 0.5406424401145723}}