File size: 2,654 Bytes
d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed d30410b e8d24ed |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
from pathlib import Path
# Directory where request by models are stored
DIR_OUTPUT_REQUESTS = Path("requested_models")
EVAL_REQUESTS_PATH = Path("eval_requests")
##########################
# Text definitions #
##########################
banner_url = "https://huggingface.co/datasets/vargha/persian_asr_leaderboard/resolve/main/banner.png"
BANNER = f'<div style="display: flex; justify-content: space-around;"><img src="{banner_url}" alt="Banner" style="width: 40vw; min-width: 300px; max-width: 600px;"> </div>'
INTRODUCTION_TEXT = "π The π€ Persian ASR Leaderboard ranks and evaluates speech recognition models \
on the Hugging Face Hub using the Persian Common Voice dataset. \
\nWe report the [WER](https://huggingface.co/spaces/evaluate-metric/wer) and [CER](https://huggingface.co/spaces/evaluate-metric/cer) metrics (β¬οΈ lower the better). Models are ranked based on their WER, from lowest to highest. Check the π Metrics tab to understand how the models are evaluated. \
\nIf you want results for a model that is not listed here, you can submit a request for it to be included βοΈβ¨."
CITATION_TEXT = """@misc{persian-asr-leaderboard,
title = {Persian Automatic Speech Recognition Leaderboard},
author = {Your Name},
year = 2024,
publisher = {Hugging Face},
howpublished = "\\url{https://huggingface.co/spaces/your-username/persian_asr_leaderboard}"
}
"""
METRICS_TAB_TEXT = """
# Metrics and Dataset
## Metrics
We evaluate models using the Word Error Rate (WER) and Character Error Rate (CER) metrics. Both metrics are used to measure the accuracy of automatic speech recognition systems.
- **Word Error Rate (WER)**: Calculates the percentage of words that were incorrectly predicted. A lower WER indicates better performance.
- **Character Error Rate (CER)**: Similar to WER but operates at the character level, which can be more informative for languages with rich morphology like Persian.
## Dataset
We use the [Persian Common Voice](https://huggingface.co/datasets/vargha/common_voice_fa) dataset for evaluation. The dataset consists of diverse speech recordings from various speakers, making it a good benchmark for Persian ASR models.
## How to Submit Your Model
To submit your model for evaluation, go to the "βοΈβ¨ Request a model here!" tab and enter your model's name in the format `username/model_name`. Your model should be hosted on the Hugging Face Hub.
## Reproducing Results
To reproduce the results or to see how the evaluation is conducted, you can visit our [GitHub repository](https://github.com/your-username/persian_asr_leaderboard).
"""
|