tabedini commited on
Commit
4b74893
1 Parent(s): cb9dff8

Update utils.py

Browse files
Files changed (1) hide show
  1. utils.py +1 -1
utils.py CHANGED
@@ -136,7 +136,7 @@ LLM_BENCHMARKS_ABOUT_TEXT = f"""
136
  > A sample of the evaluation dataset is hosted on [Hugging Face Datasets](https://huggingface.co/datasets/PartAI/llm-leaderboard-datasets-sample), offering the AI community a glimpse of the benchmark content and format. This sample allows developers to pre-assess their models against representative data before a full leaderboard evaluation.
137
  >
138
  > 4. **Collaborative Development**
139
- > This leaderboard represents a significant collaboration between Part AI and Professor Saeedeh Momtazi of Amirkabir University of Technology, leveraging industrial expertise and academic research to create a high-quality, open benchmarking tool. The partnership underscores a shared commitment to advancing Persian LLMs.
140
  >
141
  > 5. **Comprehensive Evaluation Pipeline**
142
  > By integrating a standardized evaluation pipeline, models are assessed across a variety of data types, including text, mathematical formulas, and numerical data. This multi-faceted approach enhances the evaluation’s reliability and allows for precise, nuanced assessment of model performance across multiple dimensions.
 
136
  > A sample of the evaluation dataset is hosted on [Hugging Face Datasets](https://huggingface.co/datasets/PartAI/llm-leaderboard-datasets-sample), offering the AI community a glimpse of the benchmark content and format. This sample allows developers to pre-assess their models against representative data before a full leaderboard evaluation.
137
  >
138
  > 4. **Collaborative Development**
139
+ > This leaderboard represents a significant collaboration between Part AI and Professor Saeedeh Momtazi of Amirkabir University of Technology (with key contributions from [Shahriar Shariati](https://huggingface.co/shahriarshm) and [Farhan Farsi](https://huggingface.co/FarhanFarsi)), leveraging industrial expertise and academic research to create a high-quality, open benchmarking tool. The partnership underscores a shared commitment to advancing Persian LLMs.
140
  >
141
  > 5. **Comprehensive Evaluation Pipeline**
142
  > By integrating a standardized evaluation pipeline, models are assessed across a variety of data types, including text, mathematical formulas, and numerical data. This multi-faceted approach enhances the evaluation’s reliability and allows for precise, nuanced assessment of model performance across multiple dimensions.