File size: 1,208 Bytes
ffebbfb
 
42ec372
 
 
ffebbfb
 
42ec372
 
ffebbfb
 
42ec372
 
d7dc735
42ec372
 
4c8cf84
42ec372
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
---
title: README
emoji: 🤗
colorFrom: green
colorTo: purple
sdk: static
pinned: false
tags:
- evaluate
---

🤗 Evaluate provides access to a wide range of evaluation tools. It covers a range of modalities such as text, computer vision, audio, etc. as well as tools to evaluate models or datasets. 


It has three types of evaluations:
- **Metric**: measures the performance of a model on a given dataset, usually by comparing the model's predictions to some ground truth labels -- these are covered in this space.
- **Comparison**: used to compare the performance of two or more models on a single test dataset., e.g. by comparing their predictions to ground truth labels and computing their agreement -- covered in the [Evaluate Comparison](https://huggingface.co/spaces/evaluate-comparison) Spaces.
- **Measurement**: for gaining more insights on datasets and model predictions based on their properties and characteristics -- covered in the [Evaluate Measurement](https://huggingface.co/evaluate-measurement) Spaces.

All three types of evaluation supported by the 🤗 Evaluate library are meant to be mutually complementary, and help our community carry out more mindful and responsible evaluation!