|
--- |
|
title: README |
|
emoji: 📚 |
|
colorFrom: indigo |
|
colorTo: purple |
|
sdk: static |
|
pinned: false |
|
--- |
|
|
|
# hmTEAMS |
|
|
|
Historical Multilingual TEAMS Models. Following languages are currently covered: |
|
|
|
* English (British Library Corpus - Books) |
|
* German (Europeana Newspaper) |
|
* French (Europeana Newspaper) |
|
* Finnish (Europeana Newspaper, Digilib) |
|
* Swedish (Europeana Newspaper, Digilib) |
|
* Dutch (Delpher Corpus) |
|
* Norwegian (NCC Corpus) |
|
|
|
More details can be found in [our GitHub repository](https://github.com/stefan-it/hmTEAMS). |
|
|
|
# Leaderboard |
|
|
|
We test our pretrained language models on various datasets from HIPE-2020, HIPE-2022 and Europeana. The following table |
|
shows an overview of used datasets. |
|
|
|
|
|
| Language | Datasets | |
|
|----------|------------------------------------------------------------------| |
|
| English | [AjMC] - [TopRes19th] | |
|
| German | [AjMC] - [NewsEye] - [HIPE-2020] | |
|
| French | [AjMC] - [ICDAR-Europeana] - [LeTemps] - [NewsEye] - [HIPE-2020] | |
|
| Finnish | [NewsEye] | |
|
| Swedish | [NewsEye] | |
|
| Dutch | [ICDAR-Europeana] | |
|
|
|
[AjMC]: https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-ajmc.md |
|
[NewsEye]: https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-newseye.md |
|
[TopRes19th]: https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-topres19th.md |
|
[ICDAR-Europeana]: https://github.com/stefan-it/historic-domain-adaptation-icdar |
|
[LeTemps]: https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-letemps.md |
|
[HIPE-2020]: https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-hipe2020.md |
|
|
|
All results can be found in the [`hmLeaderboard`](https://huggingface.co/spaces/hmbench/hmLeaderboard). |
|
|
|
# Acknowledgements |
|
|
|
We thank [Luisa März](https://github.com/LuisaMaerz), [Katharina Schmid](https://github.com/schmika) and |
|
[Erion Çano](https://github.com/erionc) for their fruitful discussions about Historical Language Models. |
|
|
|
Research supported with Cloud TPUs from Google's [TPU Research Cloud](https://sites.research.google/trc/about/) (TRC). |
|
Many Thanks for providing access to the TPUs ❤️ |