Update README.md
Browse files
README.md
CHANGED
@@ -3,4 +3,36 @@ datasets:
|
|
3 |
- nilq/babylm-100M
|
4 |
language:
|
5 |
- en
|
6 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
- nilq/babylm-100M
|
4 |
language:
|
5 |
- en
|
6 |
+
---
|
7 |
+
|
8 |
+
- GPT-2 model submitted by team CLAUSE Bielefeld to the BabyLM challenge 2023
|
9 |
+
- implements a very naive curriculum learning approach inspired by usage-based linguistics: training examples are ordered according to complexity measures from research on child-directed speech (please consult paper for more info)
|
10 |
+
|
11 |
+
Citation:
|
12 |
+
```
|
13 |
+
@inproceedings{bunzeck-zarriess-2023-gpt,
|
14 |
+
title = "{GPT}-wee: How Small Can a Small Language Model Really Get?",
|
15 |
+
author = "Bunzeck, Bastian and
|
16 |
+
Zarrie{\ss}, Sina",
|
17 |
+
editor = "Warstadt, Alex and
|
18 |
+
Mueller, Aaron and
|
19 |
+
Choshen, Leshem and
|
20 |
+
Wilcox, Ethan and
|
21 |
+
Zhuang, Chengxu and
|
22 |
+
Ciro, Juan and
|
23 |
+
Mosquera, Rafael and
|
24 |
+
Paranjabe, Bhargavi and
|
25 |
+
Williams, Adina and
|
26 |
+
Linzen, Tal and
|
27 |
+
Cotterell, Ryan",
|
28 |
+
booktitle = "Proceedings of the BabyLM Challenge at the 27th Conference on Computational Natural Language Learning",
|
29 |
+
month = dec,
|
30 |
+
year = "2023",
|
31 |
+
address = "Singapore",
|
32 |
+
publisher = "Association for Computational Linguistics",
|
33 |
+
url = "https://aclanthology.org/2023.conll-babylm.2",
|
34 |
+
doi = "10.18653/v1/2023.conll-babylm.2",
|
35 |
+
pages = "35--46",
|
36 |
+
}
|
37 |
+
|
38 |
+
```
|