bbunzeck commited on
Commit
76d971f
1 Parent(s): 8b56581

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -1
README.md CHANGED
@@ -3,4 +3,36 @@ datasets:
3
  - nilq/babylm-100M
4
  language:
5
  - en
6
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  - nilq/babylm-100M
4
  language:
5
  - en
6
+ ---
7
+
8
+ - GPT-2 model submitted by team CLAUSE Bielefeld to the BabyLM challenge 2023
9
+ - implements a very naive curriculum learning approach inspired by usage-based linguistics: training examples are ordered according to complexity measures from research on child-directed speech (please consult paper for more info)
10
+
11
+ Citation:
12
+ ```
13
+ @inproceedings{bunzeck-zarriess-2023-gpt,
14
+ title = "{GPT}-wee: How Small Can a Small Language Model Really Get?",
15
+ author = "Bunzeck, Bastian and
16
+ Zarrie{\ss}, Sina",
17
+ editor = "Warstadt, Alex and
18
+ Mueller, Aaron and
19
+ Choshen, Leshem and
20
+ Wilcox, Ethan and
21
+ Zhuang, Chengxu and
22
+ Ciro, Juan and
23
+ Mosquera, Rafael and
24
+ Paranjabe, Bhargavi and
25
+ Williams, Adina and
26
+ Linzen, Tal and
27
+ Cotterell, Ryan",
28
+ booktitle = "Proceedings of the BabyLM Challenge at the 27th Conference on Computational Natural Language Learning",
29
+ month = dec,
30
+ year = "2023",
31
+ address = "Singapore",
32
+ publisher = "Association for Computational Linguistics",
33
+ url = "https://aclanthology.org/2023.conll-babylm.2",
34
+ doi = "10.18653/v1/2023.conll-babylm.2",
35
+ pages = "35--46",
36
+ }
37
+
38
+ ```