omarmomen commited on
Commit
59d81d5
1 Parent(s): abf5c11

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -1
README.md CHANGED
@@ -1 +1,20 @@
1
- Submitted Model to BabyLM Shared Task
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - omarmomen/babylm_10M
5
+ language:
6
+ - en
7
+ metrics:
8
+ - perplexity
9
+ library_name: transformers
10
+ ---
11
+ # Model Card for omarmomen/structroberta_sx_final
12
+
13
+ This model is part of the experiments in the published paper at the BabyLM workshop in CoNLL 2023.
14
+ The paper titled "Increasing The Performance of Cognitively Inspired Data-Efficient Language Models via Implicit Structure Building" (https://aclanthology.org/2023.conll-babylm.29/)
15
+
16
+ <strong>omarmomen/structroberta_sx_final</strong> is a modification on the Roberta Model to incorporate syntactic inductive bias using an unsupervised parsing mechanism.
17
+
18
+ This model variant places the parser network ahead of all attention blocks, and increase the number of convolution layers from 4 to 6.
19
+
20
+ The model is pretrained on the BabyLM 10M dataset using a custom pretrained RobertaTokenizer (https://huggingface.co/omarmomen/babylm_tokenizer_32k).