--- language: en license: apache-2.0 datasets: - wikipedia --- # BERT Large Uncased (CDA) - Counterfactual Data Augmentation Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1810.04805) and first released in [this repository](https://github.com/google-research-datasets/Zari). The model is pre-trained from scratch over Wikipedia. Word substitutions for data augmentation are determined using the word lists provided at [corefBias](https://github.com/uclanlp/corefBias) ([Zhao et al. (2018)](https://arxiv.org/abs/1804.06876)). Disclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by the FairNLP team. ### BibTeX entry and citation info ``` @misc{zari, title={Measuring and Reducing Gendered Correlations in Pre-trained Models}, author={Kellie Webster and Xuezhi Wang and Ian Tenney and Alex Beutel and Emily Pitler and Ellie Pavlick and Jilin Chen and Slav Petrov}, year={2020}, eprint={2010.06032}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```