--- task_categories: - text-generation language: - en size_categories: - 1B The following data mix was used to train OLMoE-1B-7B, a Mixture-of-Experts LLM with 1B active and 7B total parameters released in September 2024. The base version of OLMoE-1B-7B can be found at [this page](https://huggingface.co/allenai/OLMoE-1B-7B-0924), the SFT of OLMoE-1B-7B is available [here](https://huggingface.co/allenai/OLMoE-1B-7B-0924-SFT), and a version combining SFT and DPO is available following [this link](https://huggingface.co/allenai/OLMoE-1B-7B-0924-Instruct). ## Statistics | Subset | Tokens | Words | Bytes | Docs | |--------------------------------------------------------------|:----------:|:----------:|:----------:|:----------:| | [DCLM Baseline 1.0](https://huggingface.co/datasets/mlfoundations/dclm-baseline-1.0) | 3.86 T | 3.38 T | 16.7 T | 2.95 B | | [Starcoder](https://huggingface.co/datasets/bigcode/starcoderdata) | 101 B | 63.9 B | 325 B | 78.7 M | | [peS2o](https://huggingface.co/datasets/allenai/peS2o)
([Dolma](https://huggingface.co/datasets/allenai/dolma)) | 57.2 B | 51.3 B | 268 B | 38.8 M | | Arxiv
([RedPajama v1](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T)
via [Proof Pile II](https://huggingface.co/datasets/EleutherAI/proof-pile-2)) | 21.1 B | 23.5 B | 88.8 B | 1.55 M | | OpenWebMath
([Proof Pile II](https://huggingface.co/datasets/EleutherAI/proof-pile-2)) | 12.7 B | 10.2 B | 42.4 B | 2.91 M | | Algebraic Stack
([Proof Pile II](https://huggingface.co/datasets/EleutherAI/proof-pile-2)) | 12.6 B | 9.6 B | 39.3 B | 2.83 M | | En Wikipedia +
Wikibooks
([Dolma](https://huggingface.co/datasets/allenai/dolma)) | 3.69 B | 3.16 B | 16.2 B | 6.17 M | | **Total** | **4.07 T** | **3.53 T** | **17.4 T** | **3.08 B** | ## Preprocessing All subsets were pre-processed to remove documents with a *sequence* of 32 or more repeated *ngrams*. - a *ngram* is a span of 1 to 13 tokens, included; - *tokens* are obtained using the model tokenizer; - a *sequence* is a contiguous span of repeated ngrams. In addition of the above, Starcoder dataset was further processed by removing any document meeting any of the following rules: - document is from a repository with fewer than 2 stars on GitHub; - the top most frequent word in the document constitutes over 30% of the document; - the two most frequent words in the document constitutes over 50% of the document. ## Licensing Information This mix is licensed under [Open Data Commons Attribution License (ODC-By) v1.0](https://opendatacommons.org/licenses/by/1-0/). By using this dataset, you are bound to licenses and Terms of Services of underlying datasets, which you can access by clicking on the links in the table above. ## Citation ```bibtex @misc{muennighoff2024olmoeopenmixtureofexpertslanguage, title={OLMoE: Open Mixture-of-Experts Language Models}, author={Niklas Muennighoff and Luca Soldaini and Dirk Groeneveld and Kyle Lo and Jacob Morrison and Sewon Min and Weijia Shi and Pete Walsh and Oyvind Tafjord and Nathan Lambert and Yuling Gu and Shane Arora and Akshita Bhagia and Dustin Schwenk and David Wadden and Alexander Wettig and Binyuan Hui and Tim Dettmers and Douwe Kiela and Ali Farhadi and Noah A. Smith and Pang Wei Koh and Amanpreet Singh and Hannaneh Hajishirzi}, year={2024}, eprint={2409.02060}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2409.02060}, } ```