Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
https://arxiv.org/abs/2402.00159