--- license: odc-by task_categories: - text-generation language: - en tags: - math - code - pre-training - synthesis size_categories: - 1B Megamath is part of TxT360, curated by LLM360 Team.
MegaMath Collection
We introduce MegaMath, an open math pretraining dataset curated from diverse, math-focused sources, with over 300B tokens. MegaMath is curated via the following three efforts: - **Revisiting web data**: We re-extracted mathematical documents from Common Crawl with math-oriented HTML optimizations, fasttext-based filtering and deduplication, all for acquiring higher-quality data on the Internet. - **Recalling Math-related code data**: We identified high quality math-related code from large code training corpus, Stack-V2, further enhancing data diversity. - **Exploring Synthetic data**: We synthesized QA-style text, math-related code, and interleaved text-code blocks from web data or code data. ## MegaMath Compared to Existing Datasets MegaMath is the largest open math pre-training dataset to date, surpassing DeepSeekMath (120B) tokens.
## MegaMath Delivers with High Quality During development, we use extensive experiments to find optimal practice for text extraction, deduplication, fasttext training, etc. Training MegaMath data shows better performance than existing open datasets.
## Training MegaMath on Latest LMs We also release two proof-of-concept models which is based on [Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) and [LLama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B). Training MegaMath on Llama-3.2-1B and LLama-3.2-3B brings about 15% ~ 20% performance boost on 10 downstream benchmarks, demonstrateing its high data quality.
## Detailed Statistics | **Category** | **# Sample(M)** | **# Toks(B)** | **Avg. (# Toks)** | |------------------------|----------------:|--------------:|------------------:| | **Web Domain** | **121.5** | **279.0** | **2296.9** | | Web | 106.5 | 263.9 | 2478.7 | | Web-Pro | 15.0 | 15.1 | 1006.0 | | **Code Domain** | **13.4** | **28.1** | **2102.7** | | **Synthetic Data** | **80.2** | **64.5** | **804.5** | | Translated Code | 7.4 | 7.2 | 979.5 | | Q&A | 22.6 | 7.0 | 308.3 | | Text&Code Block | 50.2 | 50.3 | 1002.1 | | **Total** | **215.1** | **371.6** | **1727.6** | ## Citation If you use our dataset or find our work useful, please cite ```bibtex @article{zhou2025megamath, title = {MegaMath: Pushing the Limits of Open Math Corpora}, author = {Zhou, Fan and Wang, Zengzhi and Ranjan, Nikhil and Cheng, Zhoujun and Tang, Liping and He, Guowei and Liu, Zhengzhong and Xing, Eric P.}, journal = {arXiv preprint arXiv:2504.02807}, year = {2025}, note = {Preprint} } ```