craffel's picture
craffel HF Staff
Update README.md
5afc546 verified
# Comma v0.1 dataset
This repository contains the dataset used to train [Comma v0.1-1T](https://huggingface.co/common-pile/comma-v0.1-1t) and [Comma v0.1-2T](https://huggingface.co/common-pile/comma-v0.1-2t).
It is a slightly modified and consolidated version of the [Common Pile v0.1 "filtered" data](https://huggingface.co/collections/common-pile/common-pile-v01-filtered-data-68300bb0a946d10dda697663).
If you are looknig for the raw Common Pile v0.1 data, please see [this collection](https://huggingface.co/collections/common-pile/common-pile-v01-raw-data-6826b454a5a6a445d0b51b37).
You can learn more about Common Pile in [our paper](https://huggingface.co/papers/2506.05209).
## Mixing rates and token counts
The Comma v0.1 models were trained in two stages, a "main" stage and a "cooldown" stage.
During each stage, we heuristically set mixing rates to up or downweight different sources.
In the two tables below, we provide per-source token count, repeat rate, and effective token count (after up/downweighting) for the main and cooldown stage of the Comma v0.1-1T training run.
For the Comma v0.1-2T training run, all sources are repeated 2x as many times in both stages.
Token counts are as provided by the Comma v0.1 tokenizer; using a different tokenizer may change these counts significantly.
| Main stage | Tokens (B) | Repeats | Effective tokens (B) |
|-------------------------------|------------|---------|----------------------|
| arxiv_abstracts | 0.57 | 6 | 3.4 |
| arxiv_papers | 6.0 | 6 | 35.8 |
| biodiversity_heritage_library | 9.8 | 0.25 | 2.5 |
| caselaw_access_project | 19.7 | 1 | 19.7 |
| cccc | 15.2 | 6 | 91.4 |
| data_provenance_initiative | 0.92 | 6 | 5.5 |
| doab | 3.0 | 6 | 18.2 |
| foodista | 0.025 | 6 | 0.15 |
| github_archive | 11.0 | 6 | 66.1 |
| library_of_congress | 9.5 | 0.25 | 2.4 |
| libretexts | 0.093 | 6 | 0.56 |
| news | 0.064 | 6 | 0.38 |
| oercommons | 0.012 | 6 | 0.07 |
| peS2o | 43.3 | 6 | 260.0 |
| pre_1929_books | 12.4 | 1 | 12.4 |
| pressbooks | 0.14 | 6 | 0.86 |
| project_gutenberg | 5.7 | 1 | 5.7 |
| public_domain_review | 0.0017 | 6 | 0.010 |
| pubmed | 36.6 | 1 | 36.6 |
| python_enhancement_proposals | 0.0027 | 6 | 0.016 |
| regulations | 1.4 | 6 | 8.2 |
| stackexchange | 23.9 | 6 | 143.2 |
| stackv2_edu | 67.8 | 2 | 135.5 |
| stackv2_html | 1.2 | 2 | 2.5 |
| ubuntu_irc | 1.9 | 6 | 11.1 |
| uk_hansard | 2.3 | 6 | 14.0 |
| usgpo | 8.8 | 0.25 | 2.2 |
| uspto | 157.4 | 0.25 | 39.4 |
| wikimedia | 15.8 | 6 | 94.7 |
| wikiteam | 4.3 | 4 | 17.2 |
| youtube | 4.7 | 1 | 4.7 |
| Total | 463.6 | | 1034.4 |
| Cooldown stage | Tokens (B) | Repeats | Effective tokens (B) |
|------------------------------|------------|---------|----------------------|
| arxiv_papers | 6.0 | 0.5 | 3.0 |
| cccc | 15.2 | 0.3 | 4.6 |
| data_provenance_initiative | 0.92 | 2 | 1.8 |
| doab | 3.0 | 2 | 6.1 |
| foodista | 0.025 | 2 | 0.05 |
| libretexts | 0.093 | 2 | 0.19 |
| news | 0.064 | 2 | 0.13 |
| oercommons | 0.012 | 2 | 0.02 |
| peS2o | 43.3 | 0.1 | 4.3 |
| pressbooks | 0.14 | 2 | 0.29 |
| public_domain_review | 0.0017 | 2 | 0.003 |
| python_enhancement_proposals | 0.0027 | 2 | 0.005 |
| stackexchange | 23.9 | 0.25 | 6.0 |
| stackv2_edu | 67.8 | 0.1 | 6.8 |
| wikimedia | 15.8 | 0.4 | 6.3 |
| Total | 176.2 | | 39.5 |