|
# Comma v0.1 dataset |
|
|
|
This repository contains the dataset used to train [Comma v0.1-1T](https://huggingface.co/common-pile/comma-v0.1-1t) and [Comma v0.1-2T](https://huggingface.co/common-pile/comma-v0.1-2t). |
|
It is a slightly modified and consolidated version of the [Common Pile v0.1 "filtered" data](https://huggingface.co/collections/common-pile/common-pile-v01-filtered-data-68300bb0a946d10dda697663). |
|
If you are looknig for the raw Common Pile v0.1 data, please see [this collection](https://huggingface.co/collections/common-pile/common-pile-v01-raw-data-6826b454a5a6a445d0b51b37). |
|
You can learn more about Common Pile in [our paper](https://huggingface.co/papers/2506.05209). |
|
|
|
## Mixing rates and token counts |
|
|
|
The Comma v0.1 models were trained in two stages, a "main" stage and a "cooldown" stage. |
|
During each stage, we heuristically set mixing rates to up or downweight different sources. |
|
In the two tables below, we provide per-source token count, repeat rate, and effective token count (after up/downweighting) for the main and cooldown stage of the Comma v0.1-1T training run. |
|
For the Comma v0.1-2T training run, all sources are repeated 2x as many times in both stages. |
|
Token counts are as provided by the Comma v0.1 tokenizer; using a different tokenizer may change these counts significantly. |
|
|
|
|
|
| Main stage | Tokens (B) | Repeats | Effective tokens (B) | |
|
|-------------------------------|------------|---------|----------------------| |
|
| arxiv_abstracts | 0.57 | 6 | 3.4 | |
|
| arxiv_papers | 6.0 | 6 | 35.8 | |
|
| biodiversity_heritage_library | 9.8 | 0.25 | 2.5 | |
|
| caselaw_access_project | 19.7 | 1 | 19.7 | |
|
| cccc | 15.2 | 6 | 91.4 | |
|
| data_provenance_initiative | 0.92 | 6 | 5.5 | |
|
| doab | 3.0 | 6 | 18.2 | |
|
| foodista | 0.025 | 6 | 0.15 | |
|
| github_archive | 11.0 | 6 | 66.1 | |
|
| library_of_congress | 9.5 | 0.25 | 2.4 | |
|
| libretexts | 0.093 | 6 | 0.56 | |
|
| news | 0.064 | 6 | 0.38 | |
|
| oercommons | 0.012 | 6 | 0.07 | |
|
| peS2o | 43.3 | 6 | 260.0 | |
|
| pre_1929_books | 12.4 | 1 | 12.4 | |
|
| pressbooks | 0.14 | 6 | 0.86 | |
|
| project_gutenberg | 5.7 | 1 | 5.7 | |
|
| public_domain_review | 0.0017 | 6 | 0.010 | |
|
| pubmed | 36.6 | 1 | 36.6 | |
|
| python_enhancement_proposals | 0.0027 | 6 | 0.016 | |
|
| regulations | 1.4 | 6 | 8.2 | |
|
| stackexchange | 23.9 | 6 | 143.2 | |
|
| stackv2_edu | 67.8 | 2 | 135.5 | |
|
| stackv2_html | 1.2 | 2 | 2.5 | |
|
| ubuntu_irc | 1.9 | 6 | 11.1 | |
|
| uk_hansard | 2.3 | 6 | 14.0 | |
|
| usgpo | 8.8 | 0.25 | 2.2 | |
|
| uspto | 157.4 | 0.25 | 39.4 | |
|
| wikimedia | 15.8 | 6 | 94.7 | |
|
| wikiteam | 4.3 | 4 | 17.2 | |
|
| youtube | 4.7 | 1 | 4.7 | |
|
| Total | 463.6 | | 1034.4 | |
|
|
|
| Cooldown stage | Tokens (B) | Repeats | Effective tokens (B) | |
|
|------------------------------|------------|---------|----------------------| |
|
| arxiv_papers | 6.0 | 0.5 | 3.0 | |
|
| cccc | 15.2 | 0.3 | 4.6 | |
|
| data_provenance_initiative | 0.92 | 2 | 1.8 | |
|
| doab | 3.0 | 2 | 6.1 | |
|
| foodista | 0.025 | 2 | 0.05 | |
|
| libretexts | 0.093 | 2 | 0.19 | |
|
| news | 0.064 | 2 | 0.13 | |
|
| oercommons | 0.012 | 2 | 0.02 | |
|
| peS2o | 43.3 | 0.1 | 4.3 | |
|
| pressbooks | 0.14 | 2 | 0.29 | |
|
| public_domain_review | 0.0017 | 2 | 0.003 | |
|
| python_enhancement_proposals | 0.0027 | 2 | 0.005 | |
|
| stackexchange | 23.9 | 0.25 | 6.0 | |
|
| stackv2_edu | 67.8 | 0.1 | 6.8 | |
|
| wikimedia | 15.8 | 0.4 | 6.3 | |
|
| Total | 176.2 | | 39.5 | |