Dolma
DatasetFreeAllen AI's 3T token dataset for fully reproducible LLM training.
Capabilities10 decomposed
multi-source pretraining corpus assembly with documented curation
Medium confidenceAggregates 3 trillion tokens from 7 heterogeneous sources (Common Crawl, The Stack, peS2o, Project Gutenberg, Wikipedia, Wikibooks, C4) into a unified pretraining dataset with published filtering rules, deduplication strategies, and source mixing ratios. The assembly process applies source-specific quality filters and fuzzy deduplication via Duplodocus before combining sources at documented proportions, enabling reproducible dataset composition for LLM training.
Dolma publishes exact filtering rules, deduplication methods (via Duplodocus fuzzy matching), and source mixing ratios alongside the dataset itself, enabling researchers to independently audit and reproduce curation decisions—a level of transparency uncommon in large pretraining corpora where composition details are typically proprietary
More transparent and reproducible than proprietary datasets (GPT-3, Chinchilla) and more comprehensively documented than C4 alone, with explicit multi-source composition and published deduplication strategies
fuzzy deduplication at scale via duplodocus
Medium confidenceApplies ultra-efficient fuzzy deduplication across the 3 trillion token corpus using the Duplodocus tool, which identifies and removes near-duplicate documents within and across source domains without requiring exact string matching. The fuzzy matching approach reduces redundancy while preserving legitimate diversity, operating at scale to handle the full dataset volume without prohibitive computational overhead.
Duplodocus performs fuzzy (approximate) deduplication rather than exact-match deduplication, enabling removal of near-duplicates and paraphrased content while scaling to 3 trillion tokens; most commodity deduplication tools use exact matching or simple hashing, which miss semantic redundancy
More efficient than naive pairwise comparison and more comprehensive than exact-match deduplication, though specific algorithmic advantages over MinHash or LSH-based approaches are not documented
large-scale data cleaning and quality filtering via datamap-rs
Medium confidenceApplies domain-specific quality filters and cleaning rules to each of the 7 source corpora using the Datamap-rs tool, which performs large-scale text normalization, content filtering, and quality assessment. The tool enables source-specific filtering strategies (e.g., code quality metrics for The Stack, academic rigor for peS2o) while maintaining computational efficiency across the full 3 trillion token dataset.
Datamap-rs enables source-specific filtering strategies within a single pipeline, allowing different quality thresholds and content criteria for web text vs. code vs. academic papers vs. books, rather than applying uniform filters across all sources
More flexible than generic text cleaning tools (e.g., ftfy, NFKD normalization) by supporting domain-specific quality metrics, though specific filtering algorithms and thresholds are not publicly documented
dataset variant composition with configurable source mixing
Medium confidenceProvides multiple pretraining dataset variants (Standard Pool, Long Context Mix) with different source mixing ratios optimized for different training objectives. The variants are pre-composed and documented, allowing researchers to select a dataset variant matching their training goals without manually adjusting source proportions. The composition strategy reflects decisions about optimal balance between web text, code, academic content, and other domains.
Dolma provides pre-composed, documented dataset variants with explicit source mixing ratios rather than requiring users to manually combine sources or tune proportions, reducing configuration complexity and enabling reproducible comparisons across research teams
More structured than ad-hoc dataset composition and more transparent than proprietary models' undocumented mixing strategies, though less flexible than fully customizable composition systems
training data provenance tracing via olmotrace
Medium confidenceEnables researchers to trace model outputs back to specific training documents and source domains using the OlmoTrace tool, which maps model predictions to the training data that influenced them. This capability supports interpretability research, bias analysis, and data attribution by linking model behavior to specific training examples and sources within the Dolma corpus.
OlmoTrace integrates with Dolma's documented source composition and deduplication metadata to enable fine-grained tracing of model behavior to specific training sources, leveraging the dataset's transparency to support interpretability research that would be impossible with proprietary training data
More practical than generic influence functions because it leverages Dolma's explicit source composition and deduplication metadata; more comprehensive than document-level attribution because it can trace to specific source domains and filtering decisions
test set contamination detection and removal via decon
Medium confidenceIdentifies and removes test set data from the pretraining corpus using the Decon tool, which detects overlap between training data and evaluation benchmarks. This prevents data leakage that would artificially inflate model performance on standard benchmarks, ensuring that reported model performance reflects genuine capability rather than memorization of test examples.
Decon is specifically designed for pretraining dataset curation and integrates with Dolma's documented source composition, enabling systematic detection and removal of benchmark contamination before training rather than post-hoc analysis of model performance
More proactive than post-training contamination analysis and more comprehensive than manual benchmark checking, though specific detection algorithms and benchmark coverage are not documented
reproducible llm training integration via olmocore framework
Medium confidenceIntegrates Dolma with the OlmoCore training framework, which provides fast, easy configuration for pretraining language models with documented data composition, hyperparameters, and training procedures. The framework enables researchers to reproduce model training exactly by specifying dataset variant, mixing ratios, and training configuration, supporting fully reproducible LLM development from data through model weights.
OlmoCore is designed specifically for reproducible pretraining with Dolma, providing integrated configuration management for dataset composition, deduplication, filtering, and training hyperparameters in a single framework rather than requiring manual orchestration of separate tools
More integrated and reproducible than generic training frameworks (Hugging Face Transformers, DeepSpeed) because it bundles Dolma's documented data curation with training configuration; more transparent than proprietary training pipelines that don't expose data composition or filtering decisions
reproducible evaluation via olmes utility
Medium confidenceProvides the OLMES utility for running reproducible evaluations on models trained with Dolma and OlmoCore, enabling standardized benchmark testing with documented evaluation procedures. The utility ensures consistent evaluation methodology across research teams and model variants, supporting fair performance comparisons and preventing evaluation methodology drift.
OLMES is designed specifically for evaluating models trained with Dolma and OlmoCore, providing integrated evaluation procedures that document benchmark selection, metric definitions, and evaluation methodology to support reproducible model comparison
More integrated with Dolma/OlmoCore than generic evaluation frameworks (lm-evaluation-harness) and more transparent about evaluation procedures than proprietary model evaluation, though specific benchmarks and metrics are not documented
post-training data composition for instruction tuning and preference optimization
Medium confidenceProvides separate post-training corpora (distinct from the pretraining Dolma dataset) for instruction tuning and preference optimization, enabling researchers to fine-tune base models trained on Dolma with supervised instruction-following and reinforcement learning from human feedback (RLHF). The post-training data is composed and documented separately from pretraining data, supporting the full pipeline from base model training through instruction-tuned and preference-optimized variants.
Dolma provides separate, documented post-training data composition alongside the pretraining corpus, enabling full-pipeline reproducibility from base model training through instruction-tuned variants rather than requiring external post-training data sources
More integrated than using external instruction-tuning datasets (Alpaca, ShareGPT) because post-training data is composed and documented specifically for Dolma-trained models; more transparent than proprietary models' undocumented fine-tuning procedures
open-source model family training with documented variants
Medium confidenceSupports training the OLMo family of language models with multiple documented variants (7B, 32B base models; Instruct and Think configurations) using Dolma pretraining data and OlmoCore framework. Each model variant is trained with published hyperparameters, data composition, and training procedures, enabling researchers to reproduce or extend the model family with full transparency.
OLMo models are trained entirely on Dolma with fully documented data composition, hyperparameters, and training procedures, enabling researchers to reproduce model training or understand model behavior by tracing it back to specific training data and decisions
More transparent and reproducible than proprietary models (GPT, Claude) and more comprehensively documented than most open-source models (LLaMA, Mistral) regarding training data composition and curation decisions
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with Dolma, ranked by overlap. Discovered automatically through the match graph.
RedPajama v2
30 trillion token web dataset with 40+ quality signals per document.
C4 (Colossal Clean Crawled Corpus)
Google's cleaned Common Crawl corpus used to train T5.
c4
Dataset by allenai. 6,98,456 downloads.
CulturaX
6.3T token multilingual dataset across 167 languages.
FineWeb
Hugging Face's 15T token dataset, new standard for LLM training.
fineweb-edu
Dataset by HuggingFaceFW. 3,52,917 downloads.
Best For
- ✓LLM researchers training models from scratch with reproducibility requirements
- ✓organizations building custom language models who need transparent data sourcing
- ✓open-source ML practitioners requiring fully documented training data
- ✓teams studying data composition effects on model capabilities and biases
- ✓data engineers preparing large-scale pretraining corpora
- ✓researchers studying the impact of deduplication on model performance
- ✓teams building custom datasets who need efficient duplicate removal
- ✓data engineers curating large pretraining datasets with multiple source domains
Known Limitations
- ⚠3 trillion tokens requires substantial storage infrastructure (estimated 2-3TB uncompressed); not suitable for resource-constrained environments
- ⚠dataset composition is static; cannot dynamically adjust source mixing ratios without full reprocessing
- ⚠exact filtering thresholds and deduplication similarity metrics not exposed in public documentation, limiting independent reproduction of curation decisions
- ⚠inherits source-specific biases: Common Crawl skews toward English and contemporary web content, The Stack reflects GitHub distribution, peS2o selection criteria undocumented
- ⚠no explicit temporal coverage information or data cutoff dates provided
- ⚠appears English-dominant with no documented multilingual composition breakdown
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Allen AI's 3 trillion token open dataset used to train the OLMo family of language models. Curated from 7 sources: Common Crawl (web), The Stack (code), peS2o (academic), Project Gutenberg (books), Wikipedia, Wikibooks, and C4. Extensive documentation of data curation decisions including exact filtering rules, deduplication methods, and mixing ratios. Released alongside the OLMo toolkit for fully reproducible LLM training research.
Categories
Alternatives to Dolma
The GitHub for AI — 500K+ models, datasets, Spaces, Inference API, hub for open-source AI.
Compare →FLUX, Stable Diffusion, SDXL, SD3, LoRA, Fine Tuning, DreamBooth, Training, Automatic1111, Forge WebUI, SwarmUI, DeepFake, TTS, Animation, Text To Video, Tutorials, Guides, Lectures, Courses, ComfyUI, Google Colab, RunPod, Kaggle, NoteBooks, ControlNet, TTS, Voice Cloning, AI, AI News, ML, ML News,
Compare →Are you the builder of Dolma?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →