RedPajama v2
DatasetFree30 trillion token web dataset with 40+ quality signals per document.
Capabilities11 decomposed
multi-language web-scale document collection with 40+ quality annotations
Medium confidenceAggregates 100+ billion deduplicated documents (30 trillion tokens) from 84 CommonCrawl dumps across 5 languages (English, German, French, Spanish, Italian). Each document is pre-annotated with 40+ quality signals including perplexity scores, deduplication hashes, content classifiers, and toxicity ratings computed via a standardized pipeline. The architecture processes raw CommonCrawl HTML through text extraction, deduplication, and multi-dimensional quality scoring, enabling downstream users to apply custom filtering strategies without reprocessing the raw data.
Processes 84 CommonCrawl dumps (claimed as most complete coverage vs. C4, Refinedweb, Dolma, SlimPajama) with 40+ pre-computed quality annotations per document, enabling fine-grained data curation research without requiring users to reprocess raw CommonCrawl. Open-source processing scripts allow reproducibility and custom filtering strategies on a standardized base dataset.
Larger scale (30 trillion tokens vs. C4's 156B tokens, RedPajama-1T's 1T tokens) with richer quality annotations (40+ signals vs. minimal metadata in competitors) and multilingual coverage, making it superior for comparative curation research and training diverse language models.
document-level deduplication with hash-based matching
Medium confidenceImplements deduplication across 100+ billion documents using hash-based matching to identify and remove duplicate content from CommonCrawl. The pipeline computes deduplication hashes for each document and filters the raw 100+ trillion token corpus down to 30 trillion deduplicated tokens. This approach preserves document boundaries (unlike token-level deduplication) and produces deterministic, reproducible results across reprocessing runs.
Uses document-level hash-based deduplication (preserving document boundaries) rather than token-level or fuzzy matching, enabling reproducible filtering and transparent deduplication hashes that users can inspect and verify. Processes 84 CommonCrawl dumps with consistent deduplication methodology.
Document-level deduplication is more interpretable and reproducible than token-level approaches, and the published deduplication hashes enable users to understand and verify which documents were removed, unlike proprietary datasets that hide deduplication decisions.
free and open-source corpus access
Medium confidenceProvides the entire 30 trillion token corpus, processing scripts, and quality annotations as free, open-source resources with no licensing restrictions. Users can download, modify, redistribute, and use the data for any purpose including commercial applications. This open approach enables broad research access and community-driven improvements without vendor lock-in.
Provides complete 30 trillion token corpus with processing scripts as free, open-source resources with no licensing restrictions, whereas competitors (C4, RefinedWeb) may have usage restrictions or require commercial licensing
Eliminates licensing costs and vendor lock-in through open-source distribution, enabling broad access for academic and commercial use versus competitors with restricted access or licensing requirements
perplexity-based quality scoring for language model fitness
Medium confidenceComputes perplexity scores for each document using a reference language model, enabling quantitative assessment of text quality and language model fitness. The perplexity metric measures how well a pre-trained model predicts the document; lower perplexity indicates higher-quality, more coherent text. These pre-computed scores allow users to filter documents by quality threshold without running inference themselves, and to study the relationship between perplexity and downstream model performance.
Pre-computes perplexity scores for 100+ billion documents, eliminating the computational cost of running inference for quality assessment. Enables comparative studies of how perplexity thresholds affect training outcomes without requiring users to implement their own scoring pipeline.
Provides pre-computed perplexity scores (eliminating inference cost) whereas competitors like C4 use heuristic filters (URL patterns, line-ending ratios); perplexity is a more principled, model-based quality metric but requires understanding of the reference model used.
content classification and toxicity annotation across documents
Medium confidenceAnnotates each document with content classifiers and toxicity ratings, enabling category-based filtering and safety-aware data curation. The pipeline applies pre-trained classifiers to categorize document content (e.g., news, forums, documentation) and compute toxicity scores. These annotations are pre-computed and stored with each document, allowing users to filter by content type or toxicity threshold without running inference themselves.
Pre-computes both content classifiers and toxicity ratings for 100+ billion documents, enabling multi-dimensional safety and content-based filtering without requiring users to implement or run their own classifiers. Supports comparative studies of how content filtering affects model behavior.
Provides pre-computed toxicity and content annotations (eliminating inference cost) whereas most web datasets require downstream filtering; enables safety-aware curation at scale without custom classifier implementation.
open-source reproducible data processing pipeline
Medium confidencePublishes end-to-end processing scripts on GitHub that convert raw CommonCrawl HTML to deduplicated, annotated documents. The pipeline is fully open-source, enabling users to understand, verify, and reproduce the data processing methodology. Scripts handle HTML-to-text conversion, deduplication, quality signal computation, and filtering, allowing researchers to reprocess data with custom parameters or apply the same methodology to new CommonCrawl dumps.
Publishes complete, open-source processing scripts enabling full reproducibility and transparency of data processing methodology. Users can inspect, verify, and reapply the pipeline to new data, unlike proprietary datasets where processing is opaque.
Open-source pipeline enables reproducibility and auditability vs. proprietary datasets (C4, Refinedweb) where processing methodology is proprietary or partially documented; enables research on data processing methodology itself.
fine-grained data curation via quality signal filtering
Medium confidenceEnables users to apply custom filtering strategies by combining 40+ pre-computed quality signals (perplexity, toxicity, content classifiers, deduplication hashes, etc.). Rather than providing pre-filtered 'ready-to-train' datasets, RedPajama v2 provides the raw signals and lets users define their own filtering logic. This architecture supports comparative studies of curation strategies and enables organizations to apply domain-specific or value-aligned filtering without reprocessing the base dataset.
Provides 40+ pre-computed quality signals enabling fine-grained, user-defined curation strategies rather than pre-filtered datasets. This architecture supports comparative research on curation methodology and enables organizations to apply custom filtering without reprocessing the base dataset.
Enables comparative curation research (studying how different filtering strategies affect outcomes) whereas competitors provide pre-filtered datasets; gives users control over filtering logic but requires more implementation effort.
multilingual web corpus with consistent annotation across 5 languages
Medium confidenceProvides 30 trillion tokens across 5 languages (English, German, French, Spanish, Italian) with consistent quality signal annotations applied uniformly across all languages. The architecture processes each language through the same deduplication, quality scoring, and classification pipeline, enabling comparative studies of language-specific data characteristics and training multilingual models on a standardized base dataset. Language-specific processing details are not documented, but the consistent annotation methodology enables cross-language analysis.
Provides 30 trillion tokens across 5 languages with identical quality signal annotations, enabling comparative studies of language-specific data characteristics and training multilingual models on a standardized base. Consistent annotation methodology across languages enables cross-language analysis.
Larger multilingual coverage (5 languages, 30 trillion tokens) than RedPajama-1T (English-only, 1 trillion tokens) and most competitors; consistent annotation enables comparative language research, but limited to European languages vs. competitors with broader language coverage.
commoncrawl-scale data aggregation from 84 dumps
Medium confidenceAggregates data from 84 CommonCrawl dumps (100+ trillion raw tokens) into a single, deduplicated, consistently-annotated dataset. The architecture handles the complexity of processing massive-scale web data including deduplication across dumps, consistent quality signal computation, and language-specific filtering. This enables users to work with a unified, large-scale web corpus without managing multiple CommonCrawl dumps or implementing their own aggregation pipeline.
Processes 84 CommonCrawl dumps (claimed as most complete coverage vs. C4, Refinedweb, Dolma, SlimPajama) into a single, consistently-annotated dataset. Eliminates user burden of managing multiple dumps and implementing aggregation logic.
Larger scale (30 trillion tokens, 84 dumps) than competitors (C4: 156B tokens, Refinedweb: limited dumps, Dolma: limited dumps); unified dataset eliminates user aggregation burden but inherits web biases from CommonCrawl.
open-source processing pipeline and transparency
Medium confidencePublishes processing scripts on GitHub enabling users to understand, validate, and extend the data processing pipeline. Scripts cover HTML-to-text conversion, deduplication, quality signal computation, and filtering. This transparency enables reproducible research, allows users to apply custom modifications, and supports community contributions. Users can inspect the exact methodology used for corpus creation and adapt it for their own data sources.
Publishes complete processing scripts on GitHub enabling users to validate, reproduce, and extend the data processing pipeline, whereas competitors typically keep processing methodology proprietary or undocumented
Provides full transparency into data processing through open-source scripts, enabling reproducible research and community contributions, versus competitors that hide processing methodology or provide only final datasets
huggingface dataset distribution and streaming
Medium confidenceDistributes the 30 trillion token corpus via HuggingFace Datasets, enabling users to download, stream, or access subsets without managing raw files directly. HuggingFace integration provides standardized data loading APIs compatible with PyTorch, TensorFlow, and other ML frameworks. Users can load documents with quality annotations, apply filters, and create training dataloaders with minimal code.
Distributes 30 trillion token corpus through HuggingFace Datasets with standardized APIs for PyTorch/TensorFlow integration, whereas competitors require custom data loading code or proprietary distribution mechanisms
Enables seamless integration with standard ML frameworks through HuggingFace Datasets, reducing engineering overhead versus competitors requiring custom data loading implementations
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with RedPajama v2, ranked by overlap. Discovered automatically through the match graph.
mC4
Multilingual web corpus covering 101 languages.
OPUS
Massive parallel corpus for machine translation.
C4 (Colossal Clean Crawled Corpus)
Google's cleaned Common Crawl corpus used to train T5.
Hebbia
Revolutionize document analysis: AI collaboration, transparency, vast data...
FineFineWeb
Dataset by m-a-p. 4,59,057 downloads.
CulturaX
6.3T token multilingual dataset across 167 languages.
Best For
- ✓LLM researchers training foundation models at scale
- ✓organizations studying data curation and filtering strategies
- ✓teams building open-source language models across multiple languages
- ✓data scientists analyzing web content quality distributions
- ✓LLM training teams concerned with data quality and training efficiency
- ✓researchers studying the impact of deduplication on model performance
- ✓organizations building custom datasets from CommonCrawl
- ✓academic researchers with limited budgets
Known Limitations
- ⚠Web-only source (CommonCrawl) inherits web biases, spam, and low-quality content; requires downstream filtering to achieve production quality
- ⚠40+ quality signals are pre-computed but specific signal definitions and validation methodology not publicly documented
- ⚠No domain-specific data (code, scientific papers, books) — coverage limited to web content
- ⚠5 languages only; no coverage for non-Latin scripts or low-resource languages
- ⚠Raw data (100+ trillion tokens) is 3.3× larger than processed data, indicating significant filtering already applied; original filtering thresholds not transparent
- ⚠No temporal metadata or freshness guarantees for CommonCrawl dumps
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Together AI's massive 30 trillion token web dataset with over 100 billion documents across 5 languages (English, German, French, Spanish, Italian). Each document annotated with 40+ quality signals enabling fine-grained data curation. Includes perplexity scores, deduplication hashes, content classifiers, and toxicity ratings. Designed to enable reproducible LLM training research. The quality signal annotations make it uniquely valuable for studying data curation strategies.
Categories
Alternatives to RedPajama v2
Open-source image generation — SD3, SDXL, massive ecosystem of LoRAs, ControlNets, runs locally.
Compare →Are you the builder of RedPajama v2?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →