FLAN CollectionDataset44/100 via “multi-task instruction-tuning dataset aggregation”
Google's 1,836-task instruction mixture for broad generalization.
Unique: Aggregates four heterogeneous instruction datasets (Flan 2021, P3, Super-Natural Instructions, CoT) into a single unified mixture with explicit task-level composition tracking, enabling reproducible instruction-tuning at scale. Uses multiple prompt templates per task (3-10 variants) to improve robustness to prompt phrasing variations, a technique not consistently applied across individual source datasets.
vs others: Larger and more diverse than any single instruction dataset (1,836 vs ~500 tasks in P3 alone), and explicitly designed for multi-task generalization rather than task-specific optimization, making it more suitable for training general-purpose instruction-following models than domain-specific alternatives.