WildChat
DatasetFree1M+ real user-AI conversations with demographic metadata.
Capabilities9 decomposed
real-world conversation dataset collection and curation
Medium confidenceAggregates over 1 million authentic user conversations with ChatGPT and GPT-4 captured through a research chatbot interface, preserving full conversation threads with metadata including timestamps, user demographics (country, browser type), and conversation-level toxicity annotations. The dataset captures genuine, unfiltered user intents across diverse domains without synthetic generation or prompt engineering, enabling analysis of actual AI usage patterns in production environments.
Captures unfiltered, real-world conversations from production ChatGPT/GPT-4 deployments rather than synthetic or crowdsourced data, preserving authentic user intents, failure modes, and edge cases with demographic metadata (country, browser) enabling stratified analysis across user populations
Larger scale (1M+ conversations) and more authentic than crowdsourced datasets like ShareGPT, with explicit demographic metadata absent from most open conversation corpora, though less curated and safety-filtered than instruction-tuning datasets like FLAN or Alpaca
demographic-stratified conversation analysis and filtering
Medium confidenceEnables filtering and analysis of conversations by user demographics (country, browser type) and conversation-level metadata, allowing researchers to slice the dataset by geographic region, device type, or other user attributes. The dataset structure preserves demographic fields as queryable attributes, supporting cohort analysis, geographic bias detection, and population-specific model evaluation without requiring external demographic inference.
Provides explicit demographic metadata (country, browser) at conversation level, enabling direct stratified analysis without requiring external demographic inference or proxy models, though limited to coarse-grained attributes compared to crowdsourced alternatives
More direct demographic stratification than ShareGPT or other conversation corpora, though less granular than purpose-built fairness datasets with rich demographic annotations
toxicity annotation and content safety labeling
Medium confidenceProvides conversation-level toxicity labels assigned through automated or human annotation, enabling researchers to identify and filter harmful content, study safety patterns, and train content moderation models. Labels are attached at the conversation level (not per-message), allowing downstream filtering of unsafe conversations or stratified analysis of toxicity distribution across user demographics and conversation types.
Provides real-world toxicity annotations from production ChatGPT/GPT-4 conversations rather than synthetic or crowdsourced toxic examples, capturing authentic harmful content patterns without artificial prompt engineering, though at conversation-level granularity rather than message-level
More authentic toxicity examples than synthetic safety datasets, though coarser-grained labeling and less detailed harm taxonomy than purpose-built safety datasets like ToxiGen or RealToxicityPrompts
multilingual conversation corpus extraction and analysis
Medium confidenceProvides access to non-English conversations within the dataset, enabling analysis of how users in different languages interact with English-trained LLMs and supporting training of multilingual or cross-lingual models. Conversations are preserved in original language with metadata indicating language or country of origin, allowing language-specific filtering and comparative analysis across linguistic communities.
Includes real-world multilingual conversations from production ChatGPT/GPT-4 deployments, capturing authentic non-English user interactions and code-switching patterns, though limited in coverage and requiring language detection for explicit language identification
More authentic multilingual examples than synthetic multilingual datasets, though smaller and less balanced than purpose-built multilingual corpora like FLORES or mC4
conversation metadata extraction and temporal analysis
Medium confidenceProvides structured metadata for each conversation including timestamps, conversation IDs, user IDs, and conversation length, enabling temporal analysis of usage patterns, trend detection, and time-series studies of how user needs and LLM interactions evolved. Metadata is queryable and filterable, supporting cohort analysis by time period and correlation analysis between temporal patterns and conversation characteristics.
Preserves conversation-level timestamps from production ChatGPT/GPT-4 deployments, enabling temporal analysis of real-world usage evolution without synthetic time-shifting, though limited to conversation-level granularity without turn-level timing
More authentic temporal data than synthetic datasets, though coarser-grained than specialized time-series conversation corpora with explicit turn-level timestamps
domain and use-case diversity sampling and stratification
Medium confidenceProvides conversations spanning diverse user intents and domains (coding help, creative writing, sensitive topics, general Q&A, etc.) captured from real users without prompt engineering, enabling researchers to sample representative conversations across use cases and train models on realistic domain distributions. The dataset's scale and authenticity allow stratified sampling by inferred domain or use case without requiring explicit domain labels.
Captures authentic domain diversity from real ChatGPT/GPT-4 users without synthetic prompt engineering, preserving natural distribution of use cases and user intents, though requiring post-hoc domain inference rather than explicit labels
More authentic domain diversity than synthetic instruction-tuning datasets, though less explicitly labeled and curated than purpose-built domain-specific corpora
conversation metadata extraction and statistical summarization
Medium confidenceThe dataset includes structured metadata for each conversation (user demographics, browser/device info, conversation length, timestamps, toxicity labels) that can be extracted and aggregated for statistical analysis. Researchers can compute summary statistics (e.g., average conversation length by country, toxicity prevalence by domain) without processing full conversation text, enabling efficient exploratory analysis and dataset characterization. Metadata is stored in queryable fields, supporting both individual record lookup and bulk aggregation.
Provides structured metadata fields (country, browser, device, toxicity label) linked to each conversation, enabling efficient statistical summarization without processing full conversation text. Metadata is captured at collection time, preserving temporal and contextual information.
More efficient for statistical analysis than processing full conversation text, but metadata quality and completeness are not explicitly documented compared to explicitly validated datasets
instruction-following and user intent distribution analysis
Medium confidenceThe dataset captures authentic user requests and model responses, enabling analysis of instruction-following patterns, user intent distribution, and how well models address diverse user needs. Researchers can analyze which types of instructions users provide, how models interpret and respond to them, and where misalignment or misunderstanding occurs. This supports studying instruction-following quality, identifying common user frustrations, and understanding the diversity of real-world use cases beyond typical benchmarks.
Captures authentic user instructions and model responses from production ChatGPT/GPT-4 deployments, reflecting real instruction-following challenges and user intent distribution rather than synthetic instruction-tuning data. Includes edge cases and sensitive topics that users genuinely request.
More representative of real-world instruction-following patterns than synthetic instruction-tuning datasets, but lacks explicit success metrics or user satisfaction labels compared to explicitly validated instruction-following benchmarks
model behavior and response quality comparative analysis
Medium confidenceThe dataset includes conversations with both ChatGPT and GPT-4, enabling direct comparison of model behavior, response quality, and user satisfaction across model versions. Researchers can analyze how model improvements manifest in real-world usage, identify domains where newer models perform better, and study whether user satisfaction or request patterns differ by model. This supports understanding model evolution, identifying model-specific failure modes, and studying how users adapt to model capabilities.
Provides direct comparison of ChatGPT and GPT-4 behavior on identical user requests in production, capturing how model improvements manifest in real-world usage rather than controlled benchmarks. Includes user reactions and follow-up requests that reveal satisfaction and adaptation patterns.
More representative of real-world model comparison than synthetic benchmarks, but lacks explicit quality labels or user satisfaction metrics compared to explicitly annotated model evaluation datasets
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with WildChat, ranked by overlap. Discovered automatically through the match graph.
OpenAssistant Conversations (OASST)
161K human-written messages in 35 languages with quality ratings.
UltraChat 200K
200K high-quality multi-turn dialogues for instruction tuning.
RedPajama v2
30 trillion token web dataset with 40+ quality signals per document.
ShareGPT
Real ChatGPT conversations used to train Vicuna.
Capybara
Multi-turn conversation dataset for steerable models.
ToxiGen
Microsoft's dataset for implicit toxicity detection.
Best For
- ✓ML researchers studying LLM behavior and user interaction patterns
- ✓teams building instruction-tuned models requiring diverse, authentic training data
- ✓researchers analyzing geographic and demographic variations in AI usage
- ✓safety researchers studying real-world toxicity, jailbreaks, and edge cases
- ✓researchers studying geographic variation in AI usage and user needs
- ✓teams building localized or region-specific AI products
- ✓fairness researchers analyzing demographic disparities in AI interactions
- ✓product teams understanding device-specific usage patterns
Known Limitations
- ⚠Dataset is English-dominant with limited multilingual coverage despite claims of multilingual conversations
- ⚠Toxicity labels are coarse-grained (binary or limited categories) rather than fine-grained harm taxonomy
- ⚠No explicit consent from original ChatGPT/GPT-4 users — raises privacy and licensing questions for derivative use
- ⚠Conversation metadata is limited to country and browser; lacks temporal distribution analysis or user segmentation by expertise level
- ⚠No conversation quality scores or user satisfaction ratings — cannot distinguish high-value from low-value interactions
- ⚠Demographic data is limited to country and browser type — no age, education, expertise level, or socioeconomic indicators
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Allen AI's collection of over 1 million real user conversations with ChatGPT and GPT-4 captured through a research chatbot interface. Includes user demographics (country, browser), conversation metadata, and toxicity labels. Covers genuine user needs from coding help to creative writing to sensitive topics. Uniquely valuable for understanding real-world AI usage patterns. Includes both English and multilingual conversations, providing insight into how diverse populations interact with AI.
Categories
Alternatives to WildChat
Open-source image generation — SD3, SDXL, massive ecosystem of LoRAs, ControlNets, runs locally.
Compare →Are you the builder of WildChat?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →