roberta-base-openai-detectorModel45/100 via “binary-classification-of-ai-generated-text”
text-classification model by undefined. 9,16,951 downloads.
Unique: Fine-tuned specifically on GPT-2 generated text paired with BookCorpus/Wikipedia human text, making it one of the earliest publicly available detectors trained on a controlled synthetic dataset rather than heuristic rules or proprietary data. Uses RoBERTa's masked language modeling pretraining as a foundation, which captures deeper syntactic and semantic patterns than bag-of-words or n-gram baselines.
vs others: More accurate than rule-based detectors (perplexity thresholds, entropy analysis) on GPT-2 outputs, but significantly less effective than newer detectors trained on GPT-3.5/4 outputs; trades generalization for interpretability since it's a standard transformer classifier rather than a black-box ensemble.