Capability
Long Context Text Generation With 128k Token Window
20 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
via “128k context window for long-document processing”
Mistral's efficient 24B model for production workloads.
Unique: Combines 128K context window with 24B parameter efficiency, enabling long-document processing on single GPU without cloud API costs, though context window claim not independently verified
vs others: Larger context window than many 24B models while maintaining single-GPU deployability, though smaller than some 70B+ models and context window claim lacks independent verification