Capability
Session Based Document History And Re Summarization
20 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
via “long-context understanding and summarization”
text-generation model by undefined. 1,06,54,004 downloads.
Unique: DeepSeek-V3.2 uses sparse mixture-of-experts with efficient attention patterns (e.g., grouped-query attention) to handle longer contexts with lower memory overhead than dense models, enabling 4K-8K token processing without proportional VRAM increases
vs others: Processes 4K-token documents with 30-40% lower VRAM than Llama-2-70B due to sparse MoE and efficient attention, while maintaining comparable summarization quality on CNN/DailyMail and XSum benchmarks