oneformer_ade20k_swin_largeModel41/100 via “task-conditioned-query-generation”
image-segmentation model by undefined. 1,02,623 downloads.
Unique: Implements task conditioning via learnable query tokens (e.g., 100 queries for panoptic, 150 for semantic) that are concatenated with positional encodings and processed through the same transformer decoder stack. This differs from multi-head approaches (separate decoder heads per task) by forcing shared feature representations while allowing task-specific query distributions.
vs others: Reduces model parameters by 25-30% vs separate task-specific decoders while maintaining within 0.5 mIoU of task-specific models, enabling efficient multi-task deployment. However, task-specific models can be independently optimized, potentially achieving 1-2 mIoU higher performance if model size is not constrained.