Routed Attention is an attention pattern proposed as part of the Routing Transformer architecture. Each attention module considers a clustering of the space: the current timestep only attends to context belonging to the same cluster. In other word, the current time-step query is routed to a limited number of context through its cluster assignment. This can be contrasted with strided attention patterns and those proposed with the Sparse Transformer.
In the image to the right, the rows represent the outputs while the columns represent the inputs. The different colors represent cluster memberships for the output token.
Source: Efficient Content-Based Sparse Attention with Routing TransformersPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Image Segmentation | 1 | 5.88% |
Medical Image Segmentation | 1 | 5.88% |
Semantic Segmentation | 1 | 5.88% |
2D Object Detection | 1 | 5.88% |
Medical Diagnosis | 1 | 5.88% |
medical image detection | 1 | 5.88% |
Medical Object Detection | 1 | 5.88% |
Object Detection | 1 | 5.88% |
Real-Time Object Detection | 1 | 5.88% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |