What is: Gated Positional Self-Attention?
| Source | ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases |
| Year | 2000 |
| Data Source | CC BY-SA - https://paperswithcode.com |
Gated Positional Self-Attention (GPSA) is a self-attention module for vision transformers, used in the ConViT architecture, that can be initialized as a convolutional layer -- helping a ViT learn inductive biases about locality.
