What is: Gated Positional Self-Attention?
Source | ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Gated Positional Self-Attention (GPSA) is a self-attention module for vision transformers, used in the ConViT architecture, that can be initialized as a convolutional layer -- helping a ViT learn inductive biases about locality.