Vision Transformers

Compact Convolutional Transformers

Introduced by Hassani et al. in Escaping the Big Data Paradigm with Compact Transformers

Compact Convolutional Transformers utilize sequence pooling and replace the patch embedding with a convolutional embedding, allowing for better inductive bias and making positional embeddings optional. CCT achieves better accuracy than ViT-Lite (smaller ViTs) and increases the flexibility of the input parameters.

Source: Escaping the Big Data Paradigm with Compact Transformers

Papers


Paper Code Results Date Stars

Categories