Vision Transformers have achieved impressive performance in many vision
tasks. While the token mixer or attention block has been studied in great
detail, much less research has been devoted to the channel mixer or feature
mixing block (FFN or MLP), which accounts for a significant portion of of the
model parameters and computation. In this work, we show that the dense MLP
connections can be replaced with a block diagonal MLP structure that supports
larger expansion ratios by splitting MLP features into groups. To improve the
feature clusters formed by this structure we propose the use of a lightweight,
parameter-free, channel covariance attention (CCA) mechanism as a parallel
branch during training. This enables gradual feature mixing across channel
groups during training whose contribution decays to zero as the training
progresses to convergence. In result, the CCA block can be discarded during
inference, enabling enhanced performance at no additional computational cost.
The resulting Scalable CHannEl MixEr (SCHEME) can be plugged into
any ViT architecture to obtain a gamut of models with different trade-offs
between complexity and performance by controlling the block diagonal MLP
structure. This is shown by the introduction of a new family of SCHEMEformer
models. Experiments on image classification, object detection, and semantic
segmentation, with different ViT backbones, consistently demonstrate
substantial accuracy gains over existing designs, especially for lower
complexity regimes. The SCHEMEformer family is shown to establish new Pareto
frontiers for accuracy vs FLOPS, accuracy vs model size, and accuracy vs
throughput, especially for fast transformers of small size.Comment: Preprin