Yujin ChoA. R. LeeByung‐Gyu KimJan Plato
Abstract The Vision Transformer (ViT) has demonstrated remarkable performance in a wide range of computer vision tasks, such as image classification, object detection, and image generation. Unlike convolutional neural networks (CNNs), ViT benefits from a global receptive field, which enables more effective modeling of relationships between image patches. However, the lack of inductive biases makes ViT models difficult to train stably, especially on limited datasets. Without access to large-scale pre-trained weights, performance often degrades significantly. To address this issue, we propose a novel architecture called RMSF-ViT. It employs a progressive fusion strategy that incorporates fine-grained patch information beyond the fixed single patch size used in conventional ViT architectures. In addition, RMSF-ViT reduces the number of attention heads by half compared to vanilla ViT models. This design improves both performance and computational efficiency, as demonstrated on the CIFAR-10, CIFAR-100, Flowers, and Pets datasets.
Hao TangDawei LiuChengchao Shen
Yujie MaHanfa XingZurui AoCixian Yu
Zhi JinGuoheng HuangFeng ZhangXiaochen YuanDing ZhuZhe TanChi‐Man PunGuo Zhong
Chunlong XiaXinliang WangFeng LvXin HaoYifeng Shi
Zhou HuangZhigang LiMing ZhuBo WuPaul Chen