Zhao YangJiaqi WangYansong TangKai ChenHengshuang ZhaoPhilip Torr
Referring image segmentation is a fundamental vision-language task that aims to segment out an object referred to by a natural language expression from an image. One of the key challenges behind this task is leveraging the referring expression for highlighting relevant positions in the image. A paradigm for tackling this problem is to leverage a powerful vision-language ("cross-madal") decoder to fuse features independently extracted from a vision encoder and a language encoder. Recent methods have made remarkable advancements in this paradigm by exploiting Transformers as cross-modal decoders, concurrent to the Transformer's overwhelming success in many other vision-language tasks. Adopting a different approach in this work, we show that significantly better cross-modal alignments can be achieved through the early fusion of linguistic and visual features in intermediate layers of a vision Transformer encoder network. By conducting cross-modal feature fusion in the visual feature encoding stage, we can leverage the well-proven correlation modeling power of a Transformer encoder for excavating helpful multi-modal context. This way, accurate segmentation results are readily harvested with a light-weight mask predictor. Without bells and whistles, our method surpasses the previous state-of-the-art methods on Ref CoCo, RefCOCO+, and G-Ref by large margins.
Zhao YangJiaqi WangXubing YeYansong TangKai ChenHengshuang ZhaoPhilip H. S. Torr
Fayou XuBing LuoChao ZhangLi XuMingxing PuBo Li
Nisarg A. ShahVibashan VSVishal M. Patel
Shuyi OuyangHongyi WangShiao XieZiwei NiuRuofeng TongYen‐Wei ChenLanfen Lin
Yubin ChoHyunwoo YuSuk‐Ju Kang