JOURNAL ARTICLE

Multi-Stream Attention-Aware Graph Convolution Network for Video Salient Object Detection

Mingzhu XuPing FuBing LiuJun-Bao Li

Year: 2021 Journal:   IEEE Transactions on Image Processing Vol: 30 Pages: 4183-4197   Publisher: Institute of Electrical and Electronics Engineers

Abstract

Recent advances in deep convolution neural networks (CNNs) boost the development of video salient object detection (SOD), and many remarkable deep-CNNs video SOD models have been proposed. However, many existing deep-CNNs video SOD models still suffer from coarse boundaries of the salient object, which may be attributed to the loss of high-frequency information. The traditional graph-based video SOD models can preserve object boundaries well by conducting superpixels/supervoxels segmentation in advance, but they perform weaker in highlighting the whole object than the latest deep-CNNs models, limited by heuristic graph clustering algorithms. To tackle this problem, we find a new way to address this issue under the framework of graph convolution networks (GCNs), taking advantage of graph model and deep neural network. Specifically, a superpixel-level spatiotemporal graph is first constructed among multiple frame-pairs by exploiting the motion cues implied in the frame-pairs. Then the graph data is imported into the devised multi-stream attention-aware GCN, where a novel Edge-Gated graph convolution (GC) operation is proposed to boost the saliency information aggregation on the graph data. A novel attention module is designed to encode the spatiotemporal sematic information via adaptive selection of graph nodes and fusion of the static-specific and the motion-specific graph embedding. Finally, a smoothness-aware regularization term is proposed to enhance the uniformity of salient object. Graph nodes (superpixels) inherently belonging to the same class will be ideally clustered together in the learned embedding space. Extensive experiments have been conducted on three widely used datasets. Compared with fourteen state-of-the-art video SOD models, our proposed method can well retain the salient object boundaries and possess a strong learning ability, which shows that this work is a good practice for designing GCNs for video SOD.

Keywords:
Computer science Artificial intelligence Pattern recognition (psychology) Graph Computer vision Theoretical computer science

Metrics

59
Cited By
4.91
FWCI (Field Weighted Citation Impact)
83
Refs
0.96
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Visual Attention and Saliency Detection
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Image and Video Quality Assessment
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

SAGNet: Synergistic Attention-Graph Network For video salient object detection

Huo LinaXueyuan GaoWei WangKe ChenKe Wang

Journal:   Image and Vision Computing Year: 2025 Vol: 160 Pages: 105570-105570
JOURNAL ARTICLE

Three-Stream Attention-Aware Network for RGB-D Salient Object Detection

Hao ChenYoufu Li

Journal:   IEEE Transactions on Image Processing Year: 2019 Vol: 28 (6)Pages: 2825-2835
JOURNAL ARTICLE

Multi-Stream Temporally Enhanced Network for Video Salient Object Detection

Dan XuJiale RuJinlong Shi

Journal:   Computers, materials & continua/Computers, materials & continua (Print) Year: 2023 Vol: 78 (1)Pages: 85-104
JOURNAL ARTICLE

Edge-Aware Convolution Neural Network Based Salient Object Detection

Wenlong GuanTiantian WangJinqing QiLihe ZhangHuchuan Lu

Journal:   IEEE Signal Processing Letters Year: 2018 Vol: 26 (1)Pages: 114-118
© 2026 ScienceGate Book Chapters — All rights reserved.