Weichen XuYezhi ShenQian LinJan P. AllebachFengqing Zhu
Virtual background has become an increasingly important feature of online video conferencing due to the popularity of remote work in recent years. To enable virtual background, a segmentation mask of the participant needs to be extracted from the real-time video input. Most previous works have focused on image based methods for portrait segmentation. However, portrait video segmentation poses additional challenges due to complicated background, body motion, and inter-frame consistency. In this paper, we utilize temporal guidance to improve video segmentation, and propose several methods to address these challenges including prior mask, optical flow, and visual memory. We leverage an existing portrait segmentation model PortraitNet to incorporate our temporal guided methods. Experimental results show that our methods can achieve improved segmentation performance on portrait videos with minimum latency.
Weichen XuYezhi ShenQian LinJan P. AllebachFengqing Zhu
Yifan WangWenbo ZhangLijun WangFenghua YangHuchuan Lu
Jesús BescósA. MovillaJosé Manuel MenéndezGuillermo Cisneros
Wujun WenYunheng LiZhuben DongLin FengWanxiao YangShenlan Liu
Janko ĆalićSorin SavEbroul IzquierdoSeán MarlowN. MurphyN. O'Connor