In 360-degree video streaming, Most solutions are based on tile-based streaming that divides videos into tiles and streams the high-quality tiles corresponding to the user's viewport areas. However, these methods cannot transmit different combinations of tile coding efficiently. In this paper, we experimented with streaming 360-degree videos using a motion-constrained tile set (MCTS) technique that allows encoding with constraining motion vectors such that each tile can be decoded and transmitted independently. Moreover, we have used a tile-based approach using a saliency map that integrates the information of human visual attention with the contents to deliver high-quality tiles to the region of interest (ROI). We encoded the 360-degree videos at various quality representations with MCTS techniques and assigned a tile quality representation using a saliency map predicted by the existing convolutional neural network (CNN) model. We proposed a novel heuristic algorithm to assign appropriate quality to the tiles on the centerline. Consequently, mixed quality videos based on the saliency map enable efficient streaming in 360-degree videos. Using the Salient360! dataset, the proposed method shows an improvement in terms of bandwidth with little loss of viewport image quality.
Jong-Beom JeongSoonbin LeeInae KimSangSoon LeeEun Seok Ryu
Wenxuan GaoChenglin LiHaoran LvWenrui DaiJunni ZouHongkai XiongXinlong PanHaipeng Wang
Chiao‐Wen LinYi‐Zih ChenYong-Siang WangYu-Hsiu ChouYu-Ching YangNien-Chen LinWanjiun Liao
Chuong Hoang VoJui‐Chiu ChiangDuy H. LeThu NguyenTuan Van Pham