Convolutional neural networks are designed for dense data, but vision data is\noften sparse (stereo depth, point clouds, pen stroke, etc.). We present a\nmethod to handle sparse depth data with optional dense RGB, and accomplish\ndepth completion and semantic segmentation changing only the last layer. Our\nproposal efficiently learns sparse features without the need of an additional\nvalidity mask. We show how to ensure network robustness to varying input\nsparsities. Our method even works with densities as low as 0.8% (8 layer\nlidar), and outperforms all published state-of-the-art on the Kitti depth\ncompletion benchmark.\n
Maximilian JaritzRaoul de CharetteÉmilie WirbelXavier PerrottonFawzi Nashashibi
Bohao PengXiaoyang WuJiang LiYukang ChenHengshuang ZhaoZhuotao TianJiaya Jia
Sadaf FarkhaniMikkel Fly KraghPeter ChristiansenRasmus Nyholm JørgensenHenrik Karstoft
Yan XuXinge ZhuJianping ShiGuofeng ZhangHujun BaoHongsheng Li
Shuo GuJiacheng LuJian YangChengzhong XuHui Kong