Given a set of images of the same scene, we propose an efficient method for realistically synthesizing a new image seen from a new viewpoint. Compared with some existing techniques which explicitly reconstruct the 3D geometry of the scene, we directly reconstruct the colour of each pixel in the new image by utilizing a weighted photoconsistency constraint. Global photoconsistency constraint of one pixel is firstly utilized to generate a list of plausible colours for each rendered pixel in the new image. Considering the existing of partial occlusion or the deficiencies in the image-formation model, we iteratively update the colour for each rendered pixel based on local texture statistics similar to the input images. Experimental results on the generation of a new image from a new viewing position from a set of input images show that our proposed method is promising and satisfactory.
Ricardo MarroquimMartin KrausPaulo Roma Cavalcanti
Chang LiuXinzhu SangXunbo YuXin GaoLi LiuKuiru WangBinbin YanChongxiu Yu