An arbitrary view synthesis method from 2D-Plus-Depth image for real-time auto-stereoscopic display is presented.
Traditional methods use depth image based rendering (DIBR) technology, which is a process of synthesizing “virtual”
views of a scene from still or moving images and associated per-pixel depth information. All the virtual view images are
generated and then the ultimate stereo-image is synthesized. DIBR can greatly decrease the number of reference images
and is flexible and efficient as the depth images are used. However it causes some problems such as the appearance of
holes in the rendered image, and the occurrence of depth discontinuity on the surface of the object at virtual image plane.
Here, reversed disparity shift pixel rendering is used to generate the stereo-image directly, and the target image won’t
generate holes. To avoid duplication of calculation and also to be able to match with any specific three-dimensional
display, a selecting table is designed to pick up appropriate virtual viewpoints for auto-stereoscopic display. According to
the selecting table, only sub-pixels of the appropriate virtual viewpoints are calculated, so calculation amount is
independent of the number of virtual viewpoints. In addition, 3D image warping technology is used to translate depth
information to parallax between virtual viewpoints and parallax, and the viewer can adjust the
zero-parallax-setting-plane (ZPS) and change parallax conveniently to suit his/her personal preferences. The proposed
method is implemented with OPENGL and demonstrated on a laptop computer with a 2.3 GHz Intel Core i5 CPU and
NVIDA GeForce GT540m GPU. We got a frame rate 30 frames per second with 4096×2340 video. High synthesis
efficiency and good stereoscopic sense can be obtained. The presented method can meet the requirements of
real-time ultra-HD super multi-view auto-stereoscopic display.