1 April 2008 Prediction structure and quantization parameter selection for efficient multiview video coding
Pil-Kyu Park, Yo-Sung Ho
Author Affiliations +
Abstract
Multiview video is obtained by capturing a three-dimensional scene using many cameras simultaneously. Since adjacent view sequences have strong spatial redundancy as well as temporal redundancy, we can achieve additional coding gain by eliminating the spatial redundancy. In this paper, we propose view-temporal prediction structures that can be adjusted to various characteristics of general multiview video by separating them into temporal and view prediction structures. The proposed temporal prediction structure minimizes the average distance between a frame to be coded and its reference frames. The proposed view prediction structure considers the location of the I frame in anchor and the number of B frames to be inserted in anchor. We also propose a quantization parameter selection method for enhancing coding efficiency of views whose anchor frames are coded as B frames. Experimental results show that the proposed algorithms achieved 0.10 to 0.34 dB of peak SNR (PSNR) gain and reduced the variances of PSNR values of views by 28.26% on average against the results by the reference prediction.
©(2008) Society of Photo-Optical Instrumentation Engineers (SPIE)
Pil-Kyu Park and Yo-Sung Ho "Prediction structure and quantization parameter selection for efficient multiview video coding," Optical Engineering 47(4), 047401 (1 April 2008). https://doi.org/10.1117/1.2909666
Published: 1 April 2008
Lens.org Logo
CITATIONS
Cited by 4 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Video

Quantization

Cameras

Video coding

Holmium

Optical engineering

3D video streaming

Back to Top