Reference Picture Synthesis for Video Sequences Captured with a Monocular Moving Camera

H. B. Golestani, Christian Rohlfing, J.-R. Ohm

To be presented at the 2019 IEEE International Conference Visual Communications and Image Processing (VCIP 2019), 1-4 December 2019, Sydney, Australia.
Conference homepage


Inter-frame prediction plays an important role in video coding by predicting the current frame from previously encoded pictures, called reference pictures. In the case of camera motion, the content of a current frame could be very different from its reference pictures and may consequently lead to a more difficult motion compensation. The main idea of this paper is to process the input 2D video sequence in order to estimate the 3D geometry of the scene and then employ this data to virtually synthesize “geometrically compensated” reference pictures. Since these virtual reference pictures are more similar to the current frame, motion estimation and consequently coding efficiency could be enhanced. The proposed method is tested over six different video sequences and around 11% bitrate reduction is achieved compared to the High Efficiency Video Coding (HEVC) standard.

Download a Pre-Print


Supplementary Material

Sequences, 3D Models, Point Clouds and BD-PSNRs

NOTICE FOR IEEE PUBLICATIONS: © IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works, must be obtained from the IEEE. Contact: Manager, Copyrights and Permissions / IEEE Service Center / 445 Hoes Lane / P.O. Box 1331 / Piscataway, NJ 08855-1331, USA. Telephone: + Intl. 908-562-3966.