Paper ID | 3D-4.10 | ||
Paper Title | LEARNING TO COMPOSE 6-DOF OMNIDIRECTIONAL VIDEOS USING MULTI-SPHERE IMAGES | ||
Authors | Jisheng Li, Yuze He, Yubin Hu, Tsinghua University, China; Yuxing Han, Research Institute of Tsinghua University in Shenzhen, China; Jiangtao Wen, Tsinghua University, China | ||
Session | 3D-4: 3D Image and Video Processing | ||
Location | Area J | ||
Session Time: | Tuesday, 21 September, 13:30 - 15:00 | ||
Presentation Time: | Tuesday, 21 September, 13:30 - 15:00 | ||
Presentation | Poster | ||
Topic | Three-Dimensional Image and Video Processing: Image and video processing augmented and virtual reality | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Omnidirectional video is an essential component of Virtual Reality. Although various methods have been proposed to generate content that can be viewed with six degrees of freedom (6-DoF), existing systems usually involve complex depth estimation, image in-painting or stitching pre-processing. In this paper, we propose a system that uses a 3D ConvNet to generate a multi-sphere images (MSI) representation that can be experienced in 6-DoF VR. The system utilizes conventional omnidirectional VR camera footage directly without the need for a depth map or segmentation mask, thereby significantly simplifying the overall complexity of the 6-DoF omnidirectional video composition. By using a newly designed weighted sphere sweep volume (WSSV) fusing technique, our approach is compatible with most panoramic VR camera setups. A ground truth generation approach for high-quality artifact-free 6-DoF contents is proposed and can be used by the research and development community for 6-DoF content generation. |