Multi-layer Scene Representation from Composed Focal Stacks

Reina Ishikawa, Hideo Saito, Denis Kalkofen, Shohei Mori

Research output: Contribution to journalArticlepeer-review

Abstract

Multi-layer images are a powerful scene representation for high-performance rendering in virtual/augmented reality (VR/AR). The major approach to generate such images is to use a deep neural network trained to encode colors and alpha values of depth certainty on each layer using registered multi-view images. A typical network is aimed at using a limited number of nearest views. Therefore, local noises in input images from a user-navigated camera deteriorate the final rendering quality and interfere with coherency over view transitions. We propose to use a focal stack composed of multi-view inputs to diminish such noises. We also provide theoretical analysis for ideal focal stacks to generate multi-layer images. Our results demonstrate the advantages of using focal stacks in coherent rendering, memory footprint, and AR-supported data capturing. We also show three applications of imaging for VR.
Original languageEnglish
Pages (from-to)4719-4729
Number of pages11
JournalIEEE Transactions on Visualization and Computer Graphics
Volume29
Issue number11
DOIs
Publication statusPublished - 1 Nov 2023

Keywords

  • Apertures
  • AR-supported imaging
  • Cameras
  • focal stack
  • Image reconstruction
  • Lenses
  • Light fields
  • Multi-layered scene representation
  • Rendering (computer graphics)
  • Three-dimensional displays
  • view synthesis

ASJC Scopus subject areas

  • Software
  • Signal Processing
  • Computer Vision and Pattern Recognition
  • Computer Graphics and Computer-Aided Design

Fingerprint

Dive into the research topics of 'Multi-layer Scene Representation from Composed Focal Stacks'. Together they form a unique fingerprint.

Cite this