Scene Understanding using Lightfield Cameras and Neural Networks

 

Description

 

Lightfield cameras can be represented as an array of cameras. This array of cameras can be represented compactly by an intrinsic matrix with only 6 parameters [Dansereau13,Zhang18]. However, approaches that allow to perform self-calibration of these cameras are scarce. Furthermore, these cameras allow to recover the position of a point in the scene in a single acquisition. Only recently started to appear deep neural network strategies to recover the depth information of the scene.

 

In this work, we aim at developing a self-calibration procedure for lightfield cameras for example based on lines in the scene [Hartley2003], and study the reconstruction capabilities of a recently proposed deep neural network [Shin18].

 

Objectives

-         Study the camera models for lightfield cameras.

-         Calibrate lightfield camera from lines in the scene.

-         Reconstruct scene using epipolar plane images geometry and neural networks.

 

References:

[Dansereau13] Dansereau, Donald G., Oscar Pizarro, and Stefan B. Williams. "Decoding, calibration and rectification for lenselet-based plenoptic cameras." Proceedings of the IEEE conference on computer vision and pattern recognition. 2013.
[Zhang18] Zhang, Qi, et al. "A Generic Multi-Projection-Center Model and Calibration Method for Light Field Cameras." IEEE transactions on pattern analysis and machine intelligence (2018).

[Hartley2003] Hartley, Richard, and Andrew Zisserman. Multiple view geometry in computer vision. Cambridge university press, 2003.

[Shin18] Shin, Changha, et al. "EPINET: A Fully-Convolutional Neural Network Using Epipolar Geometry for Depth from Light Field Images." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018.

 

Requirements (grades, required courses, etc):

Current average grade >= 15

 

Place for conducting the work-proposal:

ISR / IST