Octomap by K. Wurm et al, MS Kinect

 

MSc dissertation proposal 2013/2014

 

Integrating Color-Depth Images into World Representations

 

 

Introduction:

 

Recent vision sensors, namely the Microsoft Kinect [MS-Kinect] comprising not only video but also depth information, promise to be sensors capable to provide the required high quality data allowing to construct from scratch complex scene representations. This MSc project proposal is precisely concerned with designing scene representations integrating video and depth information.

 

 

Objectives:

 

The objectives of this work are twofold: (i) Building scene representations encompassing multiple data acquisitions done with the Kinect camera, and (ii) Creating fast browsing methodologies of the scene representations.

 

 

Detailed description:

 

Creating scene representations is nowadays facilitated by the recent introduction in the market of the MS Kinect camera [MS-Kinect] which provides not only visual information of the scenario, but also depth information (3D information). Integrating this 3D information is still an issue due mainly to the large amounts of data involved and the required (large) computational resources.

 

Some recent research works show that by choosing the proper scene representations it is possible to do data acquisition, processing and integration with a standard PC. One such reference is the work by Wurm et al [Wurm10], where a mobile robot maps a number of campuses. Another reference is the work by Shen et al [Shen11], where a flying robot (quadrotor) maps a multi-floor scenario using a Kinect camera (click here to see a video).

 

The referred works give conceptual starting points and software libraries which facilitate doing experiments. This constitutes therefore the starting point of the MSc project. Some other aspects involve doing data acquisition and displaying the resulting scenarios. An interesting point to explore involves using the acquired data for navigation experiments.

 

In summary, the work is organized in the following main steps:

1) data acquisition, implies moving the robot within the scene and capturing scene information

2) data integration into a single scene representation

3) demonstration of a virtual tour inside the scene representation or trajectory following with the robot

 

 

References:

 

[HRI2007] - 2nd ACM/IEEE International Conference on Human-Robot Interaction, http://hri2007.org/

 

[Gaspar00] Vision-based Navigation and Environmental Representations with an Omnidirectional Camera, José Gaspar, Niall Winters, José Santos-Victor, IEEE Transaction on Robotics and Automation, Vol 16, number 6, December 2000

 

[ISR-galery] Some images of unicycle type robots at ISR: see labmate, scouts, pioneers, ... in "Mini galeria de fotografias de projectos @ IST/ISR", http://users.isr.ist.utl.pt/~jag/infoforum/isr_galeria/

 

[MS-Kinect] http://en.wikipedia.org/wiki/Kinect

 

[Wurm10] Kai M. Wurm, Armin Hornung, Maren Bennewitz, Cyrill Stachniss, Wolfram Burgard, "OctoMap: A Probabilistic, Flexible, and Compact 3D Map Representation for Robotic Systems", in Proceedings of the ICRA 2010 Workshop on Best Practice in 3D Perception and Modeling for Mobile Manipulation, 2010, http://www.informatik.uni-freiburg.de/~wurm/papers/wurm10octomap.pdf

 

[Shen11] Shaojie Shen, Nathan Michael, and Vijay Kumar , "3D Indoor Exploration with a Computationally Constrained MAV", ICRA 2011. See also video in http://www.youtube.com/watch?v=cOeCZDBHrJs&feature=channel_video_title

 

 

Expected results:

 

At the end of the work the students will have enriched their knowledge in:

* Computer vision

* Building scene representations

 

 

Observations:

--

 

 

More MSc dissertation proposals on Computer and Robot Vision in:

 

http://omni.isr.ist.utl.pt/~jag