Pioneer
P3-AT mobile robot (rightmost photos: config-1 at ISR/IST)
MSc dissertation proposal 2010/2011
Vision-based
Motion Control of a
Introduction:
"Although a few of the robots of tomorrow may resemble the
anthropomorphic devices seen in Star Wars, most will look nothing like the
humanoid C-3PO. In fact, as mobile peripheral devices become more and more
common, it may be increasingly difficult to say exactly what a robot is.
Because the new machines will be so specialized and ubiquitous--and look so
little like the two-legged automatons of science fiction--we probably will not
even call them robots. But as these devices become affordable to consumers,
they could have just as profound an impact on the way we work, communicate,
learn and entertain ourselves as the PC has had over the past 30 years."
Excerpt of A Robot in Every Home, Bill Gates,
ScientificAmerican.com, January 2007.
Future robots will be more than mere tools doing isolated tasks: they
will be quasi-team members whose tasks have to be integrated with those of
humans or other robots [HRI2007]. One of the key aspects of the integration is
collision free motion among the various robots and people working in the
environment. This work is therefore mainly concerned with designing motion controllers
capable of responding efficiently, while considering the own dynamic
limitations of the robots and the constraints to motion imposed by other robots
or people.
Objectives:
The objectives of this work are twofold: (i) Visual tracking of patterns by their colours
and/or shapes; (ii) Controlling a unicycle-type
robot in order to reach the visually tracked target-pattern. At the
end is expected to demonstrate a point to point navigation capability by a
mobile robot under the view of vision sensors.
Detailed description:
Controlling the motion of a mobile robot is still a very active
research-and-development topic mainly because the large variety of applications
that intrinsically implies a large diversity of locomotion methods and sensors.
In this work-proposal the robots are of the very common unicycle-type, i.e.
carts or cars moving in a two-dimensional world having two parallel driven
wheels, one mounted on each side of their centre, and one or two offset castors
to maintain balance. These vehicles allow simultaneous arbitrary rotation and
translation being only constrained in the sideways motion (the sideways motion
cannot be instantaneous). The main sensors to use are computer vision and odometry.
More specifically, the work-proposal consists of controlling one unicycle robot to navigate to a
target location following a reference path [Carona08]. The target location is
defined by an off-board video camera.
The path following, or tracking, can be aided by both the off-board and
on-board sensors. The controller must consider the dynamic
constraints of the mobile robot and re-adjust the motion according
to unexpected obstacles (e.g. other
robots or people). In other words, the on-board sensors will be allowed to a
certain extent to redefine the path to follow, based on the characteristics of
the robot and the state of the environment. The control methodologies may
involve some visual path following [Gaspar00], visual servoing
[VISP], or dynamic control [Aguiar04] as considered to be the most adequate for
the tasks at hand.
The work is therefore organized in the following main steps:
1) visual tracking of patterns marking the
robot and the target, based on colours and/or shapes
2) combining the sensors information in common reference
frames, using information fusion filters
3) controller design for the unicycle type
robot, in order to move it towards the target
References:
[Carona08] Control of Unicycle Type Robots: Tracking, Path Following and
Point Stabilization, Ricardo Carona, A. Pedro Aguiar, José Gaspar, in Proc. of IV Jornadas
de Engenharia Electrónica e
Telecomunicações e de Computadores,
pp180-185, November 2008, Lisbon, Portugal.
[HRI2007] - 2nd ACM/IEEE International Conference on Human-Robot
Interaction, http://hri2007.org/
[Aguiar04] "Pose Estimation of Autonomous Vehicles using Visual
Information: A Minimum-Energy Estimator Approach", A. Pedro Aguiar and João P. Hespanha, in Proc. of IAV2004 - 5th IFAC/EURON Symposium on
Intelligent Autonomous Vehicles, Lisbon, Portugal, Jul. 2004.
[Gaspar00] Vision-based Navigation and Environmental Representations
with an Omnidirectional Camera, José Gaspar, Niall Winters, José Santos-Victor, IEEE Transaction on Robotics and Automation, Vol 16, number 6, December 2000
[ISR-galery] Some images of unicycle type
robots at ISR: see labmate, scouts, pioneers, ... in "Mini galeria
de fotografias de projectos
@ IST/ISR", http://users.isr.ist.utl.pt/~jag/infoforum/isr_galeria/
[OpenCV] Open Computer Vision Library, http://sourceforge.net/projects/opencv/
[VISP] Visual servoing videos, http://www.irisa.fr/lagadic/visp/video.html
Expected results:
At the end of the work the students will have enriched their knowledge
in:
* Computer vision
* Control of unicycle type vehicles
Examples of expected demonstrations in simulated and/or real
environments:
* remotely controlling the unicycle-type robot
by pointing to one reference target-object to be reached by the robot
* guiding the robot towards a target, while the robot negotiates its own
navigation in order to avoid collisions with other robots
Observations:
--
More MSc dissertation
proposals on Computer and Robot Vision in:
http://omni.isr.ist.utl.pt/~jag