11:30 am to 12:00 am
Event Location: NSH 3002
Abstract: Society is becoming more automated with robots beginning to perform most tasks in factories and starting to help out in home and office environments. Arguably, one of the most important functions of robots is the ability to manipulate objects in their environment to accomplish primitive tasks. Because the space of possible robot designs, sensor modalities, and target tasks is huge, researchers end up having to manually create many models, databases, and programs for their specific task, an effort that is repeated whenever the task changes. This thesis introduces a manipulation framework that, given a specification for a robot and a task, can automatically construct the necessary databases and programs required for the robot to reliably execute the task. It addresses problems in three main areas critical for manipulation.
We present a geometric-based planning framework that analyses all necessary modalities of manipulation planning and offers efficient algorithms to solve each. This allows identification of the necessary information needed from the task and robot specifications. Using this set of analyses, the construction process then builds a planning knowledge-base that allows the planners to make informative geometric decisions about the structure of the scene and the robot’s goals. We show how to efficiently generate and query the information for planners.
In order to reliably complete the task, we present efficient algorithms that consider the visibility of objects in cameras when choosing manipulation goals. We show real-world results with robots using cameras attached to their grippers to boost accuracy of the detected objects and to reliably complete the tasks. Furthermore, we use the visibility theory to develop a completely automated extrinsic camera calibration method and a present a new measure for computing confidence of the final results.
For the perception side of manipulation, we present a vision-centric database that can analyze a rigid object’s surface for stable and discriminable features to use in pose extraction programs. Furthermore, we show work towards a new object pose extraction algorithm that does not rely on 2D/3D feature correspondences and thus reduces the early-commitment problem plaguing the generality of traditional vision-based pose extraction algorithms.
Committee:Takeo Kanade, Co-chair
James Kuffner, Co-chair
Paul Rybski
Kei Okada, University of Tokyo