This Blog is maintained by the Robot Perception and Learning lab at CSIE, NTU, Taiwan. Our scientific interests are driven by the desire to build intelligent robots and computers, which are capable of servicing people more efficiently than equivalent manned systems in a wide variety of dynamic and unstructured environments.
Tuesday, May 29, 2012
Lab Meeting May 30th, 2012 (Wei-Shin): Progress Report
I will present my current progress on chair manipulation at lab meeting. (Borrowed account for posting.)
Tuesday, May 22, 2012
Lab Meeting May 22th, 2012 (Mark):Strong supervision from weak annotation: Interactive training of deformable part models
We propose a framework for large scale learning and annotation of
structured models. The system interleaves interactive labeling (where
the current model is used to semi-automate the labeling of a new
example) and online learning (where a newly labeled example is used to
update the current model parameters). This framework is scalable to
large datasets and complex image models and is shown to have excellent
theoretical and practical properties in terms of train time, optimality
guarantees, and bounds on the amount of annotation effort per image. We
apply this framework to part-based detection, and introduce a novel
algorithm for interactive labeling of deformable part models. The
labeling tool updates and displays in real-time the maximum likelihood
location of all parts as the user clicks and drags the location of one
or more parts. We demonstrate that the system can be used to efficiently
and robustly train part and pose detectors on the CUB Birds-200-a
challenging dataset of birds in unconstrained pose and environment.
paper link
paper link
Tuesday, May 01, 2012
[Robot Perception and Learning] Meeting 2012/05/02 (Andi):Energy Based Multiple Model Fitting for Non-Rigid Structure from Motion
Energy Based Multiple Model Fitting for Non-Rigid Structure from Motion
Authors: Chris Russell, Joao Fayad, Lourdes Agapito
From: CVPR '11
Abstract:
In this paper we reformulate the 3D reconstruction of deformable surfaces from monocular video sequences as a labeling problem. We solve simultaneously for the assignment of feature points to multiple local deformation models and the fitting of models to points to minimize a geometric cost, subject to a spatial constraint that neighboring points should also belong to the same model.
Piecewise reconstruction methods rely on features shared between models to enforce global consistency on the 3D surface. To account for this overlap between regions, we consider a super-set of the classic labeling problem in which a set of labels, instead of a single one, is assigned to each variable. We propose a mathematical formulation of this new model and show how it can be efficiently optimized with a variant of -expansion. We demonstrate how this framework can be applied to Non-Rigid Structure from Motion and leads to simpler explanations of the same data. Compared to existing methods run on the same data, our approach has up to half the reconstruction error, and is more robust to over-fitting and outliers.
Link: get paper here