TR2010-024

Pose Estimation in Heavy Clutter Using a Multi-Flash Camera


    •  Liu, M.-Y., Tuzel, C.O., Veeraraghavan, A.N., Chellappa, R., Agrawal, A.K., Okuda, H., "Pose Estimation in Heavy Clutter Using a Multi-Flash Camera", IEEE International Conference on Robotics and Automation (ICRA), May 2010.
      BibTeX TR2010-024 PDF
      • @inproceedings{Liu2010may,
      • author = {Liu, M.-Y. and Tuzel, C.O. and Veeraraghavan, A.N. and Chellappa, R. and Agrawal, A.K. and Okuda, H.},
      • title = {Pose Estimation in Heavy Clutter Using a Multi-Flash Camera},
      • booktitle = {IEEE International Conference on Robotics and Automation (ICRA)},
      • year = 2010,
      • month = may,
      • url = {https://www.merl.com/publications/TR2010-024}
      • }
  • Research Areas:

    Computer Vision, Robotics

Abstract:

We propose a novel solution to object detection, localization and pose estimation with applications in robot vision. The proposed method is especially applicable when the objects of interest may not be richly textured and are immersed in heavy clutter. We show that a multi-flash camera (MFC) provides accurate separation of depth edges and texture edges in such scenes. Then, we reformulate the problem, as one finding matches between the depth edges obtained in one or more MFC images to the rendered depth edges that are computed offline using 3D CAD model of the objects. In order to facilitate accurate matching of these binary depth edge maps, we introduce a novel cost function that respects both the position and the local orientation of each edge pixel. This cost function is significantly superior to traditional Chamfer cost and leads to accurate matching even in heavily cluttered scenes where traditional methods are unreliable. We present a sub-linear time algorithm to compute the cost function using techniques from 3D distance transforms and integral images. Finally, we also propose a multi-view based pose-refinement algorithm to improve the estimated pose. We implemented the algorithm on an industrial robot arm and obtained location and angular estimation accuracy of the order of 1 mm and 2 degrees respectively for a variety of parts with minimal texture.

 

  • Related News & Events

    •  NEWS    MERL researcher, Oncel Tuzel, gives keynote talk at 2016 International Symposium on Visual Computing
      Date: December 14, 2015 - December 16, 2015
      Where: Las Vegas, NV, USA
      Research Area: Machine Learning
      Brief
      • MERL researcher, Oncel Tuzel, gave a keynote talk at 2016 International Symposium on Visual Computing in Las Vegas, Dec. 16, 2015. The talk was titled: "Machine vision for robotic bin-picking: Sensors and algorithms" and reviewed MERL's research in the application of 2D and 3D sensing and machine learning to the problem of general pose estimation.

        The talk abstract was: For over four years, at MERL, we have worked on the robot "bin-picking" problem: using a 2D or 3D camera to look into a bin of parts and determine the pose, 3D rotation and translation, of a good candidate to pick up. We have solved the problem several different ways with several different sensors. I will briefly describe the sensors and the algorithms. In the first half of the talk, I will describe the Multi-Flash camera, a 2D camera with 8 flashes, and explain how this inexpensive camera design is used to extract robust geometric features, depth edges and specular edges, from the parts in a cluttered bin. I will present two pose estimation algorithms, (1) Fast directional chamfer matching--a sub-linear time line matching algorithm and (2) specular line reconstruction, for fast and robust pose estimation of parts with different surface characteristics. In the second half of the talk, I will present a voting-based pose estimation algorithm applicable to 3D sensors. We represent three-dimensional objects using a set of oriented point pair features: surface points with normals and boundary points with directions. I will describe a max-margin learning framework to identify discriminative features on the surface of the objects. The algorithm selects and ranks features according to their importance for the specified task which leads to improved accuracy and reduced computational cost.
    •  
    •  NEWS    ICRA 2010: 3 publications by Yuichi Taguchi, Amit K. Agrawal, C. Oncel Tuzel, Tim K. Marks and others
      Date: May 3, 2010
      Where: IEEE International Conference on Robotics and Automation (ICRA)
      MERL Contact: Tim K. Marks
      Research Area: Computer Vision
      Brief
      • The papers "Pose Estimation in Heavy Clutter Using a Multi-Flash Camera" by Liu, M.-Y., Tuzel, C.O., Veeraraghavan, A.N., Chellappa, R., Agrawal, A.K. and Okuda, H., "Rao-Blackwellized Particle Filtering for Probing-based 6-DOF Localization in Robotic Assembly" by Taguchi, Y., Marks, T.K. and Okuda, H. and "Multi-Class Batch-mode Active Learning for Image Classification" by Joshi, A.J., Porikli, F. and Papanikolopoulos, N. were presented at the IEEE International Conference on Robotics and Automation (ICRA).
    •