1
Using Vision for Pre- and Post-Grasping Object Localization for Soft Hands Changhyun Choi, Joseph DelPreto, and Daniela Rus Abstract—In this poster, we present soft hands guided by an RGB-D object perception algorithm which is capable of localizing the pose of an object before and after grasping. The soft hands can perform manipulation operations such as grasping and connecting two parts. The flexible soft grippers grasp objects reliably under high uncertainty but the poses of the objects after grasping are subject to high uncertainty. Visual sensing ameliorates the increased uncertainty by means of in- hand object localization. The combination of soft hands and visual object perception enables our Baxter robot, augmented with soft hands, to perform object assembly tasks which require high precision. The effectiveness of our approach is validated by comparing it to the Baxter’s original hard hands with and without the in-hand object localization. I. I NTRODUCTION An important prerequisite for object manipulation is esti- mating the pose of an object and coping with the uncertainty of the pose estimates. Various sensing modalities, such as proprioception [1], [2], visual exteroception [3], and con- tact/force sensing [4] have been employed. Visual sensing allows passive perception as it does not require contact, and is thus useful in the pre-grasping phase. Tactile, contact, force, and proprioceptive sensing modalities are useful when robots interact with objects in the post-grasping phase. The pose of a grasped object can be quite uncertain as the act of grasping tends to move the object and increase the uncertainty. Many prior works have combined vision and contact to decrease uncertainty [5]. Soft grippers are more compliant and easier to control than their hard counterparts [6]. The flexible materials of soft hands enable compliance with discrepancy between their belief space and the real environment; this compliance allows soft hands to be more tolerant of errors in the pose estimates of objects. Softness, however, often reduces the confidence of the object state in the gripper since the pose of the object is more uncertain due to the flexibility of the soft fingers. In- hand object localization is thus needed for advanced object manipulations requiring accurate pose. The goal of this work is to develop a reliable object manipulation system using soft hands and visual pose feed- back and to evaluate its effectiveness. Fig. 1a illustrates our system setup in which the Baxter robot is augmented with two soft hands and an RGB-D sensor. We use vision for localizing objects presented to the robot on a tabletop and This work is to appear at International Symposium on Experimental Robotics (ISER), Tokyo, Japan, Oct. 2016. The authors are with Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, MA 02139, USA cchoi, delpreto, [email protected] then determining the pose of a grasped object in the hand. Fig. 1b shows one of our soft hands, which is composed of four pneumatically actuated fingers [2]. An RGB-D sensor is employed to localize objects in the workspace of the robot in the pre-grasping phase and to detect soft fingers and a grasped object in the post-grasping phase. Our approach does not rely on proprioceptive force sensing, yet it is capable of assembly operations requiring precision. To the best of our knowledge, this is the first attempt to use a vision-based object localization for soft hands capable of assembly tasks. The full poster will describe the details of our technical approach and present the experimental results. RGB-D Sensor Soft Hands Assembly Parts (a) System setup (b) Soft hand Fig. 1: System overview. Our system is composed of the Baxter robot augmented with two soft hands and an RGB-D sensor. Assembly parts are randomly placed on the table, so the positions and orientations of the parts are unknown. The RGB-D sensor localizes the parts on the table and inside the hand during the pre-grasping and post-grasping phases, respectively. The RGB channels are used for identification of the soft fingers, while the depth channel is employed for depth-based object localization. REFERENCES [1] M. T. Mason, A. Rodriguez, S. S. Srinivasa, and A. S. Vazquez, “Autonomous manipulation with a general-purpose simple hand,” International Journal of Robotics Research, vol. 31, no. 5, pp. 688–703, 2012. [2] B. S. Homberg, R. K. Katzschmann, M. R. Dogar, and D. Rus, “Haptic identification of objects using a modular soft robotic gripper,” in Proc. IEEE/RSJ Int’l Conf. Intelligent Robots Systems (IROS), Sept. 2015, pp. 1698–1705. [3] A. Collet, M. Martinez, and S. S. Srinivasa, “The MOPED framework: Object recognition and pose estimation for manipulation,” International Journal of Robotics Research, vol. 30, no. 10, pp. 1284–1306, 2011. [4] J. Bimbo, S. Luo, K. Althoefer, and H. Liu, “In-hand Object Pose Estimation using Covariance-Based Tactile To Geometry Matching,” IEEE Robotics and Automation Letters, vol. 1, no. 1, 2016. [5] T. Schmidt, K. Hertkorn, R. Newcombe, Z. Marton, M. Suppa, and D. Fox, “Depth-Based Tracking with Physical Constraints for Robot Manipulation,” in Proceedings of IEEE International Conference on Robotics and Automation (ICRA), 2015. [6] R. Deimel and O. Brock, “A compliant hand based on a novel pneumatic actuator,” in Proceedings of IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2013, pp. 2047–2053.

Using Vision for Pre- and Post-Grasping Object Localizationpeople.csail.mit.edu/cchoi/pub/choi16irosw_softhand.pdf · Using Vision for Pre- and Post-Grasping Object Localization for

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Using Vision for Pre- and Post-Grasping Object Localizationpeople.csail.mit.edu/cchoi/pub/choi16irosw_softhand.pdf · Using Vision for Pre- and Post-Grasping Object Localization for

Using Vision for Pre- and Post-Grasping Object Localizationfor Soft Hands†

Changhyun Choi, Joseph DelPreto, and Daniela Rus

Abstract— In this poster, we present soft hands guided byan RGB-D object perception algorithm which is capable oflocalizing the pose of an object before and after grasping.The soft hands can perform manipulation operations such asgrasping and connecting two parts. The flexible soft grippersgrasp objects reliably under high uncertainty but the poses ofthe objects after grasping are subject to high uncertainty. Visualsensing ameliorates the increased uncertainty by means of in-hand object localization. The combination of soft hands andvisual object perception enables our Baxter robot, augmentedwith soft hands, to perform object assembly tasks which requirehigh precision. The effectiveness of our approach is validatedby comparing it to the Baxter’s original hard hands with andwithout the in-hand object localization.

I. INTRODUCTION

An important prerequisite for object manipulation is esti-mating the pose of an object and coping with the uncertaintyof the pose estimates. Various sensing modalities, such asproprioception [1], [2], visual exteroception [3], and con-tact/force sensing [4] have been employed. Visual sensingallows passive perception as it does not require contact, andis thus useful in the pre-grasping phase. Tactile, contact,force, and proprioceptive sensing modalities are useful whenrobots interact with objects in the post-grasping phase. Thepose of a grasped object can be quite uncertain as theact of grasping tends to move the object and increase theuncertainty. Many prior works have combined vision andcontact to decrease uncertainty [5].

Soft grippers are more compliant and easier to controlthan their hard counterparts [6]. The flexible materials ofsoft hands enable compliance with discrepancy between theirbelief space and the real environment; this compliance allowssoft hands to be more tolerant of errors in the pose estimatesof objects. Softness, however, often reduces the confidenceof the object state in the gripper since the pose of the objectis more uncertain due to the flexibility of the soft fingers. In-hand object localization is thus needed for advanced objectmanipulations requiring accurate pose.

The goal of this work is to develop a reliable objectmanipulation system using soft hands and visual pose feed-back and to evaluate its effectiveness. Fig. 1a illustrates oursystem setup in which the Baxter robot is augmented withtwo soft hands and an RGB-D sensor. We use vision forlocalizing objects presented to the robot on a tabletop and

†This work is to appear at International Symposium on ExperimentalRobotics (ISER), Tokyo, Japan, Oct. 2016.

The authors are with Computer Science & Artificial Intelligence Lab,Massachusetts Institute of Technology, Cambridge, MA 02139, USAcchoi, delpreto, [email protected]

then determining the pose of a grasped object in the hand.Fig. 1b shows one of our soft hands, which is composed offour pneumatically actuated fingers [2]. An RGB-D sensor isemployed to localize objects in the workspace of the robotin the pre-grasping phase and to detect soft fingers and agrasped object in the post-grasping phase. Our approach doesnot rely on proprioceptive force sensing, yet it is capable ofassembly operations requiring precision. To the best of ourknowledge, this is the first attempt to use a vision-basedobject localization for soft hands capable of assembly tasks.

The full poster will describe the details of our technicalapproach and present the experimental results.

RGB-D Sensor

Soft Hands

Assembly Parts

(a) System setup (b) Soft handFig. 1: System overview. Our system is composed of the Baxter robotaugmented with two soft hands and an RGB-D sensor. Assembly parts arerandomly placed on the table, so the positions and orientations of the partsare unknown. The RGB-D sensor localizes the parts on the table and insidethe hand during the pre-grasping and post-grasping phases, respectively.The RGB channels are used for identification of the soft fingers, while thedepth channel is employed for depth-based object localization.

REFERENCES

[1] M. T. Mason, A. Rodriguez, S. S. Srinivasa, and A. S. Vazquez,“Autonomous manipulation with a general-purpose simple hand,”International Journal of Robotics Research, vol. 31, no. 5, pp.688–703, 2012.

[2] B. S. Homberg, R. K. Katzschmann, M. R. Dogar, and D. Rus, “Hapticidentification of objects using a modular soft robotic gripper,” in Proc.IEEE/RSJ Int’l Conf. Intelligent Robots Systems (IROS), Sept. 2015, pp.1698–1705.

[3] A. Collet, M. Martinez, and S. S. Srinivasa, “The MOPED framework:Object recognition and pose estimation for manipulation,” InternationalJournal of Robotics Research, vol. 30, no. 10, pp. 1284–1306, 2011.

[4] J. Bimbo, S. Luo, K. Althoefer, and H. Liu, “In-hand Object PoseEstimation using Covariance-Based Tactile To Geometry Matching,”IEEE Robotics and Automation Letters, vol. 1, no. 1, 2016.

[5] T. Schmidt, K. Hertkorn, R. Newcombe, Z. Marton, M. Suppa, andD. Fox, “Depth-Based Tracking with Physical Constraints for RobotManipulation,” in Proceedings of IEEE International Conference onRobotics and Automation (ICRA), 2015.

[6] R. Deimel and O. Brock, “A compliant hand based on a novelpneumatic actuator,” in Proceedings of IEEE International Conferenceon Robotics and Automation (ICRA). IEEE, 2013, pp. 2047–2053.