Abstract
The task of picking up and handling objects is a great robotic challenge. Estimating the best point where the gripper fingers should come into contact with the object before performing the pick-up task is essential to avoid failures. This study presents a new approach to estimating the grasping pose of objects using a database generated by a gripper through its proximity sensors. The grasping pose estimation simulates the points where the fingers should be positioned to obtain the best grasp of the object. In this study, we used a database generated by a reconfigurable gripper with three fingers that can scan different objects through distance sensors attached to the fingers and palm of the gripper. The grasping pose of 13 objects was estimated, which were classified according to their geometries. The analysis of the grasping pose estimates considered the versatility of the gripper used. These object grasping pose estimates were validated using the CoppeliaSim software, where it was possible to configure the gripper according to the estimates generated and pick up the objects using just two or three fingers of the reconfigurable gripper.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Brahmbhatt, S., Tang, C., Twigg, C.D., Kemp, C.C., Hays, J.: Contactpose: a dataset of grasps with object contact and hand pose. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 361–378. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58601-0_22
Cao, H., Chen, G., Li, Z., Hu, Y., Knoll, A.: NeuroGrasp: multimodal neural network with Euler region regression for neuromorphic vision-based grasp pose estimation. IEEE Trans. Instrum. Meas. 71, 1–11 (2022)
Corona, E., Pumarola, A., Alenyà, G., Moreno-Noguer, F., Rogez, G.: Ganhand: predicting human grasp affordances in multi-object scenes. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2020)
Cox, R., Correll, N.: Merging local and global 3D perception using contact sensing. AAAI Spring Symposium - Technical report (2017)
Dong, Z., Tian, H., Bao, X., Yan, Y., Chen, F.: GraspVDN: scene-oriented grasp estimation by learning vector representations of grasps. Complex Intell. Syst. 8, 2911–2922 (2022)
Du, G., Wang, K., Lian, S., Zhao, K.: Vision-based robotic grasping from object localization, object pose estimation to grasp estimation for parallel grippers: a review. Artif. Intell. Rev. 54, 1677–1734 (2021)
Eppner, C., Mousavian, A., Fox, D.: Acronym: a large-scale grasp dataset based on simulation. In: Proceedings - IEEE International Conference on Robotics and Automation (2021)
Fang, H.S., Wang, C., Gou, M., Lu, C.: Graspnet-1billion: a large-scale benchmark for general object grasping. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2020)
Guo, N., Zhang, B., Zhou, J., Zhan, K., Lai, S.: Pose estimation and adaptable grasp configuration with point cloud registration and geometry understanding for fruit grasp planning. Comput. Electron. Agric. 179, 105818 (2020)
Kroon, D.J.: Polygon2voxel (2023). https://www.mathworks.com/matlabcentral/fileexchange/24086-polygon2voxel. Accessed 30 June 2023
Kulecki, B., Młodzikowski, K., Staszak, R., Belter, D.: Practical aspects of detection and grasping objects by a mobile manipulating robot. Ind. Robot 48, 688–699 (2021)
Li, Y., Kong, T., Chu, R., Li, Y., Wang, P., Li, L.: Simultaneous semantic and collision learning for 6-DoF grasp pose estimation. In: IEEE International Conference on Intelligent Robots and Systems (2021)
Liu, M., Pan, Z., Xu, K., Ganguly, K., Manocha, D.: Generating grasp poses for a high-DoF gripper using neural networks. In: IEEE International Conference on Intelligent Robots and Systems (2019)
Pereira, J.H., Joventino, C.F., Fabro, J.A., Oliveira, A.S.: Non-contact tactile perception for hybrid-active gripper. IEEE Robot. Autom. Lett. 8, 3047–3054 (2023)
Vohra, M., Prakash, R., Behera, L.: Real-time grasp pose estimation for novel objects in densely cluttered environment. 2019 28th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN 2019 (2019)
Yu, Y., Cao, Z., Liang, S., Geng, W., Yu, J.: A novel vision-based grasping method under occlusion for manipulating robotic system. IEEE Sens. J. 20, 10996–11006 (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Pereira, J.H.M., Joventino, C.F., Fabro, J.A., de Oliveira, A.S. (2024). Estimation of Optimal Gripper Configuration Through an Embedded Array of Proximity Sensors. In: Filipe, J., Röning, J. (eds) Robotics, Computer Vision and Intelligent Systems. ROBOVIS 2024. Communications in Computer and Information Science, vol 2077. Springer, Cham. https://doi.org/10.1007/978-3-031-59057-3_26
Download citation
DOI: https://doi.org/10.1007/978-3-031-59057-3_26
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-59056-6
Online ISBN: 978-3-031-59057-3
eBook Packages: Computer ScienceComputer Science (R0)