Abstract
This paper presents a robotic demonstrator for manipulation and distribution of objects. The demonstrator relies on robust 3D vision-based solutions for navigation, object detection and detection of graspable surfaces using the rc_visard, a self-registering stereo vision sensor. Suitable software modules were developed for SLAM and for model-free suction gripping. The modules run onboard the sensor, which enables creating the presented demonstrator as a standalone application that does not require an additional host PC. The modules are interfaced with ROS, which allows a quick implementation of a fully functional robotic application.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
References
Bay H, Ess A, Tuytelaars T, Gool L (2008) Speeded-up robust features (SURF). Comput Vis Image Underst 110(3):346–359
Bohg J, Morales A, Asfour T, Kragic D (2014) Data-driven grasp synthesis—a survey. IEEE Trans Robot 30(2):289–309
Correll N, Bekris K, Berenson D, Brock O, Causo A, Hauser K, Okada K, Rodriguez A, Romano J, Wurman P (2018) Analysis and observations from the first Amazon picking challenge. IEEE Trans Autom Sci Eng 15(1):172–188
DHL Trend Research (2016) Robotics in logistics: a DPDHL perspective on implications and use cases for the logistics industry. DHL Customer Solutions & Innovation
Döllinger A, Larsson T (2005) Selection of automated order picking systems. Master thesis, Chalmers University of Technology, Sweden
EHI Retail Institute (2017) Robotics 4 retail: status quo, potenziale und herausforderungen. EHI-Whitepaper
Falco J, Sun Y, Roa M (2018) Robotic grasping and manipulation competition: competitor feedback and lessons learned. In: Sun Y, Falco J (eds) Robotic grasping and manipulation: first robotic grasping and manipulation challenge. Springer, Berlin, pp 180–189
Galvez-Lopez D, Tardos JD (2012) Bags of binary words for fast place recognition in image sequences. IEEE Trans Robot 28(5):1188–1197
Gambaro E, Emmerich C, Muenster K, Schaller R, Suppa M (2018) Verfahren zum erstellen eines objektmodells zum greifen eines objekts, computerlesbares speichermedium und robotersystem. In: German Patent Office
Gualtierei M, Pas A, Platt R (2018) Pick and place without geometric object models. In: Proceeding of IEEE international conference on robotics and automation—ICRA, Brisbane, Australia, pp 7433–7440
Hillenbrand U (2008) Pose clustering from stereo data. In: Proceedings of VISAPP international workshop on robotic perception, Madeira, Portugal, pp 23–32
Hinterstoisser S, Lepetit V, Ilic S, Holzer S, Bradski G, Konolige K, Navab N (2012) Model based training, detection and pose estimation of texture-less 3D objects in heavily cluttered scenes. In: Proceedings of ACCV Asian conference on computer vision, Daejeon, Korea, pp 548–562
Hirschmüller H (2008) Stereo processing by semi-global matching and mutual information. IEEE Trans Pattern Anal Mach Intell 30(2):328–341
Lowe D (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91–110
Mahler J, Matl M, Liu X, Li A, Gealy D, Goldberg K (2018) Dex-Net 3.0: computing robust vacuum suction grasp targets in point clouds using a new analytic model and deep learning. In: Proceedings of IEEE international conference on robotics and automation—ICRA, Brisbane, Australia, pp 5620–5627
Mur R, Tardos J (2017) ORB-SLAM2: an open-source SLAM system for monocular, stereo and RGB-D cameras. IEEE Trans Robot 33(5):1255–1262
Olson E (2011) AprilTag: a robust and flexible visual fiducial system. In: Proceedings of IEEE international conference on robotics and automation—ICRA, Shanghai, China, pp 3400–3407
Pauwels K, Kragic D (2015) Simtrack: a simulation-based framework for scalable real-time object pose detection and tracking. In: Proceedings of IEEE/RSJ international conference on intelligent robots and systems—IROS, Hamburg, Germany, pp 1300–1307
Porges O, Stouraitis T, Borst C, Roa MA (2014) Reachability and capability analysis for manipulation tasks. In: ROBOT2013: first Iberian robotics conference. Springer, Madrid, Spain, pp 703–718
Redmon J, Farhadi A (2018) YOLOv3: an incremental improvement. arXiv:1804.02767
Roy N, Newman M, Srinivasa S (2013) Recognition and pose estimation of rigid transparent objects with a kinect sensor. In: Proceedings of robotics science and systems—RSS, Berlin, Germany
Rusu R, Blodow N, Beetz M (2009) Fast point feature histograms (FPFH) for 3D registration. In: Proceedings of IEEE international conference on Robotics and Automation—ICRA, Kobe, Japan, pp 3212–3217
Rusu R, Cousins S (2011) 3D is here: point cloud library (PCL). In: Proceedings of IEEE international conference on robotics and automation—ICRA, Shanghai, China
Sepp W, Fuchs S, Hirzinger G (2006) Hierarchical featureless tracking for position-based 6-DoF visual servoing. In: Proceedings of IEEE/RSJ international conference on intelligent robots and systems—IROS, Beijing, China, pp 4310–4315
Styleintelligence (2018) Market report: goods-to-person ecommerce fulfilment robotics
Xiang Y, Schmidt T, Narayanan V, Fox D (2018) PoseCNN: a convolutional neural network for 6D object pose estimation in cluttered scenes. In: Proceedings of robotics: science and systems—RSS, Pittsburgh, USA
Xiao J, Hays J, Ehinger K, Oliva A, Torralba A (2010) SUN database: large-scale scene recognition from abbey to zoo. In: Proceedings of IEEE conference on computer vision and pattern recognition—CVPR, San Francisco, USA, pp 3485–3492
Funding
This project was partially funded by the European Union’s Horizon 2020 research and innovation programme under the project ROSIN, Grant agreement no. 732287, with the FTP (Focused Technical Project) VISARD4ROS.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Roa-Garzón, M.A., Gambaro, E.F., Florek-Jasinska, M. et al. Vision-Based Solutions for Robotic Manipulation and Navigation Applied to Object Picking and Distribution. Künstl Intell 33, 171–180 (2019). https://doi.org/10.1007/s13218-019-00588-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13218-019-00588-z