Environment modeling and localization from datasets of omnidirectional scenes using machine learning techniques

Neural Computing and Applications - Tập 35 - Trang 16487-16508 - 2023
Sergio Cebollada1,2, Luis Payá1, Adrián Peidró1, Walterio Mayol3, Oscar Reinoso1,2
1Department of Systems Engineering and Automation, Miguel Hernández University, Elche, Spain
2Valencian Graduate School and Research Network for Artificial Intelligence, Valencian Community, Spain
3Department of Computer Science, University of Bristol, Bristol, UK

Tóm tắt

This work presents a framework to create a visual model of the environment which can be used to estimate the position of a mobile robot by means of artificial intelligence techniques. The proposed framework retrieves the structure of the environment from a dataset composed of omnidirectional images captured along it. These images are described by means of global-appearance approaches. The information is arranged in two layers, with different levels of granularity. The first layer is obtained by means of classifiers and the second layer is composed of a set of data fitting neural networks. Subsequently, the model is used to estimate the position of the robot, in a hierarchical fashion, by comparing the image captured from the unknown position with the information in the model. Throughout this work, five classifiers are evaluated (Naïve Bayes, SVM, random forest, linear discriminant classifier and a classifier based on a shallow neural network) along with three different global-appearance descriptors (HOG, gist, and a descriptor calculated from an intermediate layer of a pre-trained CNN). The experiments have been tackled with some publicly available datasets of omnidirectional images captured indoors with the presence of dynamic changes. Several parameters are used to assess the efficiency of the proposal: the ability of the algorithm to estimate coarsely the position (hit ratio), the average error (cm) and the necessary computing time. The results prove the efficiency of the framework to model the environment and localize the robot from the knowledge extracted from a set of omnidirectional images with the proposed artificial intelligence techniques.

Tài liệu tham khảo

Amorós F, Payá L, Marín JM, Reinoso O (2018) Trajectory estimation and optimization through loop closure detection, using omnidirectional imaging and global-appearance descriptors. Expert Syst Appl 102:273–290. https://doi.org/10.1016/j.eswa.2018.02.042 Amorós F, Payá L, Mayol-Cuevas W, Jiménez LM, Reinoso O (2020) Holistic descriptors of omnidirectional color images and their performance in estimation of position and orientation. IEEE Access 8:81822–81848. https://doi.org/10.1109/access.2020.2990996 Ballesta M, Payá L, Cebollada S, Reinoso O, Murcia F (2021) A CNN regression approach to mobile robot localization using omnidirectional images. Appl Sci. https://doi.org/10.3390/app11167521 Barshan B, Ayrulu B, Utete SW (2000) Neural network-based target differentiation using sonar for robotics applications. IEEE Trans Robot Autom 16(4):435–442. https://doi.org/10.1109/70.864239 Bay H, Tuytelaars T, Gool LV (2006) Surf: speeded up robust features. In: European conference on computer vision, Springer, pp 404–417. https://doi.org/10.1007/11744023_32 Cebollada S, Payá L, Flores M, Peidró A, Reinoso O (2021) A state-of-the-art review on mobile robotics tasks using artificial intelligence and visual data. Expert Syst Appl 167:114195. https://doi.org/10.1016/j.eswa.2020.114195 Cebollada S, Payá L, Flores M, Román V, Peidró A, Reinoso O (2020) A deep learning tool to solve localization in mobile autonomous robotics. In: ICINCO 2020, 17th international conference on informatics in control, automation and robotics, INSTICC, pp 232–241 (Online streaming, 7–9 July 2020) Cebollada S, Payá L, Mayol W, Reinoso O (2019) Evaluation of clustering methods in compression of topological models and visual place recognition using global appearance descriptors. Appl Sci 9(3):377. https://doi.org/10.3390/app9030377 Cebollada S, Payá L, Román V, Reinoso O (2019) Hierarchical localization in topological models under varying illumination using holistic visual descriptors. IEEE Access 7:49580–49595. https://doi.org/10.1109/ACCESS.2019.2910581 Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273–297. https://doi.org/10.1007/BF00994018 Cruz Ulloa C, Prieto Sánchez G, Barrientos A, Del Cerro J (2021) Autonomous thermal vision robotic system for victims recognition in search and rescue missions. Sensors 21(21):7346. https://doi.org/10.3390/s21217346 da Silva SPP, da Nbrega RVM, Medeiros AG, Marinho LB, Almeida JS, Filho PPR (2018) Localization of mobile robots with topological maps and classification with reject option using convolutional neural networks in omnidirectional images. In: 2018 international joint conference on neural networks (IJCNN), pp 1–8. https://doi.org/10.1109/IJCNN.2018.8489328 Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: In Proceedings of the IEEE conference on computer vision and pattern recognition, San Diego, vol II, pp 886–893. https://doi.org/10.1109/CVPR.2005.177 Duguleana M, Mogan G (2016) Neural networks based reinforcement learning for mobile robots obstacle avoidance. Expert Syst Appl 62:104–115. https://doi.org/10.1016/j.eswa.2016.06.021 Dymczyk M, Gilitschenski I, Nieto J, Lynen S, Zeisl B, Siegwart R (2018) Landmarkboost: efficient visualcontext classifiers for robust localization. In: 2018 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp 677–684. https://doi.org/10.1109/IROS.2018.8594100 Faessler M, Fontana F, Forster C, Mueggler E, Pizzoli M, Scaramuzza D (2016) Autonomous, vision-based flight and live dense 3D mapping with a quadrotor micro aerial vehicle. J Field Robot 33(4):431–450. https://doi.org/10.1002/rob.21581 Fan F, Ma Q, Ge J, Peng Q, Riley WW, Tang S (2013) Prediction of texture characteristics from extrusion food surface images using a computer vision system and artificial neural networks. J Food Eng 118(4):426–433. https://doi.org/10.1016/j.jfoodeng.2013.04.015 Gonzalez R, Apostolopoulos D, Iagnemma K (2018) Slippage and immobilization detection for planetary exploration rovers via machine learning and proprioceptive sensing. J Field Robot 35(2):231–247. https://doi.org/10.1002/rob.21736 Guo J, Gould S (2015) Deep CNN ensemble with data augmentation for object detection. https://doi.org/10.48550/ARXIV.1506.07224. arXiv:1506.07224 Ho TK (1995) Random decision forests. In: Proceedings of 3rd international conference on document analysis and recognition, IEEE, vol 1, pp 278–282. https://doi.org/10.1109/ICDAR.1995.598994 Horst M, Möller R (2017) Visual place recognition for autonomous mobile robots. Robotics 6(2):9. https://doi.org/10.3390/robotics6020009 Iagnemma K, Ward CC (2009) Classification-based wheel slip detection and detector fusion for mobile robots on outdoor terrain. Auton Robot 26(1):33–46. https://doi.org/10.1007/s10514-008-9105-8 Kang I, Molinaro DD, Choi G, Camargo J, Young AJ (2022) Subject-independent continuous locomotion mode classification for robotic hip exoskeleton applications. IEEE Trans Biomed Eng 69(10):3234–3242. https://doi.org/10.1109/TBME.2022.3165547 Korrapati H, Mezouar Y (2017) Multi-resolution map building and loop closure with omnidirectional images. Auton Robot 41(4):967–987. https://doi.org/10.1007/s10514-016-9560-6 Leonardis A, Bischof H (2000) Robust recognition using eigenimages. Comput Vis Image Underst 78(1):99–118. https://doi.org/10.1006/cviu.1999.0830 Li S, Chou L, Chang T, Yang C, Chang Y (2019) Obstacle avoidance of mobile robot based on hyperomni vision. Sens Mater 31(3):1021–1036. https://doi.org/10.18494/SAM.2019.2226 Liu R, Zhang J, Yin K, Pan Z, Lin R, Chen S (2018) Absolute orientation and localization estimation from an omnidirectional image. In: pacific rim international conference on artificial intelligence, Springer, pp 309–316. https://doi.org/10.1007/978-3-319-97310-4_35 Lowe DG (1999) Object recognition from local scale-invariant features. In: The proceedings of the seventh IEEE international conference on Computer vision, 1999, vol 2, pp 1150–1157. https://doi.org/10.1109/ICCV.1999.790410 Luxburg U (2007) A tutorial on spectral clustering. Stat Comput 17:395–416. https://doi.org/10.1007/s11222-007-9033-z Mancini M, Bulò SR, Ricci E, Caputo B (2017) Learning deep NBNN representations for robust place categorization. IEEE Robot Autom Lett 2(3):1794–1801. https://doi.org/10.1109/LRA.2017.2705282 Marinho LB, Rebouças Filho PP, Almeida JS, Souza JWM, Souza Junior AH, de Albuquerque VHC (2018) A novel mobile robot localization approach based on classification with rejection option using computer vision. Comput Electr Eng 68:26–43. https://doi.org/10.1016/j.compeleceng.2018.03.047 Maron ME (1961) Automatic indexing: an experimental inquiry. J ACM 8(3):404–417. https://doi.org/10.1145/321075.321084 Meattini R, Benatti S, Scarcia U, De Gregorio D, Benini L, Melchiorri C (2018) An sEMG-based human-robot interface for robotic hands using machine learning and synergies. IEEE Trans Compon Packag Manuf Technol 8(7):1149–1158. https://doi.org/10.1109/TCPMT.2018.2799987 Murthy GRS, Jadon RS (2010) Hand gesture recognition using neural networks. In: 2010 IEEE 2nd international advance computing conference (IACC), pp 134–138. https://doi.org/10.1109/IADCC.2010.5423024 Oliva A, Torralba A (2001) Modeling the shape of the scene: a holistic representation of the spatial envelope. Int J Comput Vis 42(3):145–175. https://doi.org/10.1023/A:1011139631724 Oliva A, Torralba A (2006) Building the gist of a scene: the role of global image features in recognition. In: Visual perception, progress in brain research, vol 155, Elsevier, pp 23–36. https://doi.org/10.1016/S0079-6123(06)55002-2. https://www.sciencedirect.com/science/article/pii/S0079612306550022 Payá L, Peidró A, Amorós F, Valiente D, Reinoso O (2018) Modeling environments hierarchically with omnidirectional imaging and global-appearance descriptors. Remote Sens 10(4):522 Payá L, Reinoso O, Berenguer Y, Úbeda D (2016) Using omnidirectional vision to create a model of the environment: a comparative evaluation of global-appearance descriptors. J Sens 2016:1209507. https://doi.org/10.1155/2016/1209507 Posada LF, Narayanan KK, Hoffmann F, Bertram T (2010) Floor segmentation of omnidirectional images for mobile robot visual navigation. In: 2010 IEEE/RSJ international conference on intelligent robots and systems, IEEE, pp 804–809. IROS.2010.5652869 Pronobis A, Caputo B (2009) COLD: COsy localization database. Int J Robot Res (IJRR) 28(5):588–594. https://doi.org/10.1177/0278364909103912 Rahimi A, Recht B (2008) Random features for large-scale kernel machines. In: Advances in neural information processing systems, pp 1177–1184 Rebouças Filho PP, da Silva SPP, Ohata EF, Almeida JS, de Sousa PHF, Nascimento NMM, dos Santos Silva FH (2019) A new strategy for mobile robots localization based on omnidirectional sonar images and machine learning. In: Anais Estendidos da XXXII conference on graphics, patterns and images, SBC, pp 168–171. https://doi.org/10.5753/sibgrapi.est.2019.8321 Reich S, Seer M, Berscheid L, Wörgötter F, Braun JM (2018) Omnidirectional visual odometry for flying robots using low-power hardware. In: VISIGRAPP (5: VISAPP), pp 499–507. https://doi.org/10.5220/0006509704990507 Reinoso O, Payá L (2020) Special issue on visual sensors. Sensors. https://doi.org/10.3390/s20030910 Reinoso O, Payá L (2020) Special issue on mobile robots navigation. Appl Sci. https://doi.org/10.3390/app10041317 Rituerto A, Murillo AC, Guerrero J (2014) Semantic labeling for indoor topological mapping using a wearable catadioptric system. Robot Auton Syst 62(5):685–695. https://doi.org/10.1016/j.robot.2012.10.002 Rituerto A, Puig L, Guerrero J (2010) Visual slam with an omnidirectional camera. In: 2010 international conference on pattern recognition, IEEE, pp 348–351. https://doi.org/10.1109/ICPR.2010.94 Román V, Payá L, Peidró A, Ballesta M, Reinoso O (2021) The role of global appearance of omnidirectional images in relative distance and orientation retrieval. Sensors. https://doi.org/10.3390/s21103327 Rublee E, Rabaud V, Konolige K, Bradski G (2011) Orb: An efficient alternative to sift or surf. In: International conference on computer vision, IEEE, pp 2564–2571 Shi X, Shen Y, Wang Y, Bai L (2018) Differential-clustering compression algorithm for real-time aerospace telemetry data. IEEE Access 6:57425–57433. https://doi.org/10.1109/ACCESS.2018.2872778 Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data 6(1):60. https://doi.org/10.1109/ACCESS.2018.2872778 Triebel R, Grimmett H, Paul R, Posner I (2016) Driven learning for driving: How introspection improves semantic mapping. In: Robotics research, Springer, pp 449–465 Valgren C, Lilienthal A (2010) Sift, surf & seasons: appearance-based long-term localization in outdoor environments. Robot Auton Syst 58:149–156. https://doi.org/10.1016/j.robot.2009.09.010 van Gerven M, Bohte S (2017) Editorial: artificial neural networks as models of neural information processing. Front Comput Neurosci 11:114. https://doi.org/10.3389/fncom.2017.00114 Wang LL, Ngan HYT, Yung NHC (2018) Automatic incident classification for large-scale traffic data by adaptive boosting SVM. Inf Sci 467:59–73. https://doi.org/10.1016/j.ins.2018.07.044 Wozniak P, Kwolek B (2021) Deep embeddings-based place recognition robust to motion blur. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1771–1779 Zhang J, Li M, Feng Y, Yang C (2020) Robotic grasp detection based on image processing and random forest. Multimed Tools Appl 79(3):2427–2446. https://doi.org/10.1007/s11042-019-08302-9 Zhou B, Lapedriza A, Xiao J, Torralba A, Oliva A (2014) Learning deep features for scene recognition using places database. In: Advances in neural information processing systems, pp 487–495