Environment modeling and localization from datasets of omnidirectional scenes using machine learning techniques
Tóm tắt
This work presents a framework to create a visual model of the environment which can be used to estimate the position of a mobile robot by means of artificial intelligence techniques. The proposed framework retrieves the structure of the environment from a dataset composed of omnidirectional images captured along it. These images are described by means of global-appearance approaches. The information is arranged in two layers, with different levels of granularity. The first layer is obtained by means of classifiers and the second layer is composed of a set of data fitting neural networks. Subsequently, the model is used to estimate the position of the robot, in a hierarchical fashion, by comparing the image captured from the unknown position with the information in the model. Throughout this work, five classifiers are evaluated (Naïve Bayes, SVM, random forest, linear discriminant classifier and a classifier based on a shallow neural network) along with three different global-appearance descriptors (HOG, gist, and a descriptor calculated from an intermediate layer of a pre-trained CNN). The experiments have been tackled with some publicly available datasets of omnidirectional images captured indoors with the presence of dynamic changes. Several parameters are used to assess the efficiency of the proposal: the ability of the algorithm to estimate coarsely the position (hit ratio), the average error (cm) and the necessary computing time. The results prove the efficiency of the framework to model the environment and localize the robot from the knowledge extracted from a set of omnidirectional images with the proposed artificial intelligence techniques.
Tài liệu tham khảo
Amorós F, Payá L, Marín JM, Reinoso O (2018) Trajectory estimation and optimization through loop closure detection, using omnidirectional imaging and global-appearance descriptors. Expert Syst Appl 102:273–290. https://doi.org/10.1016/j.eswa.2018.02.042
Amorós F, Payá L, Mayol-Cuevas W, Jiménez LM, Reinoso O (2020) Holistic descriptors of omnidirectional color images and their performance in estimation of position and orientation. IEEE Access 8:81822–81848. https://doi.org/10.1109/access.2020.2990996
Ballesta M, Payá L, Cebollada S, Reinoso O, Murcia F (2021) A CNN regression approach to mobile robot localization using omnidirectional images. Appl Sci. https://doi.org/10.3390/app11167521
Barshan B, Ayrulu B, Utete SW (2000) Neural network-based target differentiation using sonar for robotics applications. IEEE Trans Robot Autom 16(4):435–442. https://doi.org/10.1109/70.864239
Bay H, Tuytelaars T, Gool LV (2006) Surf: speeded up robust features. In: European conference on computer vision, Springer, pp 404–417. https://doi.org/10.1007/11744023_32
Cebollada S, Payá L, Flores M, Peidró A, Reinoso O (2021) A state-of-the-art review on mobile robotics tasks using artificial intelligence and visual data. Expert Syst Appl 167:114195. https://doi.org/10.1016/j.eswa.2020.114195
Cebollada S, Payá L, Flores M, Román V, Peidró A, Reinoso O (2020) A deep learning tool to solve localization in mobile autonomous robotics. In: ICINCO 2020, 17th international conference on informatics in control, automation and robotics, INSTICC, pp 232–241 (Online streaming, 7–9 July 2020)
Cebollada S, Payá L, Mayol W, Reinoso O (2019) Evaluation of clustering methods in compression of topological models and visual place recognition using global appearance descriptors. Appl Sci 9(3):377. https://doi.org/10.3390/app9030377
Cebollada S, Payá L, Román V, Reinoso O (2019) Hierarchical localization in topological models under varying illumination using holistic visual descriptors. IEEE Access 7:49580–49595. https://doi.org/10.1109/ACCESS.2019.2910581
Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273–297. https://doi.org/10.1007/BF00994018
Cruz Ulloa C, Prieto Sánchez G, Barrientos A, Del Cerro J (2021) Autonomous thermal vision robotic system for victims recognition in search and rescue missions. Sensors 21(21):7346. https://doi.org/10.3390/s21217346
da Silva SPP, da Nbrega RVM, Medeiros AG, Marinho LB, Almeida JS, Filho PPR (2018) Localization of mobile robots with topological maps and classification with reject option using convolutional neural networks in omnidirectional images. In: 2018 international joint conference on neural networks (IJCNN), pp 1–8. https://doi.org/10.1109/IJCNN.2018.8489328
Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: In Proceedings of the IEEE conference on computer vision and pattern recognition, San Diego, vol II, pp 886–893. https://doi.org/10.1109/CVPR.2005.177
Duguleana M, Mogan G (2016) Neural networks based reinforcement learning for mobile robots obstacle avoidance. Expert Syst Appl 62:104–115. https://doi.org/10.1016/j.eswa.2016.06.021
Dymczyk M, Gilitschenski I, Nieto J, Lynen S, Zeisl B, Siegwart R (2018) Landmarkboost: efficient visualcontext classifiers for robust localization. In: 2018 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp 677–684. https://doi.org/10.1109/IROS.2018.8594100
Faessler M, Fontana F, Forster C, Mueggler E, Pizzoli M, Scaramuzza D (2016) Autonomous, vision-based flight and live dense 3D mapping with a quadrotor micro aerial vehicle. J Field Robot 33(4):431–450. https://doi.org/10.1002/rob.21581
Fan F, Ma Q, Ge J, Peng Q, Riley WW, Tang S (2013) Prediction of texture characteristics from extrusion food surface images using a computer vision system and artificial neural networks. J Food Eng 118(4):426–433. https://doi.org/10.1016/j.jfoodeng.2013.04.015
Gonzalez R, Apostolopoulos D, Iagnemma K (2018) Slippage and immobilization detection for planetary exploration rovers via machine learning and proprioceptive sensing. J Field Robot 35(2):231–247. https://doi.org/10.1002/rob.21736
Guo J, Gould S (2015) Deep CNN ensemble with data augmentation for object detection. https://doi.org/10.48550/ARXIV.1506.07224. arXiv:1506.07224
Ho TK (1995) Random decision forests. In: Proceedings of 3rd international conference on document analysis and recognition, IEEE, vol 1, pp 278–282. https://doi.org/10.1109/ICDAR.1995.598994
Horst M, Möller R (2017) Visual place recognition for autonomous mobile robots. Robotics 6(2):9. https://doi.org/10.3390/robotics6020009
Iagnemma K, Ward CC (2009) Classification-based wheel slip detection and detector fusion for mobile robots on outdoor terrain. Auton Robot 26(1):33–46. https://doi.org/10.1007/s10514-008-9105-8
Kang I, Molinaro DD, Choi G, Camargo J, Young AJ (2022) Subject-independent continuous locomotion mode classification for robotic hip exoskeleton applications. IEEE Trans Biomed Eng 69(10):3234–3242. https://doi.org/10.1109/TBME.2022.3165547
Korrapati H, Mezouar Y (2017) Multi-resolution map building and loop closure with omnidirectional images. Auton Robot 41(4):967–987. https://doi.org/10.1007/s10514-016-9560-6
Leonardis A, Bischof H (2000) Robust recognition using eigenimages. Comput Vis Image Underst 78(1):99–118. https://doi.org/10.1006/cviu.1999.0830
Li S, Chou L, Chang T, Yang C, Chang Y (2019) Obstacle avoidance of mobile robot based on hyperomni vision. Sens Mater 31(3):1021–1036. https://doi.org/10.18494/SAM.2019.2226
Liu R, Zhang J, Yin K, Pan Z, Lin R, Chen S (2018) Absolute orientation and localization estimation from an omnidirectional image. In: pacific rim international conference on artificial intelligence, Springer, pp 309–316. https://doi.org/10.1007/978-3-319-97310-4_35
Lowe DG (1999) Object recognition from local scale-invariant features. In: The proceedings of the seventh IEEE international conference on Computer vision, 1999, vol 2, pp 1150–1157. https://doi.org/10.1109/ICCV.1999.790410
Luxburg U (2007) A tutorial on spectral clustering. Stat Comput 17:395–416. https://doi.org/10.1007/s11222-007-9033-z
Mancini M, Bulò SR, Ricci E, Caputo B (2017) Learning deep NBNN representations for robust place categorization. IEEE Robot Autom Lett 2(3):1794–1801. https://doi.org/10.1109/LRA.2017.2705282
Marinho LB, Rebouças Filho PP, Almeida JS, Souza JWM, Souza Junior AH, de Albuquerque VHC (2018) A novel mobile robot localization approach based on classification with rejection option using computer vision. Comput Electr Eng 68:26–43. https://doi.org/10.1016/j.compeleceng.2018.03.047
Maron ME (1961) Automatic indexing: an experimental inquiry. J ACM 8(3):404–417. https://doi.org/10.1145/321075.321084
Meattini R, Benatti S, Scarcia U, De Gregorio D, Benini L, Melchiorri C (2018) An sEMG-based human-robot interface for robotic hands using machine learning and synergies. IEEE Trans Compon Packag Manuf Technol 8(7):1149–1158. https://doi.org/10.1109/TCPMT.2018.2799987
Murthy GRS, Jadon RS (2010) Hand gesture recognition using neural networks. In: 2010 IEEE 2nd international advance computing conference (IACC), pp 134–138. https://doi.org/10.1109/IADCC.2010.5423024
Oliva A, Torralba A (2001) Modeling the shape of the scene: a holistic representation of the spatial envelope. Int J Comput Vis 42(3):145–175. https://doi.org/10.1023/A:1011139631724
Oliva A, Torralba A (2006) Building the gist of a scene: the role of global image features in recognition. In: Visual perception, progress in brain research, vol 155, Elsevier, pp 23–36. https://doi.org/10.1016/S0079-6123(06)55002-2. https://www.sciencedirect.com/science/article/pii/S0079612306550022
Payá L, Peidró A, Amorós F, Valiente D, Reinoso O (2018) Modeling environments hierarchically with omnidirectional imaging and global-appearance descriptors. Remote Sens 10(4):522
Payá L, Reinoso O, Berenguer Y, Úbeda D (2016) Using omnidirectional vision to create a model of the environment: a comparative evaluation of global-appearance descriptors. J Sens 2016:1209507. https://doi.org/10.1155/2016/1209507
Posada LF, Narayanan KK, Hoffmann F, Bertram T (2010) Floor segmentation of omnidirectional images for mobile robot visual navigation. In: 2010 IEEE/RSJ international conference on intelligent robots and systems, IEEE, pp 804–809. IROS.2010.5652869
Pronobis A, Caputo B (2009) COLD: COsy localization database. Int J Robot Res (IJRR) 28(5):588–594. https://doi.org/10.1177/0278364909103912
Rahimi A, Recht B (2008) Random features for large-scale kernel machines. In: Advances in neural information processing systems, pp 1177–1184
Rebouças Filho PP, da Silva SPP, Ohata EF, Almeida JS, de Sousa PHF, Nascimento NMM, dos Santos Silva FH (2019) A new strategy for mobile robots localization based on omnidirectional sonar images and machine learning. In: Anais Estendidos da XXXII conference on graphics, patterns and images, SBC, pp 168–171. https://doi.org/10.5753/sibgrapi.est.2019.8321
Reich S, Seer M, Berscheid L, Wörgötter F, Braun JM (2018) Omnidirectional visual odometry for flying robots using low-power hardware. In: VISIGRAPP (5: VISAPP), pp 499–507. https://doi.org/10.5220/0006509704990507
Reinoso O, Payá L (2020) Special issue on visual sensors. Sensors. https://doi.org/10.3390/s20030910
Reinoso O, Payá L (2020) Special issue on mobile robots navigation. Appl Sci. https://doi.org/10.3390/app10041317
Rituerto A, Murillo AC, Guerrero J (2014) Semantic labeling for indoor topological mapping using a wearable catadioptric system. Robot Auton Syst 62(5):685–695. https://doi.org/10.1016/j.robot.2012.10.002
Rituerto A, Puig L, Guerrero J (2010) Visual slam with an omnidirectional camera. In: 2010 international conference on pattern recognition, IEEE, pp 348–351. https://doi.org/10.1109/ICPR.2010.94
Román V, Payá L, Peidró A, Ballesta M, Reinoso O (2021) The role of global appearance of omnidirectional images in relative distance and orientation retrieval. Sensors. https://doi.org/10.3390/s21103327
Rublee E, Rabaud V, Konolige K, Bradski G (2011) Orb: An efficient alternative to sift or surf. In: International conference on computer vision, IEEE, pp 2564–2571
Shi X, Shen Y, Wang Y, Bai L (2018) Differential-clustering compression algorithm for real-time aerospace telemetry data. IEEE Access 6:57425–57433. https://doi.org/10.1109/ACCESS.2018.2872778
Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data 6(1):60. https://doi.org/10.1109/ACCESS.2018.2872778
Triebel R, Grimmett H, Paul R, Posner I (2016) Driven learning for driving: How introspection improves semantic mapping. In: Robotics research, Springer, pp 449–465
Valgren C, Lilienthal A (2010) Sift, surf & seasons: appearance-based long-term localization in outdoor environments. Robot Auton Syst 58:149–156. https://doi.org/10.1016/j.robot.2009.09.010
van Gerven M, Bohte S (2017) Editorial: artificial neural networks as models of neural information processing. Front Comput Neurosci 11:114. https://doi.org/10.3389/fncom.2017.00114
Wang LL, Ngan HYT, Yung NHC (2018) Automatic incident classification for large-scale traffic data by adaptive boosting SVM. Inf Sci 467:59–73. https://doi.org/10.1016/j.ins.2018.07.044
Wozniak P, Kwolek B (2021) Deep embeddings-based place recognition robust to motion blur. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1771–1779
Zhang J, Li M, Feng Y, Yang C (2020) Robotic grasp detection based on image processing and random forest. Multimed Tools Appl 79(3):2427–2446. https://doi.org/10.1007/s11042-019-08302-9
Zhou B, Lapedriza A, Xiao J, Torralba A, Oliva A (2014) Learning deep features for scene recognition using places database. In: Advances in neural information processing systems, pp 487–495