Journal of Computing and Natural Science


Development and Evaluation of a Computer Vision System for Robot Navigation and Object Recognition in Real-World Environments



Journal of Computing and Natural Science

Received On : 16 May 2023

Revised On : 30 July 2023

Accepted On : 06 October 2023

Published On : 05 April 2024

Volume 04, Issue 02

Pages : 064-074


Abstract


The article discusses the vision framework for computing that includes image recognition, classification, prioritization, and navigation control modules. In this framework, a user model is used to feed the robotic controllers, whose performance improves in dynamic virtual contexts. In contrast, the vision module uses a multi-level perceptual neural network capable of efficient image segmentation, object recognition, and color segmentation, using the control module Position-Based Vision Serving (PBVS) and actions such as Avoid Collision (), Go-Ahead (), and Follow( ). It controls the motion of the robot, so the system successfully tested and met the requirements of the Antimedia Robotics Pioneer I robot. In addition, it was consistent with real life. The results show the effectiveness of the system in providing effective guidance and avoiding obstacles. Furthermore, the study investigates the use of artificial neural networks for image recognition and classification. In addition, it requires the use of SpCoMapping to add language maps to useful information. In summary, studies have emphasized the potential of computer vision and neural networks to improve robotic communication and language learning.


Keywords


Navigation Control Module, Computer Vision System, Position-Based Visual Servoing, MLP Neural Networks, Feature Extraction Module


  1. A. L. De Assis Simões, A. Pinto, J. S. Baptista, S. D. Pinheiro, and D. Romero, “Designing human-robot collaboration (HRC) workspaces in industrial settings: A systematic literature review,” Journal of Manufacturing Systems, vol. 62, pp. 28–43, Jan. 2022, doi: 10.1016/j.jmsy.2021.11.007.
  2. J. Trevelyan, W. R. Hamel, and S. Kang, “Robotics in hazardous applications,” in Springer handbooks, 2016, pp. 1521–1548. doi: 10.1007/978-3-319-32552-1_58.
  3. M. Hägele, W. Schaaf, and E. Helms, “Robot Assistants at Manual Workplaces: Effective Co-operation and Safety Aspects,” Proceedings of the 33rd ISR (International Symposium on Robotics), Jan. 2002.
  4. H. Tian, T. Wang, Y. Liu, X. Qiao, and Y. Li, “Computer vision technology in agricultural automation —A review,” Information Processing in Agriculture, vol. 7, no. 1, pp. 1–19, Mar. 2020, doi: 10.1016/j.inpa.2019.09.006.
  5. A. Moniz and B.-J. Krings, “Robots Working with Humans or Humans Working with Robots? Searching for Social Dimensions in New Human-Robot Interaction in Industry,” Societies, vol. 6, no. 3, p. 23, Aug. 2016, doi: 10.3390/soc6030023.
  6. M. Quigley, “ROS: an open-source Robot Operating System,” International Conference on Robotics and Automation, Jan. 2009, [Online]. Available: http://ci.nii.ac.jp/naid/10030771702.
  7. R. M. Cyert and M. H. DeGroot, “Rational expectations and Bayesian analysis,” Journal of Political Economy, vol. 82, no. 3, pp. 521–536, May 1974, doi: 10.1086/260210.
  8. V. Vijaya, R. Valupadasu, B. R. Chunduri, C. Rekha, and B. Sreedevi, “FPGA implementation of RS232 to Universal serial bus converter,” 2011 IEEE Symposium on Computers & Informatics, Mar. 2011, doi: 10.1109/isci.2011.5958920.
  9. H. Utz, S. Sablatnög, S. Enderle, and G. K. Kraetzschmar, “Miro - middleware for mobile robot applications,” IEEE Transactions on Robotics and Automation, vol. 18, no. 4, pp. 493–497, Aug. 2002, doi: 10.1109/tra.2002.802930.
  10. R. Linker, O. Cohen, and A. Naor, “Determination of the number of green apples in RGB images recorded in orchards,” Computers and Electronics in Agriculture, vol. 81, pp. 45–57, Feb. 2012, doi: 10.1016/j.compag.2011.11.007.
  11. C. Fahn, C. Lee, and Y.-S. Yeh, “A real-time pedestrian legs detection and tracking system used for autonomous mobile robots,” 2017 International Conference on Applied System Innovation (ICASI), May 2017, doi: 10.1109/icasi.2017.7988208.
  12. A. Vijayan and S. Ashok, “Comparative study on the performance of neural networks for prediction in assisting visual servoing,” Journal of Intelligent and Fuzzy Systems, vol. 36, no. 1, pp. 675–688, Feb. 2019, doi: 10.3233/jifs-171475.
  13. B. Cheng, Z. Li, J. Jiao, and G. An, “MLP Neural Network-Based Precise Localization of robot assembly parts,” in Lecture Notes in Computer Science, 2023, pp. 608–618. doi: 10.1007/978-981-99-6480-2_50.
  14. Y. LeCun et al., “Backpropagation applied to handwritten Zip code recognition,” Neural Computation, vol. 1, no. 4, pp. 541–551, Dec. 1989, doi: 10.1162/neco.1989.1.4.541.
  15. M. Riedmiller and H. Braun, “A direct adaptive method for faster backpropagation learning: the RPROP algorithm,” IEEE International Conference on Neural Networks, Dec. 2002, doi: 10.1109/icnn.1993.298623.
  16. R. Vankdothu, M. A. Hameed, A. Ameen, and R. Unnisa, “Brain image identification and classification on Internet of Medical Things in healthcare system using support value based deep neural network,” Computers & Electrical Engineering, vol. 102, p. 108196, Sep. 2022, doi: 10.1016/j.compeleceng.2022.108196.
  17. K. J. Cios and I. Shin, “Image recognition neural network: IRNN,” Neurocomputing, vol. 7, no. 2, pp. 159–185, Mar. 1995, doi: 10.1016/0925-2312(93)e0062-i.
  18. M. G. Quiles and R. A. F. Romero, “A Computer Vision System based on Multi-Layer Perceptrons for Controlling Mobile Robots,” 18th International Congress of Mechanical Engineering, Jan. 2005, [Online]. Available: https://www.abcm.org.br/symposium-series/SSM_Vol2/Section_X_Computer_Vision/SSM2_X_05.pdf.
  19. D. F. Wolf and G. S. Sukhatme, “Semantic mapping using mobile robots,” IEEE Transactions on Robotics, vol. 24, no. 2, pp. 245–258, Apr. 2008, doi: 10.1109/tro.2008.917001.
  20. A. Klein, S. Falkner, S. Bartels, P. Hennig, and F. Hutter, “Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets,” Proceedings of Machine Learning Research, pp. 528–536, Apr. 2017.

Acknowledgements


The authors would like to thank to the reviewers for nice comments on the manuscript.


Funding


No funding was received to assist with the preparation of this manuscript.


Ethics declarations


Conflict of interest

The authors have no conflicts of interest to declare that are relevant to the content of this article.


Availability of data and materials


No data available for above study.


Author information


Contributions

All authors have equal contribution in the paper and all authors have read and agreed to the published version of the manuscript.


Corresponding author


Rights and permissions


This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article‟s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article‟s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit https://creativecommons.org/licenses/by-nc-nd/4.0/


Cite this article


Malene Helgo, “Development and Evaluation of a Computer Vision System for Robot Navigation and Object Recognition in Real-World Environments”, Journal of Computing and Natural Science, vol.4, no.2, pp. 064-074, April 2024. doi: 10.53759/181X/JCNS202404007.


Copyright


© 2024 Malene Helgo. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.