Archive Issue – Vol.4, Issue.1 (Jan-Mar 2024)

Archive Issue – Vol.4, Issue.1 (January-March 2024)


Augmented Reality: Enhancing Human-Computer Interaction Through Immersive Technology

Arun Pratap Singh

Research Paper | Journal Paper

Vol.4, Issue.1, pp.1-05, Mar-2024

Abstract

Augmented Reality (AR) is an emerging technology that overlays digital content onto the real-world environment, thereby enhancing human perception and interaction with their surroundings. Unlike Virtual Reality (VR), which immerses users in a completely virtual environment, AR seamlessly blends the physical and digital worlds. This paper explores the evolution, applications, and future potential of AR by presenting a comprehensive review of existing literature, identifying current challenges, and proposing solutions for improved AR implementations. The study highlights AR's application in healthcare, education, manufacturing, gaming, and retail while identifying major limitations such as hardware constraints, latency, user experience challenges, and data privacy issues. The proposed work emphasizes integrating advanced computer vision techniques, 5G connectivity, and artificial intelligence (AI) to enhance AR's effectiveness. Experimental outcomes suggest that AR significantly improves learning outcomes, surgical precision, and user engagement in retail environments. Finally, this paper outlines the potential scope of AR in creating immersive smart cities, training simulations, and personalized user experiences.

Key-Words / Index Term: Augmented Reality, Human-Computer Interaction, Mixed Reality, AR Applications, Real-Time Processing, Computer Vision, Mobile AR, Wearable Devices, Machine Learning, Virtual Interaction.

References

      1. R. T. Azuma, "A survey of augmented reality," Presence: Teleoperators & Virtual Environments, vol. 6, no. 4, pp. 355–385, 1997.
      2. A. B. Craig, Understanding Augmented Reality: Concepts and Applications. Morgan Kaufmann, 2013.
      3. F. Zhou, H. B. Duh, and M. Billinghurst, "Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR," in Proc. IEEE/ACM Int. Symp. Mixed and Augmented Reality, 2008, pp. 193–202.
      4. M. Billinghurst, A. Clark, and G. Lee, "A survey of augmented reality," Foundations and Trends in Human–Computer Interaction, vol. 8, no. 2–3, pp. 73–272, 2015.
      5. Microsoft, "Microsoft HoloLens: Mixed reality technology." [Online]. Available: https://www.microsoft.com/hololens
      6. Apple, "ARKit overview." [Online]. Available: https://developer.apple.com/augmented-reality/
      7. S. Thrun, W. Burgard, and D. Fox, Probabilistic Robotics. MIT Press, 2005.
      8. E. Zhu, A. Hadadgar, I. Masiello, and N. Zary, "Augmented reality in healthcare education: An integrative review," PeerJ, vol. 2, p. e469, 2014.
      9. M. B. Ibáñez and C. Delgado-Kloos, "Augmented reality for STEM learning: A systematic review," Computers & Education, vol. 123, pp. 109–123, 2018.
      10. A. Y. C. Nee, S. K. Ong, G. Chryssolouris, and D. Mourtzis, "Augmented reality applications in design and manufacturing," CIRP Annals, vol. 61, no. 2, pp. 657–679, 2012.
      11. W. Narzt, G. Pomberger, A. Ferscha, D. Kolb, R. Müller, and J. Wieghardt, "Augmented reality navigation systems," Universal Access in the Information Society, vol. 4, pp. 177–187, 2006.
      12. A. Javornik, "Augmented reality: Research agenda for studying the impact of its media characteristics on consumer behaviour," Journal of Retailing and Consumer Services, vol. 30, pp. 252–261, 2016.
      13. J. Paavilainen, H. Korhonen, K. Alha, J. Stenros, E. Koskinen, and F. Mayra, "The Pokémon GO experience: A location-based augmented reality mobile game goes mainstream," in Proc. CHI Conf. Human Factors in Computing Systems, 2017, pp. 2493–2498.
      14. C. Dünser, M. Billinghurst, J. Wen, and A. Lehtinen, "Exploring the use of handheld AR for outdoor navigation," Computers & Graphics, vol. 32, no. 2, pp. 210–220, 2008.
      15. K. Kim, M. Billinghurst, G. Bruder, H. B. Duh, and G. F. Welch, "Revisiting trends in augmented reality research: A review of the 2nd decade of ISMAR (2008–2017)," IEEE Trans. Visualization and Computer Graphics, vol. 24, no. 11, pp. 2947–2962, 2018.
      16. M. J. Piran, D. Y. Suh, and S. Cho, "A survey on wireless communication standards for augmented reality and virtual reality," IEEE Access, vol. 8, pp. 428–444, 2020.

Citation

Arun Pratap Singh, "Augmented Reality: Enhancing Human-Computer Interaction Through Immersive Technology" International Journal of Scientific Research in Technology & Management, Vol.4, Issue.1, pp.1-05, 2024.

A Deep Analysis of Moving Object Detection & Tracking for Various Machine Learning Approaches

Amit Saxena, Sitesh Kumar Sinha, Sanjeev Kumar Gupta

Research Paper | Journal Paper

Vol.4, Issue.1, pp.06-09, Mar-2024

Abstract

Object detection is a significant process for performing computer vision related task. It plays an important role in the field of visual object tracking. If it is talking about the real world then it is a challenging task to detect object with high precision rate because of the mazy in appearance. There are various deep learning approaches through which object can be detected precisely but detecting the moving object and tracking it in every frame is bit more challenging. Moving object detection and tracking have wide variety of application such as border surviellence, road activity detection, forest monitoring and many more. There are various deep learning networks such as CNN, R-CNN, SSD, YOLO and many more, these networks are very much capable to detect the object with large traning model. Object tracking is next process after object detection because it is required to target the intial point with target object that has been detected and need to track as per the motion. The intention of this paper is to deep analyze the machine learning approaches which are used to detect the moving object with tracking.

Key-Words / Index Term: Moving Object Detection, Object Tracking, CNN, R-CNN, SSD, YOLO, Pattern Recognition.

References

    1. E. J. Fernandez-Sanchez, J. Diaz, and E. Ros, "Background subtraction based on color and depth using active sensors," Sensors, vol. 13, pp. 8895–8915, 2013, doi: 10.3390/s130708895.
    2. J. Zhu, Y. Lao, and Y. F. Zheng, "Object tracking in structured environment for video surveillance applications," IEEE Transactions on Circuits and Systems for Video Technology, vol. 20, Feb. 2010.
    3. S. Mane and S. Mangale, "Moving object detection and tracking using convolutional neural networks," in Proc. 2nd Int. Conf. Intelligent Computing and Control Systems (ICICCS), Madurai, India, 2018, pp. 1809–1813, doi: 10.1109/ICCONS.2018.8662921.
    4. R. Girshick, J. Donahue, T. Darrell, and J. Malik, "Rich feature hierarchies for accurate object detection and semantic segmentation," in Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR), Columbus, OH, 2014, pp. 580–587.
    5. C. Li, G. Dobler, X. Feng, and Y. Wang, "TrackNet: Simultaneous object detection and tracking and its application in traffic video analysis," 2019.
    6. K. He, G. Gkioxari, P. Dollár, and R. Girshick, "Mask R-CNN," arXiv preprint arXiv:1703.06870, 2017.
    7. J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You only look once: Unified, real-time object detection," arXiv preprint arXiv:1506.02640, 2015.
    8. W. Liu et al., "SSD: Single shot multibox detector," in Computer Vision – ECCV 2016, LNCS, vol. 9905, B. Leibe, J. Matas, N. Sebe, and M. Welling, Eds. Cham: Springer, 2016, pp. 21–37, doi: 10.1007/978-3-319-46448-0_2.
    9. T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollár, "Focal loss for dense object detection," in Proc. IEEE Int. Conf. Computer Vision (ICCV), Venice, Italy, 2017, pp. 2999–3007, doi: 10.1109/ICCV.2017.324.
    10. G. Ning et al., "Spatially supervised recurrent convolutional neural networks for visual object tracking," in Proc. IEEE Int. Symp. Circuits and Systems (ISCAS), Baltimore, MD, USA, 2017, pp. 1–4, doi: 10.1109/ISCAS.2017.8050867.
    11. Q. Wang, L. Zhang, L. Bertinetto, W. Hu, and P. Torr, "Fast online object tracking and segmentation: A unifying approach," in Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR), 2019, pp. 1328–1338, doi: 10.1109/CVPR.2019.00142.

Citation

Amit Saxena, Sitesh Kumar Sinha, Sanjeev Kumar Gupta, "A Deep Analysis of Moving Object Detection & Tracking for Various Machine Learning Approaches" International Journal of Scientific Research in Technology & Management, Vol.4, Issue.1, pp.06-09, 2024.

A Review on Visual Object Tracking Methods & Impact for Benchmarks

Amit Saxena, Sitesh Kumar Sinha, Manish Manoria, Sanjeev Kumar Gupta

Research Paper | Journal Paper

Vol.4, Issue.1, pp.10-16, Mar-2024

Abstract

Visual Object Tracking is a significantexploration point in Computer vision pattern recognition. The aim of Visual Object Tracking is to automatically acquire the environment of the object in the ensuing video outlines. Visual tracking is now very useful for tracking various moving objects like football in a match and many more for efficiently tracking the target for better decision making. In Artificial Intelligence; visual tracking is more challengeable because of instability of object in the frames. Conventional techniques are not efficient to deal with this kind of challenges. For tracking the object more efficiently; it has been targeted to achieve it through machine learning techniques.This paper intends to review various implemented researches that accepted the challenges for visual tracking and compare to the ground-truth. There are so many researches have been done in this field but somewhere somehow missing the target in certain benchmarks. The most promising technology which has been used by various researchers is Convolutional Neural Network. A network should be trained with all occlusions and acquire better level of precision with minimal overflows.

Key-Words / Index Term: Visual Tracking, Computer Vision, CNN, ANN, OpenCV, Machine Learning,OTB.

References

    1. VOT Challenges, "Datasets," 2015. [Online]. Available: https://www.votchallenge.net/vot2015/dataset.html. [Accessed: Sept. 26, 2021].
    2. Wikipedia, "Video Tracking," Oct. 28, 2014. [Online]. Available: https://en.wikipedia.org/wiki/Video_tracking. [Accessed: Sept. 26, 2021].
    3. Linköpings Universitet, "Discriminative Scale Space Tracker (DSST)," Jun. 26, 2017. [Online]. Available: https://www.cvl.isy.liu.se/research/objrec/visualtracking/scalvistrack/index.html. [Accessed: Sept. 26, 2021].
    4. A. W. Smeulders, D. M. Chu, R. Cucchiara, S. Calderara, A. Dehghan, and M. Shah, "Visual tracking: An experimental survey," IEEE Trans. Pattern Anal. Mach. Intell., vol. 36, no. 7, pp. 1442–1468, 2014.
    5. C. Fu, Z. Huang, Y. Li, R. Duan, and P. Lu, "Boundary effect-aware visual tracking for UAV with online enhanced background learning and multi-frame consensus verification," in Proc. IROS, 2019, pp. 4415–4422.
    6. H. Li, S. Wu, S. Huang, K. Lam, and X. Xing, "Deep motion-appearance convolutions for robust visual tracking," IEEE Access, vol. 7, pp. 180451–180466, 2019, doi: 10.1109/ACCESS.2019.2958405.
    7. L. Zheng, M. Tang, and J. Wang, "Learning robust Gaussian process regression for visual tracking," in Proc. 27th Int. Joint Conf. Artificial Intelligence (IJCAI), 2018, pp. 1219–1225, doi: 10.24963/ijcai.2018/170.
    8. M. Danelljan, L. Van Gool, and R. Timofte, "Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)," 2020, pp. 7183–7192.
    9. K. Chen and W. Tao, "Convolutional regression for visual tracking," IEEE Trans. Image Process., vol. 27, no. 7, pp. 3611–3620, Jul. 2018, doi: 10.1109/TIP.2018.2819362.
    10. A. Milan, S. H. Rezatofighi, A. Dick, I. Reid, and K. Schindler, "Online multi-target tracking using recurrent neural networks," in Proc. AAAI Conf. Artificial Intelligence (AAAI'17), 2017, pp. 4225–4232.
    11. S. Yun, J. Choi, Y. Yoo, K. Yun, and J. Y. Choi, "Action-driven visual object tracking with deep reinforcement learning," IEEE Trans. Neural Netw. Learn. Syst., vol. 29, no. 6, pp. 2239–2252, Jun. 2018, doi: 10.1109/TNNLS.2018.2801826.
    12. D. Zhang, H. Maei, X. Wang, and Y.-F. Wang, "Deep reinforcement learning for visual object tracking in videos," 2017.
    13. J. F. Henriques, R. Caseiro, P. Martins, and J. Batista, "High-speed tracking with kernelized correlation filters," IEEE Trans. Pattern Anal. Mach. Intell., vol. 37, no. 3, pp. 583–596, Mar. 2015, doi: 10.1109/TPAMI.2014.2345390.
    14. D. S. Bolme, J. R. Beveridge, B. A. Draper, and Y. M. Lui, "Visual object tracking using adaptive correlation filters," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2010, pp. 2544–2550, doi: 10.1109/CVPR.2010.5539960.
    15. D. Dogra et al., "Video analysis of Hammersmith lateral tilting examination using Kalman filter guided multi-path tracking," Med. Biol. Eng. Comput., vol. 52, 2014, doi: 10.1007/s11517-014-1178-2.
    16. S. Yun, J. Choi, Y. Yoo, K. Yun, and J. Y. Choi, "Action-decision networks for visual tracking with deep reinforcement learning," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2017, pp. 1349–1358, doi: 10.1109/CVPR.2017.148.
    17. L. Zheng, M. Tang, and J. Wang, "Learning robust Gaussian process regression for visual tracking," in Proc. IJCAI'18, 2018, pp. 1219–1225.
    18. B. Li, W. Wu, Q. Wang, F. Zhang, J. Xing, and J. Yan, "SiamRPN++: Evolution of Siamese visual tracking with very deep networks," in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), 2019, pp. 4277–4286, doi: 10.1109/CVPR.2019.00441.
    19. X. Li et al., "Dual-regression model for visual tracking," Neural Netw., vol. 132, pp. 364–374, 2020, doi: 10.1016/j.neunet.2020.09.011.
    20. B. Zhang, X. Zhang, and J. Qi, "Support vector regression learning based uncalibrated visual servoing control for 3D motion tracking," in Proc. 34th Chinese Control Conf. (CCC), 2015, pp. 8208–8213, doi: 10.1109/ChiCC.2015.7260942.
    21. T. Wang and W. Zhang, "The visual-based robust model predictive control for two-DOF video tracking system," in Proc. Chinese Control and Decision Conf. (CCDC), 2016, pp. 3743–3747, doi: 10.1109/CCDC.2016.7531635.
    22. N. Djelal, N. Saadia, and A. Ramdane-Cherif, "Target tracking based on SURF and image based visual servoing," in Proc. 2nd Int. Conf. Commun., Comput. and Control Applications (CCCA), 2012, pp. 1–5, doi: 10.1109/CCCA.2012.6417913.
    23. C. H. Li and T. I. J. Tsay, "Robust visual tracking in cluttered environment using an active contour method," in Proc. SICE, 2018, pp. 53–58, doi: 10.23919/SICE.2018.8492705.
    24. Q. Guo, W. Feng, R. Gao, Y. Liu, and S. Wang, "Exploring the effects of blur and deblurring to visual object tracking," IEEE Trans. Image Process., vol. 30, pp. 1812–1824, 2021, doi: 10.1109/TIP.2020.3045630.

Citation

Amit Saxena, Sitesh Kumar Sinha, Manish Manoria, Sanjeev Kumar Gupta, "A Review on Visual Object Tracking Methods & Impact for Benchmarks" International Journal of Scientific Research in Technology & Management, Vol.4, Issue.1, pp.10-16, 2024.

A Review of Visual Trackers & its Applications over the Real World

Amit Saxena, Sitesh Kumar Sinha, Sanjeev Kumar Gupta

Research Paper | Journal Paper

Vol.4, Issue.1, pp.17-20, Mar-2024

Abstract

Visual tracking is one of the hot topics and is widely emerging research in the field of computer vision applications. Over the years, the application of visual tracking has been gradually expanded with the continuous development of technology. Researchers have developed various novel tracking methodologies to improve the performance. Although various approaches has been proposed robust visual tracking remains a great challenge.In this paper various visual tracking methods are surveyed and classified into major classifications and also future trends are identified.

Key-Words / Index Term: Moving Object Detection, Object Tracking, CNN, R-CNN, SSD, YOLO, Pattern Recognition.

References

    1. Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features, in Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, pp. I–511–I–518.
    2. Yao R, Shi Q, Shen C, Zhang Y, van den Hengel A (2013) Part-based visual tracking with online latent structural learning, in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2363–2370.
    3. Nam H, Han B (2016) Learning multi-domain convolutional neural networks for visual tracking, in The IEEE Conference on CVPR.
    4. Ning J, Yang J, Jiang S, Zhang L, Yang MH (2016) Object tracking via dual linear structured svm and explicit feature map, in The IEEE Conference on Computer Vision and Pattern Recognition –CVPR.
    5. Hare S, Saffari A, Torr P (2011) Struck: Structured output tracking with kernels, in ICCV, pp. 263–270.
    6. Liu B, Yang L, Huang J, Meer P, Gong L, Kulikowski C (2010) Robust and fast collaborative tracking with two stage sparse optimization, in IEEE ECCV, Lecture Notes in Computer Science, Springer Berlin Heidelberg, vol. 6314, pp. 624–637.
    7. Mei X, Ling H, Wu Y, Blasch E, Bai L (2011) Minimum error bounded efficient l1 tracker with occlusion detection. CVPR 2011:1257–1264.
    8. Montero AS, Lang J, Laganiere R (2015) Scalable kernel correlation filter with sparse feature integration. In IEEE International Conference on Computer Vision Workshop (ICCVW), pp. 587–594.
    9. Elman JL (1990) Finding structure in time. Cognitive Science 14(2):179–211. Link
    10. Tokola R, Bolme D (2015) Ensembles of correlation filters for object detection, in IEEE Conference on WACV, pp. 935–942.
    11. Guo X, Xiao N, Zhang L (2019) Sequential binary code selection for robust object tracking. Multimedia Tools and Applications, pp. 1–13.
    12. Unlu HU, Niehaus PS, Chirita D, Evangeliou N, Tzes A (2019) Deep learning-based visual tracking of UAVs using a PTZ camera system. In IECON 2019-45th Annual Conference of the IEEE Industrial Electronics Society 1:638–644.
    13. Wang Q, Gao J, Xing J, Zhang M, Hu W (2017) Dcfnet: Discriminant correlation filters network for visual tracking. arXiv:1704.04057
    14. Wang F, Liu G, Zhang H, Hao Z (2018) Robust long-term correlation tracking with multiple models. DOI Multimedia Tools and Applications.
    15. Zhong Wei LH, Ming-Hsuan Y (2012) Robust object tracking via sparsity-based collaborative model, in IEEE Conference on CVPR, pp. 1838–1845.
    16. Zhang J, Ma S, Sclaroff S (2014) MEEM: robust tracking via multiple experts using entropy minimization, in Proc. of the European Conference on Computer Vision (ECCV).
    17. Wang Y, Luo X, Lu D, Wu J, Shan F (2019) Robust visual tracking via a hybrid correlation filter. Multimed Tools Appl 78(22):31633–31648.
    18. Wang X, Zheng Z, He Y, Yan F, Zeng Z, Yang Y (2020) Progressive local filter pruning for image retrieval acceleration. arXiv:2001.08878
    19. Wu Y, Shen B, Ling H (2014) Visual tracking via online nonnegative matrix factorization. IEEE Transactions on Circuits and Systems for Video Technology 24(3):374–383.
    20. Wu Y, Lim J, Yang M-H (2015) Object tracking benchmark. IEEE Trans Pattern Anal Mach Intell 37(9):1834–1848.
    21. Xue X, Li Y (2019) Robust particle tracking via spatio-temporal context learning and multi-task joint local sparse representation. Multimed Tools Appl 78(15):21187–21204.
    22. Nam H, Baek M, Han B (2016) Modeling and propagating cnns in a tree structure for visual tracking. arXiv:1608.07242
    23. Unlu HU, Niehaus PS, Chirita D, Evangeliou N, Tzes A (2019) Deep learning-based visual tracking of UAVs using a PTZ camera system. In IECON 2019-45th Annual Conference of the IEEE Industrial Electronics Society 1:638–644.
    24. Zhang (2013) Attention-based Target Recognition Algorithm and Applications in the Mobile Robot, Chongqing University.
    25. Fisher RB (2004) The PETS04 surveillance ground-truth data sets, Proc. 6th IEEE International Workshop on Performance Evaluation of Tracking and Surveillance, pp. 1–5.
    26. Collins R, Zhou X, Teh SK (2005) An open source tracking testbed and evaluation web site, IEEE International Workshop on Performance Evaluation of Tracking and Surveillance (PETS 2005), pp. 35.
    27. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M (2015) Imagenet large scale visual recognition challenge. International Journal of Computer Vision 115:211–252.
    28. Tang F, et al. (2019) Robust visual tracking based on spatial context pyramid. Multimedia Tools and Applications 78(15):21065–21084.
    29. Unlu HU, et al. (2019) Deep learning-based visual tracking of UAVs using a PTZ camera system. IECON 2019-45th Annual Conference of the IEEE Industrial Electronics Society. Vol. 1. IEEE.
    30. Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features. Proceedings of CVPR 2001. Vol. 1. IEEE.
    31. Wang Q, et al. (2017) Dcfnet: Discriminant correlation filters network for visual tracking. arXiv:1704.04057
    32. Tang F, et al. (2019) Robust visual tracking based on spatial context pyramid. Multimedia Tools and Applications 78(15):21065–21084.
    33. Unlu HU, et al. (2019) Deep learning-based visual tracking of UAVs using a PTZ camera system. IECON 2019-45th Annual Conference of the IEEE Industrial Electronics Society. Vol. 1. IEEE.
    34. Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features. Proceedings of CVPR 2001. Vol. 1. IEEE.
    35. Wang Q, et al. (2017) Dcfnet: Discriminant correlation filters network for visual tracking. arXiv:1704.04057
    36. Babenko B, Yang MH, Belongie S (2009) Visual tracking with online multiple instance learning, in IEEE Conference on CVPR, pp. 983–990.
    37. Bao C, Wu Y, Ling H, Ji H (2012) Real time robust l1 tracker using accelerated proximal gradient approach, in CVPR, IEEE Conference, pp. 1830–1837.
    38. Bertinetto L, Valmadre J, Golodetz S, Miksik O, Torr PHS (2016) Staple: Complementary learners for realtime tracking, in IEEE Conference on CVPR, pp. 1401–1409.
    39. Blanco-Filgueira B, Garcia-Lesta D, Fernández-Sanjurjo M, Brea VM, López P (2019) Deep learning-based multiple object visual tracking on embedded system for IoT and mobile edge computing applications. IEEE Internet Things J 6(3):5423–5431.
    40. Bolme D, Beveridge J, Draper B, Lui YM (2010) Visual object tracking using adaptive correlation filters, in IEEE Conference on CVPR, pp. 2544–2550.
    41. Cannons K (2008) A review of visual tracking. Technical report CSE2008–07, York University, Canada.

Citation

Amit Saxena, Sitesh Kumar Sinha, Sanjeev Kumar Gupta "A Review of Visual Trackers & its Applications over the Real World" International Journal of Scientific Research in Technology & Management, Vol.4, Issue.1, pp.17-20, 2024.