Subscription Review Article

Advancements in Image Processing Techniques for Computer Vision Applications

by 
   Jayshree Dwivedi,    Pragya Rajput,
Volume :  11 | Issue :  01 | Received :  January 25, 2024 | Accepted :  February 29, 2024 | Published :  April 3, 2024
DOI :  10.37591

[This article belongs to Journal of Image Processing & Pattern Recognition Progress(joipprp)]

Keywords

Image processing, CNN, GANs, machine learning, RNN

Abstract

The fast-developing field of computer vision is transforming how people perceive and comprehend pictures and movies. Autonomous systems, robotics, healthcare, and surveillance are just a few of the many applications that have been made possible by recent significant advances in image and video processing. An overview of current developments in computer vision approaches, algorithms, and techniques for image and video analysis is given in this abstract. In conclusion, the analysis of images and videos has advanced significantly in the field of computer vision. Recurrent neural networks (RNNs) and CNNs are two examples of deep learning approaches that have been used to increase accuracy, robustness, and efficiency in a variety of applications. A richer understanding has resulted from the integration of semantic analysis with temporal and spatial data.

Full Text

References

1. Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster R-CNN: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems (pp. 91–99).
2. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC. Ssd: Single shot multibox detector. InComputer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I 14 2016 (pp. 21–37). Springer International Publishing.
3. Redmon J, Divvala S, Girshick R, Farhadi A. You only look once: Unified, real-time object detection. InProceedings of the IEEE conference on computer vision and pattern recognition 2016 (pp. 779–788).
4. Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. InProceedings of the IEEE conference on computer vision and pattern recognition 2015 (pp. 3431–3440).
5. Ronneberger, O., Fischer, P., & Brox, T. (2015). U-Net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer- assisted intervention (pp. 234–241).
6. Chen, L. C., Papandreou, G., Kokkinos, I., Murphy, K., & Yuille, A. L. (2018). Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE transactions on pattern analysis and machine intelligence, 40(4), 834–848.
7. Turk, M., & Pentland, A. (1991). Eigenfaces for recognition. Journal of cognitive neuroscience, 3(1), 71–86.
8. Viola, P., & Jones, M. (2004). Robust real-time face detection. International journal of computer vision, 57(2), 137–154.
9. Taigman, Y., Yang, M., Ranzato, M., & Wolf, L. (2014). Deepface: Closing the gap to human-level performance in face verification. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1701-1708).
10. Bolme, D. S., Beveridge, J. R., Draper, B. A., & Lui, Y. M. (2010). Visual object tracking using adaptive correlation filters. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2544–2550).
11. Danelljan, M., Hager, G., Shahbaz Khan, F., & Felsberg, M. (2015). Learning spatiallyregularized correlation filters for visual tracking. In Proceedings of the IEEE international conference on computer vision (pp. 4310–4318).
12. Wang, N., Zhang, T., Li, J., & Yang, J. (2018). Multi-cue correlation filters for robust visual tracking. International journal of computer vision, 126(5), 484–501.
13. Dong, C., Loy, C. C., He, K., & Tang, X. (2014). Learning a deep convolutional network for image super-resolution. In European conference on computer vision (pp. 184–199).
14. Ledig, C., Theis, L., Huszár, F., Caballero, J., Cunningham, A., Acosta, A., … & Shi, W.(2017). Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4681–4690).
15. Lim, B., Son, S., Kim, H., Nah, S., & Lee, K. M. (2017). Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops (pp. 136–144).
16. Simonyan, K., & Zisserman, A. (2014). Two-stream convolutional networks for action recognition in videos. In Advances in neural information processing systems (pp. 568– 576).
17. Carreira, J., & Zisserman, A. (2017). Quo vadis, action recognition? A new model and the kinetics dataset. In proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4724–4733).
18. Feichtenhofer, C., Pinz, A., & Zisserman, A. (2017). Detect to track and track to detect. In Proceedings of the IEEE international conference on computer vision (pp. 3038– 3046).
19. Quattoni, A., & Torralba, A. (2009). Recognizing indoor scenes. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 413–420).
20. Gupta, S., Arbeláez, P., & Malik, J. (2010). Perceptual organization and recognition of indoor scenes from RGB-D images. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 564–571).