Performance Analysis of Deep Neural Networks Using Computer Vision
DOI:
https://doi.org/10.4108/eai.13-10-2021.171318Keywords:
Computer Vision, Object Detection, Deep Learning, Deep Neural Network, Image classification, Artificial intelligence, Machine learning, Convolutional neural network, AccuracyAbstract
INTRODUCTION: In recent years, deep learning techniques have been made to outperform the earlier state-of-the-art machine learning techniques in many areas, with one of the most notable cases being computer vision. Deep learning is also employed to train the neural networks with the images and to perform the various tasks such as classification and segmentation using several different models. The size and depth of current deep learning models have increased to solve
certain tasks as these models provide better accuracy. As pre-trained weights may be used for further training and prevent costly computing, transfer learning is therefore of vital importance. A brief account is given of their history, structure, benefits, and drawbacks, followed by a description of their applications in the different tasks of computer vision, such as object detection, face recognition etc.
OBJECTIVE: The purpose of this paper is to train a deep neural network to properly classify the images that it has never seen before, define techniques to enhance the efficiency of deep learning and deploy deep neural networks in various applications.
METHOD: The proposed approach represents that after the reading of images, 256x256 pixel image’s random parts are extracted and noise, distortion, flip, or rotation transforms are applied. Multiple convolution and pooling steps are applied by controlling the stride lengths.
RESULT: Data analysis and research findings showed that DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. The proposed work outperforms the previous techniques in predicting the dependent variables, learning rate, image count, image mean, performance analysis of loss rate and learning rate during training, performance Analysis of Loss with respect to Epoch for Training, Validation and Accuracy.
CONCLUSION: This research encompasses a large variety of computer applications, from image recognition and machine translation to enhanced learning. DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. Extensive research has been conducted using the various deep architectures such as AlexNet, InceptionNet, etc. To the best of authors’ knowledge, this is the first work that presents a quantitative analysis of the deep architectures mentioned above.
Downloads
Downloads
Published
How to Cite
Issue
Section
License
This is an open-access article distributed under the terms of the Creative Commons Attribution CC BY 3.0 license, which permits unlimited use, distribution, and reproduction in any medium so long as the original work is properly cited.