Analysis on Deep Learning and its Types

Lakshmi Praba A.*  Preethi S. R. **
*-** Department of Electronics and Communication Engineering, SRM Valliammai Engineering College, Kattankulathur, Tamilnadu, India.

Abstract

This paper focuses on compiling necessary aspects of deep learning for future researches. Deep learning has gained massive popularity in computing research and application development. Many cognitive problems related to unstructured problems are solved using this technology. Deep learning and Artificial intelligence are the underlying paradigms on popular applications like Google Translator. Machine learning and Deep learning are the two subset of Artificial intelligence. Deep learning algorithms are integrated using different neural network layers replicating the functioning of the human brain. Deep learning uses layers of neural networks to learn data in a recursive manner from training data from structured datasets and uses this data to predict unstructured data. Deep learning has three layers, namely input layer, output layer and hidden layer. Neural networks are often usual for the image recognition. Big Data powered with Deep learning can drive innovations beyond imagination in future.

Keywords :

Introduction

Deep learning is a subgroup of machine learning (Kaur et al., 2020; Liu et al., 2020) that emulate the working of the human brain through neural networks capable to make decisions. Deep learning is from the broader family of artificial intelligence used to simulate human brain with a system of computer applications. It simulates the action of the neuron present in the human brain to process it. Machine learning (ML) and Deep learning (DL) are subsets of Artificial Intelligence (AI). Deep neural network. After training, the application works on its own to finish the task (Kaur et al., 2020). The relationship between Artificial intelligence, machine learning and deep learning is shown in Figure 1.

Figure 1. Comparison of Artificial Intelligence, Machine Learning and Deep Learning

1. Need for Deep Learning

Deep Learning has magnificent performance in computer vision and in language processing (Liu et al., 2020). Deep learning refers to networks that has various deep layers for efficient learning. Deep learning can be pictured as a set of points it decides according to the input. It is similar to the biological nervous system and works according to already known information. In neural networks, each node works as neurons in the human brain. Machine learning works by identifying the image extracts features such as height, shape and applies them in an algorithm to obtain the output. In deep learning, the raw images will be predicted and directly fed into the algorithms without manual feature extraction and obtain the output. Thus deep learning takes less time for execution.

2. Artificial Neural Networks

The biological nervous network consists of neurons. The neuron consists of a cell body, dendrites and an axon. Majorly dendrites are used for receiving signal from every neuron. In every neuron structure, the end of each axon will be connected to the dendrites (Lynn et al., 2019). The neuron generates an electrochemical pulse called an action potential. This potential travels along the axon, activates and reaches them. The neural networks have capabilities to solve the given task that machine learning can never be able to complete. A deep neural network refers to a large amount of hidden layers. The layer can learn the given raw data directly on its own (Rafi et al., 2021). Artificial neural networks are made up of connected nodes that are organised like a human brain. The numerical values of each neuron in a structure are as follows: 3.4, 7.8, 2.33, 0.49. In real life, the link between two artificial neural networks is referred to as an axon. Weights are used to represent the relationships. Every task will have a different set of numerical weights (Abiodun et al., 2019). For each task, the weights values cannot be predicted in advance. The goal of neural networks is to tackle any tasks that are similar to those performed by the human brain.

Different sorts of neural networks are Feed forward networks and Feedback networks. Feed forward networks consist of input, output and hidden layers. After calculations are performed with the input data, new values generated will be forward to the further layer to determine the output. Feedback networks signals works in both the direction in its loop.

Neural networks distinguish the depth of the deep learning and their networks. Compared to multilinear regression Artificial Neural Networks predicts more accurate output.

3. Types of Algorithms in Deep Learning

Deep learning algorithms are extensively used to solve problems. Each type has different neural networks to solve the complex problems. Ten types of algorithms are shown in Figure 2 .

Figure 2. Taxonomy of Deep Learning Algorithms

3.1 Convolution Neural Networks

Convolution neural networks are mainly used for image processing, satellite imaging, medical image analysis, video recognition, etc. It contains a multi-layer algorithm that collects features from the given data and uses them to recognise objects (Aly & Almotairi, 2020; dos Santos et al., 2018). The layers are convolution layer, the rectified linear unit, and the pooling layer. For the action of convolution, the layer has filters. The elements are subjected to operations known as rectified linear units. As a map, the output has been rectified. Down sampling, also known as the pooling layer, reduces the dimensionality of the featured map (Yin et al., 2021). It flattens two dimensional layers to create a single layer. The object is identified using the flattened matrix as input in a convolutional neural network sample.

3.2 Recurrent Neural Networks

Recurrent neural networks are time related data. It has a loop where the output is given back to the input called feedback. The basic neuron transforms into adapted complex structures for greater performance (Lynn et al., 2019; Roy et al., 2020).

In recurrent neural networks, the input data is passed into a cell that activates and persists, and the output is fed into the feedback loop (Liu et al., 2021). The neural networks then filter the data using their own set of weights.

3.3 Long Short Term Memory Networks

It is a type of recurrent neural network that is used to remember past data for a longer period of time. Its principal applications are speech recognition and music composition. Data is sent at random intervals (Liu et al., 2020, 2021). The first phase is removing all of the irrelevant data from the prior state. Second, the cell state values are updated, and finally, the output in terms of the new cell state is generated.

3.4 Generative Adversarial Networks

Generative adversarial networks are deep learning algorithms that generate new data which resembles the training data. It has two main components: Generator and Discriminator. The Generator is used to generate false data. Discriminator is used to learn from the false data and it can distinguish real and false data. It solves many problems in computer vision and image processing (Shao et al., 2020). Generative adversarial networks are used to increase the period time and are mainly used in the video games to upscale low resolution and higher resolution in image training. Finally, it sends the result to both generator and discriminator.

3.5 Radial Basic Function Networks (RBFN)

Radial basic function networks are a special type of algorithms action function that occurs with the help of radial functions. RBFN is used in speech recognition, pattern recognition and image processing (Guo et al., 2020). RBFN has three layers input layer, hidden layer and output layer used for classification and time prediction. RBFN measures the input value using the training data. The hidden layer has a Gaussian transfer function which gives the output inversely proportional to the center of the neuron.

3.6 Multilayer Perception (MLP)

It has activation functions in its multilayer perceptron consisting of fully connected input and output layer. The same number of input and output layers are used to build speech and image recognition. MLP feeds data to the input layer and compute between the input layer and the hidden layer. Large database has been constructed for testing (Mokbal et al., 2019). MLP uses activation functions to determine the nodes and functions in sigmoid and tanh.

3.7 Self Organising Maps (SOM)

High dimensional functions are used to understand with the help of SOM (Aly & Almotairi, 2020). Self-organizing maps initialize weights to the nodes from the training data. Each node has been examined and compared with the input vectors. The values of winning nodes are called the Best Matching unit.

3.8 Deep Belief Networks (DBN)

Deep belief networks are used for accuracy and detection (Wang et al., 2021). It has multiple layers with latent variables where the binary values are often called hidden units. It has both previous and subsequent layers used for image recognition and motion capture data. DBN finds values in each layer of the bottom-up pass.

Deep belief networks are built based on the model of Restricted Boltzmann machine (Rathore et al., 2019), mainly used to generate images, to recognize and to capture data.

3.9 Restricted Boltzmann Machine

The deep learning algorithm used for dimensional reduction learns from the sets of input. It has two layers called visible units and hidden units (Li et al., 2018). The visible unit is connected to all hidden units. It collects the set of inputs and encodes them in a forward pass. The algorithm passes the output to the hidden layer. The activation of the individual passes to the output of the visible layer. In the visible layer, RBM compares with the original value and analyzes the result.

3.10 Autoencoders

Autoencoders are the special type of neural network where input and output are homogenous. It depicts the input data to the output layer (Bellini et al., 2019). It is mainly used in image processing and popularity prediction. Autoencoders receive the input and transform it into different forms. It first encodes the image and reduces the size of the input. Finally, decode the image to reconstruct the image.

4. Scope of Deep Learning

4.1 Applications of Deep Learning

The applications of Deep learning are found in image detection, forecasting, computer vision, speech recognition, social networking, data visualization, detection accuracy, reducing complexity and Machine translation problems. The summary of some of the applications of Deep learning been tabulated in Table 1.

Table 1. Some of the Applications of Deep Learning

4.2 Imapcts of Deep Learning

Feature learning is a fundamental job that gives knowledge about the problem and accuracy. In deep learning, feature learning works by itself using algorithms. It scans the data, identifies the features and it promotes learning. In deep learning using neural networks, the works are completed in a short period of time and it corrects the mistakes by repeating the task more than a thousand times. Deep learning algorithms are trying to grasp highlevel features, and therefore, it eliminates domain adroit and core extraction. The neural networks can be used for different other applications and data types. Deep learning is more flexible for newer applications.

Deep learning requires a very large amount of time for better performance for any given task. The cost will be high to train the complex models. It is quite difficult for a less competent person to adopt because there is no specific theory for training, selecting the proper network topology, or strategies to complete the specified tasks.

Conclusion

Deep learning and its various kinds are addressed in this study. Ten types of deep learning algorithms are covered, along with their applications. In the near future, neural networks will not only be deep, but will also evolve gradually. The right design and development of a network will improve both efficiency and problem solutions. Deep neural networks are effective at dealing with unstructured data due to their vast number of features. Deep learning has become an indispensible component of technology.

References

[1]. Abiodun, O. I., Jantan, A., Omolara, A. E., Dada, K. V., Umar, A. M., Linus, O. U., ..., & Kiru, M. U. (2019). Comprehensive review of artificial neural network applications to pattern recognition. IEEE Access, 7, 158820-158846. https://doi.org/10.1109/ACCESS.2019.29 45545
[2]. Aly, S., & Almotairi, S. (2020). Deep convolutional selforganizing map network for robust handwritten digit recognition. IEEE Access, 8, 107035-107045. https://doi. org/10.1109/ACCESS.2020.3000829
[3]. Bellini, V., Di Noia, T., Di Sciascio, E., & Schiavone, A. (2019). Semantics-aware autoencoder. IEEE Access, 7, 166122-166137. https://doi.org/10.1109/ACCESS.2019.29 53308
[4]. dos Santos, F. F., Pimenta, P. F., Lunardi, C., Draghetti, L., Carro, L., Kaeli, D., & Rech, P. (2018). Analyzing and increasing the reliability of convolutional neural networks on GPUs. IEEE Transactions on Reliability, 68(2), 663-677. https://doi.org/10.1109/TR.2018.2878387
[5]. Guo, R., Li, Y., Zhao, L., Zhao, J., & Gao, D. (2020). Remaining useful life prediction based on the Bayesian regularized radial basis function neural network for an external gear pump. IEEE Access, 8, 107498-107509. https://doi.org/10.1109/ACCESS.2020.3001013
[6]. Kaur, S., Singla, J., Nkenyereye, L., Jha, S., Prashar, D., Joshi, G. P., ..., & Islam, S. R. (2020). Medical diagnostic systems using Artificial Intelligence (AI) algorithms: Principles and perspectives. IEEE Access, 8, 228049- 228069. https://doi.org/10.1109/ACCESS.2020.3042273
[7]. Li, T., Wang, B., Jiang, Y., Zhang, Y., & Yan, Y. (2018). Restricted Boltzmann machine-based approaches for link prediction in dynamic networks. IEEE Access, 6, 29940- 29951. https://doi.org/10.1109/ACCESS.2018.2840054
[8]. Liu, L., Wang, P., Lin, J., & Liu, L. (2020). Intrusion detection of imbalanced network traffic based on machine learning and deep learning. IEEE Access, 9, 7550-7563. https://doi.org/10.1109/ACCESS.2020.3048198
[9]. Liu, S., Xia, Y., Shi, Z., Yu, H., Li, Z., & Lin, J. (2021). Deep learning in sheet metal bending with a novel theoryguided deep neural network. IEEE/CAA Journal of Automatica Sinica, 8(3), 565-581. https://doi.org/10.1109/ JAS.2021.1003871
[10]. Lynn, H. M., Pan, S. B., & Kim, P. (2019). A deep bidirectional GRU network model for biometric electrocardiogram classification based on recurrent neural networks. IEEE Access, 7, 145395-145405. https:// doi.org/10.1109/ACCESS.2019.2939947
[11]. Mokbal, F. M. M., Dan, W., Imran, A., Jiuchuan, L., Akhtar, F., & Xiaoxi, W. (2019). MLPXSS: An integrated XSSbased attack detection scheme in web applications using multilayer perceptron technique. IEEE Access, 7, 100567- 100580. https://doi.org/10.1109/ACCESS.2019.2927417
[12]. Rafi, S. H., Deeba, S. R., & Hossain, E. (2021). A shortterm load forecasting method using integrated CNN and LSTM network. IEEE Access, 9, 32436-32448. https://doi.org/ 10.1109/ACCESS.2021.3060654
[13]. Rathore, H., Al-Ali, A. K., Mohamed, A., Du, X., & Guizani, M. (2019). A novel deep learning strategy for classifying different attack patterns for deep brain implants. IEEE Access, 7, 24154-24164. https://doi.org/ 10.1109/ACCESS.2019.2899558
[14]. Roy, S., Menapace, W., Oei, S., Luijten, B., Fini, E., Saltori, C., ..., & Demi, L. (2020). Deep learning for classification and localization of COVID-19 markers in point-of-care lung ultrasound. IEEE Transactions on Medical Imaging, 39(8), 2676-2687. https://doi.org/10.11 09/TMI.2020.2994459
[15]. Shao, G., Huang, M., Gao, F., Liu, T., & Li, L. (2020). DuCaGAN: Unified dual capsule generative adversarial network for unsupervised image-to-image translation. IEEE Access, 8, 154691-154707. https://doi.org/10.1109/ ACCESS.2020.3007266
[16]. Wang, Z., Zeng, Y., Liu, Y., & Li, D. (2021). Deep belief network integrating improved kernel-based extreme learning machine for network intrusion detection. IEEE Access, 9, 16062-16091. https://doi.org/10.1109/ACCESS. 2021.3051074
[17]. Yin, L., Zhang, C., Wang, Y., Gao, F., Yu, J., & Cheng, L. (2021). Emotional deep learning programming controller for automatic voltage control of power systems. IEEE Access, 9, 31880-31891. https://doi.org/10.1109/ACCESS. 2021.3060620