Deep Learning Techniques
Mr. G. G. Patil, Assistant Professor, Department of Computer Science and Engineering, SVERI’s College of Engineering, Pandharpur
The term Deep Learning was introduced to the machine learning
community by Rina Dechter in 1986, and to artificial neural networks by
Igor Aizenberg and colleagues in 2000, in the context of Boolean
threshold neurons. The first general, working learning algorithm for
supervised, deep, feed forward, multilayer perceptrons was published by
Alexey Ivakhnenko and Lapa in 1967. A 1971 paper described already a
deep network with 8 layers trained by the group method of data handling
algorithm.
Deep learning (also known as deep structured learning) is
part of a broader family of machine learning methods based on artificial
neural networks with representation learning. Learning can be
supervised, semi-supervised or unsupervised. Deep learning is a class of
machine learning algorithms that uses multiple layers to progressively
extract higher level features from the raw input. For example, in image
processing, lower layers may identify edges, while higher layers may
identify the concepts relevant to a human such as digits or letters or
faces.
The term Deep Learning was introduced to the machine learning
community by Rina Dechter in 1986, and to artificial neural networks by
Igor Aizenberg and colleagues in 2000, in the context of Boolean
threshold neurons. The first general, working learning algorithm for
supervised, deep, feed forward, multilayer perceptrons was published by
Alexey Ivakhnenko and Lapa in 1967. A 1971 paper described already a
deep network with 8 layers trained by the group method of data handling
algorithm.
Deep learning (also known as deep structured learning) is
part of a broader family of machine learning methods based on artificial
neural networks with representation learning. Learning can be
supervised, semi-supervised or unsupervised. Deep learning is a class of
machine learning algorithms that uses multiple layers to progressively
extract higher level features from the raw input. For example, in image
processing, lower layers may identify edges, while higher layers may
identify the concepts relevant to a human such as digits or letters or
faces.
Figure: Machine Learning Vs Deep Learning
Deep learning architectures such as deep neural networks, deep belief
networks, recurrent neural networks and convolutional neural networks
have been applied to fields including computer vision, speech
recognition, natural language processing, audio recognition, social
network filtering, machine translation, bioinformatics, drug design,
medical image analysis, material inspection and board game programs,
where they have produced results comparable to and in some cases
surpassing human expert performance.
Most modern deep learning models
are based on artificial neural networks, specifically, Convolutional
Neural Networks (CNN)s, although they can also include propositional
formulas or latent variables organized layer-wise in deep generative
models such as the nodes in deep belief networks and deep Boltzmann
machines.
In deep learning, each level learns to transform its input
data into a slightly more abstract and composite representation. In an
image recognition application, the raw input may be a matrix of pixels;
the first representational layer may abstract the pixels and encode
edges; the second layer may compose and encode arrangements of edges;
the third layer may encode a nose and eyes; and the fourth layer may
recognize that the image contains a face. Importantly, a deep learning
process can learn which features to optimally place in which level on
its own.
Applications:
Applications of deep learning are as follows:
- Colorization of Black and White Images
- Adding Sounds to Silent Movies
- Automatic Machine Translation
- Object Classification in Photographs
- Automatic Handwriting Generation
- Character Text Generation
- Image Caption Generation
- Automatic Game Playing
Challenges:
Five major challenges of deep learning are as follows: - Deep Learning Needs Enough Quality Data
- AI and Expectations
- Becoming Production-Ready
- Deep Learning Doesn’t Understand Context Very Well
- Deep Learning Security
References:
[1] Rina Dechter (1986). Learning while searching in constraint-satisfaction problems. University of California, Computer Science Department, Cognitive Systems Laboratory.
[2] Igor Aizenberg, Naum N. Aizenberg, Joos P.L. Vandewalle (2000). Multi-Valued and Universal Binary Neurons: Theory, Learning and Applications. Springer Science & Business Media.
[3] Co-evolving recurrent neurons learn deep memory POMDPs. Proc. GECCO, Washington, D. C., pp. 1795-1802, ACM Press, New York, NY, USA, 2005.
[4] Bengio, Y.; Courville, A.; Vincent, P. (2013). “Representation Learning: A Review and New Perspectives”. IEEE Transactions on Pattern Analysis and Machine Intelligence. 35 (8): 1798–1828. arXiv:1206.5538. doi:10.1109/tpami.2013.50.
[5] LeCun, Yann; Bengio, Yoshua; Hinton, Geoffrey (28 May 2015). “Deep learning”. Nature. 521 (7553): 436–444. Bibcode:2015Natur.521..436L. doi:10.1038/nature14539. PMID 26017442.
[6] Schulz, Hannes; Behnke, Sven (2012-11-01). “Deep Learning”. KI – Künstliche Intelligenz. 26 (4): 357–363. doi:10.1007/s13218-012-0198-z. ISSN 1610-1987
.
[7] Bengio, Yoshua; Lee, Dong-Hyun; Bornschein, Jorg; Mesnard, Thomas; Lin, Zhouhan (2015-02-13). “Towards Biologically Plausible Deep Learning”. arXiv:1502.04156.
[8] Abdel-Hamid, O.; et al. (2014). “Convolutional Neural Networks for Speech Recognition”. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 22 (10): 1533–1545. doi:10.1109/taslp.2014.2339736.
Comments
Post a Comment