Deep learning: an overview and main paradigms
- Authors: Golovko V.A.1,2
-
Affiliations:
- Brest State Technical University
- National Research Nuclear University MEPhI (Moscow Engineering Physics Institute)
- Issue: Vol 26, No 1 (2017)
- Pages: 1-17
- Section: Article
- URL: https://journals.rcsi.science/1060-992X/article/view/194933
- DOI: https://doi.org/10.3103/S1060992X16040081
- ID: 194933
Cite item
Abstract
In the present paper, we examine and analyze main paradigms of learning of multilayer neural networks starting with a single layer perceptron and ending with deep neural networks, which are considered regarded as a breakthrough in the field of the intelligent data processing. The baselessness of some ideas about the capacity of multilayer neural networks is shown and transition to deep neural networks is justified. We discuss the principal learning models of deep neural networks based on the restricted Boltzmann machine (RBM), an autoassociative approach and a stochastic gradient method with a Rectified Linear Unit (ReLU) activation function of neural elements.
About the authors
V. A. Golovko
Brest State Technical University; National Research Nuclear University MEPhI (Moscow Engineering Physics Institute)
Author for correspondence.
Email: gva@bstu.by
Belarus, Brest; Moscow
Supplementary files
