Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks


Cite item

Full Text

Open Access Open Access
Restricted Access Access granted
Restricted Access Subscription Access

Abstract

To reduce random access memory (RAM) requirements and to increase speed of recognition algorithms we consider a weight discretization problem for trained neural networks. We show that an exponential discretization is preferable to a linear discretization since it allows one to achieve the same accuracy when the number of bits is 1 or 2 less. The quality of the neural network VGG-16 is already satisfactory (top5 accuracy 69%) in the case of 3 bit exponential discretization. The ResNet50 neural network shows top5 accuracy 84% at 4 bits. Other neural networks perform fairly well at 5 bits (top5 accuracies of Xception, Inception-v3, and MobileNet-v2 top5 were 87%, 90%, and 77%, respectively). At less number of bits, the accuracy decreases rapidly.

About the authors

M. Yu. Malsagov

Scientific Research Institute for System Analysis, Russian Academy of Sciences

Author for correspondence.
Email: malsagov@niisi.ras.ru
Russian Federation, Moscow, 117218

E. M. Khayrov

Scientific Research Institute for System Analysis, Russian Academy of Sciences

Author for correspondence.
Email: emil.khayrov@gmail.com
Russian Federation, Moscow, 117218

M. M. Pushkareva

Scientific Research Institute for System Analysis, Russian Academy of Sciences

Author for correspondence.
Email: mariaratko@gmail.com
Russian Federation, Moscow, 117218

I. M. Karandashev

Scientific Research Institute for System Analysis, Russian Academy of Sciences; Peoples Friendship University of Russia (RUDN University Moscow)

Author for correspondence.
Email: karandashev@niisi.ras.ru
Russian Federation, Moscow, 117218; Moscow, 117198

Supplementary files

Supplementary Files
Action
1. JATS XML

Copyright (c) 2019 Allerton Press, Inc.