Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks


如何引用文章

全文:

开放存取 开放存取
受限制的访问 ##reader.subscriptionAccessGranted##
受限制的访问 订阅存取

详细

To reduce random access memory (RAM) requirements and to increase speed of recognition algorithms we consider a weight discretization problem for trained neural networks. We show that an exponential discretization is preferable to a linear discretization since it allows one to achieve the same accuracy when the number of bits is 1 or 2 less. The quality of the neural network VGG-16 is already satisfactory (top5 accuracy 69%) in the case of 3 bit exponential discretization. The ResNet50 neural network shows top5 accuracy 84% at 4 bits. Other neural networks perform fairly well at 5 bits (top5 accuracies of Xception, Inception-v3, and MobileNet-v2 top5 were 87%, 90%, and 77%, respectively). At less number of bits, the accuracy decreases rapidly.

作者简介

M. Malsagov

Scientific Research Institute for System Analysis, Russian Academy of Sciences

编辑信件的主要联系方式.
Email: malsagov@niisi.ras.ru
俄罗斯联邦, Moscow, 117218

E. Khayrov

Scientific Research Institute for System Analysis, Russian Academy of Sciences

编辑信件的主要联系方式.
Email: emil.khayrov@gmail.com
俄罗斯联邦, Moscow, 117218

M. Pushkareva

Scientific Research Institute for System Analysis, Russian Academy of Sciences

编辑信件的主要联系方式.
Email: mariaratko@gmail.com
俄罗斯联邦, Moscow, 117218

I. Karandashev

Scientific Research Institute for System Analysis, Russian Academy of Sciences; Peoples Friendship University of Russia (RUDN University Moscow)

编辑信件的主要联系方式.
Email: karandashev@niisi.ras.ru
俄罗斯联邦, Moscow, 117218; Moscow, 117198

补充文件

附件文件
动作
1. JATS XML

版权所有 © Allerton Press, Inc., 2019