Nonlinear tensor train format for deep neural network compression

التفاصيل البيبلوغرافية
العنوان: Nonlinear tensor train format for deep neural network compression
المؤلفون: Dingheng Wang, Guoqi Li, Lei Deng, Hengnu Chen, Guangshe Zhao, Zhexian Liu
المصدر: Neural networks : the official journal of the International Neural Network Society. 144
سنة النشر: 2020
مصطلحات موضوعية: Artificial neural network, Computer science, business.industry, Cognitive Neuroscience, Deep learning, Data structure, Data Compression, Convolutional neural network, Physical Phenomena, Nonlinear system, Artificial Intelligence, Compression ratio, Artificial intelligence, Tensor, Neural Networks, Computer, business, Algorithm, MNIST database, Algorithms
الوصف: Deep neural network (DNN) compression has become a hot topic in the research of deep learning since the scale of modern DNNs turns into too huge to implement on practical resource constrained platforms such as embedded devices. Among variant compression methods, tensor decomposition appears to be a relatively simple and efficient strategy owing to its solid mathematical foundations and regular data structure. Generally, tensorizing neural weights into higher-order tensors for better decomposition, and directly mapping efficient tensor structure to neural architecture with nonlinear activation functions, are the two most common ways. However, the considerable accuracy loss is still a fly in the ointment for the tensorizing way especially for convolutional neural networks (CNNs), while the number of studies in the mapping way is comparatively limited and corresponding compression ratio appears to be not considerable. Therefore, in this work, by researching multiple types of tensor decompositions, we realize that tensor train (TT), which has specific and efficient sequenced contractions, is potential to take into account both of tensorizing and mapping ways. Then we propose a novel nonlinear tensor train (NTT) format, which contains extra nonlinear activation functions embedded in sequenced contractions and convolutions on the top of the normal TT decomposition and the proposed TT format connected by convolutions, to compensate the accuracy loss that normal TT cannot give. Further than just shrinking the space complexity of original weight matrices and convolutional kernels, we prove that NTT can afford an efficient inference time as well. Extensive experiments and discussions demonstrate that the compressed DNNs in our NTT format can almost maintain the accuracy at least on MNIST, UCF11 and CIFAR-10 datasets, and the accuracy loss caused by normal TT could be compensated significantly on large-scale datasets such as ImageNet.
تدمد: 1879-2782
URL الوصول: https://explore.openaire.eu/search/publication?articleId=doi_dedup___::6f106aa42741f342c884d217bd17bdfd
https://pubmed.ncbi.nlm.nih.gov/34547670
حقوق: CLOSED
رقم الأكسشن: edsair.doi.dedup.....6f106aa42741f342c884d217bd17bdfd
قاعدة البيانات: OpenAIRE