SemiMul: Floating-Point Free Implementations for Efficient and Accurate Neural Network Training
Document Type
Conference Proceeding
Publication Date
1-1-2022
Abstract
Multiply-accumulate operation (MAC) is a fundamental component of machine learning tasks, where multiplication (either integer or float multiplication) compared to addition is costly in terms of hardware implementation or power consumption. In this paper, we approximate floating-point multiplication by converting it to integer addition while preserving the test accuracy of shallow and deep neural networks. We mathematically show and prove that our proposed method can be utilized with any floating-point format (e.g., FP8, FP16, FP32, etc.). It is also highly compatible with conventional hardware architectures and can be employed in CPU, GPU, or ASIC accelerators for neural network tasks with minimum hardware cost. Moreover, the proposed method can be utilized in embedded processors without a floating-point unit to perform neural network tasks. We evaluated our method on various datasets such as MNIST, FashionMNIST, SVHN, Cifar-10, and Cifar-100, with both FP16 and FP32 arithmetics. The proposed method preserves the test accuracy and, in some cases, overcomes the overfitting problem and improves the test accuracy.
Identifier
85152214348 (Scopus)
ISBN
[9781665462839]
Publication Title
Proceedings 21st IEEE International Conference on Machine Learning and Applications Icmla 2022
External Full Text Location
https://doi.org/10.1109/ICMLA55696.2022.00139
First Page
837
Last Page
842
Recommended Citation
Nezhadi, Ali; Angizi, Shaahin; and Roohi, Arman, "SemiMul: Floating-Point Free Implementations for Efficient and Accurate Neural Network Training" (2022). Faculty Publications. 3472.
https://digitalcommons.njit.edu/fac_pubs/3472
