Summary
Recently, sparse training methods have started to be established as a de facto approach for training and inference efficiency in artificial neural networks. In this paper, the authors introduce three novel contributions specifically designed for sparse neural networks: a parallel training algorithm, an activation function, and a hidden neurons importance metric. These contributions aim to train truly sparse neural networks to harvest their full potential. The paper discusses the importance of sparse connectivity in neural networks and the challenges faced due to the dominance of dense matrix operations in current deep learning software and hardware. The authors propose solutions to improve the scalability and efficiency of neural networks by using sparse training methods and novel algorithms. Experimental results demonstrate the performance of the proposed methods on various datasets, showcasing the potential of sparse neural networks in achieving high accuracy with reduced computational resources.