Understanding the effects of data parallelism and sparsity on neural network training

We study two factors in neural network training: data parallelism and sparsity; here, data parallelism means processing training data in parallel using distributed systems (or equivalently increasing batch size), so that training can be accelerated; for sparsity, we refer to pruning parameters in a...

Full description

Bibliographic Details
Main Authors: Lee, N, Ajanthan, T, Torr, PHS, Jaggi, M
Format: Conference item
Language:English
Published: OpenReview 2020