Accumulated decoupled learning with gradient staleness mitigation for convolutional neural networks
Gradient staleness is a major side effect in decoupled learning when training convolutional neural networks asynchronously. Existing methods that ignore this effect might result in reduced generalization and even divergence. In this paper, we propose an accumulated decoupled learning (ADL), wh...
Main Authors: | Zhuang, Huiping, Weng, Zhenyu, Luo, Fulin, Toh, Kar-Ann, Li, Haizhou, Lin, Zhiping |
---|---|
Other Authors: | School of Electrical and Electronic Engineering |
Format: | Conference Paper |
Language: | English |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/174480 https://icml.cc/virtual/2021/index.html |
Similar Items
-
Spoofing speech detection using temporal convolutional neural network
by: Xiao, Xiong, et al.
Published: (2018) -
An energy-efficient convolution unit for depthwise separable convolutional neural networks
by: Chong, Yi Sheng, et al.
Published: (2021) -
Stability analysis of delayed neural networks via compound-parameter-based integral inequality
by: Xue, Wenlong, et al.
Published: (2024) -
Annual dilated convolution neural network for newbuilding ship prices forecasting
by: Gao, Ruobin, et al.
Published: (2022) -
Ultra-high-speed accelerator architecture for convolutional neural network based on processing-in-memory using resistive random access memory
by: Wang, Hongzhe, et al.
Published: (2023)