Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks

On‐chip training of neural networks (NNs) is regarded as a promising training method for neuromorphic systems with analog synaptic devices. Herein, a novel on‐chip training method called direct gradient calculation (DGC) is proposed to substitute conventional backpropagation (BP). In this method, th...

Full description

Bibliographic Details
Main Authors: Hyungyo Kim, Joon Hwang, Dongseok Kwon, Jangsaeng Kim, Min-Kyu Park, Jiseong Im, Byung-Gook Park, Jong-Ho Lee
Format: Article
Language:English
Published: Wiley 2021-08-01
Series:Advanced Intelligent Systems
Subjects:
Online Access:https://doi.org/10.1002/aisy.202100064
_version_ 1818738042366590976
author Hyungyo Kim
Joon Hwang
Dongseok Kwon
Jangsaeng Kim
Min-Kyu Park
Jiseong Im
Byung-Gook Park
Jong-Ho Lee
author_facet Hyungyo Kim
Joon Hwang
Dongseok Kwon
Jangsaeng Kim
Min-Kyu Park
Jiseong Im
Byung-Gook Park
Jong-Ho Lee
author_sort Hyungyo Kim
collection DOAJ
description On‐chip training of neural networks (NNs) is regarded as a promising training method for neuromorphic systems with analog synaptic devices. Herein, a novel on‐chip training method called direct gradient calculation (DGC) is proposed to substitute conventional backpropagation (BP). In this method, the gradients of a cost function with respect to the weights are calculated directly by sequentially applying a small temporal change to each weight and then measuring the change in cost value. DGC achieves a similar accuracy to that of BP while performing a handwritten digit classification task, validating its training feasibility. In particular, DGC can be applied to analog hardware‐based convolutional NNs (CNNs), which is considered to be a challenging task, enabling appropriate on‐chip training. A hybrid method is also proposed that efficiently combines DGC and BP for training CNNs, and the method achieves a similar accuracy to that of BP and DGC while enhancing the training speed. Furthermore, networks utilizing DGC maintain a higher level of accuracy than those using BP in the presence of variations in hardware (such as synaptic device conductance and neuron circuit component variations) while requiring fewer circuit components.
first_indexed 2024-12-18T01:02:39Z
format Article
id doaj.art-2ad44c2088ed41dcb455687719aea898
institution Directory Open Access Journal
issn 2640-4567
language English
last_indexed 2024-12-18T01:02:39Z
publishDate 2021-08-01
publisher Wiley
record_format Article
series Advanced Intelligent Systems
spelling doaj.art-2ad44c2088ed41dcb455687719aea8982022-12-21T21:26:19ZengWileyAdvanced Intelligent Systems2640-45672021-08-0138n/an/a10.1002/aisy.202100064Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural NetworksHyungyo Kim0Joon Hwang1Dongseok Kwon2Jangsaeng Kim3Min-Kyu Park4Jiseong Im5Byung-Gook Park6Jong-Ho Lee7Department of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaDepartment of Electrical and Computer Engineering and Inter-University Semiconductor Research Center Seoul National University 1 Gwanak-ro, Gwanak-gu Seoul 08826 South KoreaOn‐chip training of neural networks (NNs) is regarded as a promising training method for neuromorphic systems with analog synaptic devices. Herein, a novel on‐chip training method called direct gradient calculation (DGC) is proposed to substitute conventional backpropagation (BP). In this method, the gradients of a cost function with respect to the weights are calculated directly by sequentially applying a small temporal change to each weight and then measuring the change in cost value. DGC achieves a similar accuracy to that of BP while performing a handwritten digit classification task, validating its training feasibility. In particular, DGC can be applied to analog hardware‐based convolutional NNs (CNNs), which is considered to be a challenging task, enabling appropriate on‐chip training. A hybrid method is also proposed that efficiently combines DGC and BP for training CNNs, and the method achieves a similar accuracy to that of BP and DGC while enhancing the training speed. Furthermore, networks utilizing DGC maintain a higher level of accuracy than those using BP in the presence of variations in hardware (such as synaptic device conductance and neuron circuit component variations) while requiring fewer circuit components.https://doi.org/10.1002/aisy.202100064analog synaptic devicesbackpropagationhardware-based neural networkson-chip training
spellingShingle Hyungyo Kim
Joon Hwang
Dongseok Kwon
Jangsaeng Kim
Min-Kyu Park
Jiseong Im
Byung-Gook Park
Jong-Ho Lee
Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks
Advanced Intelligent Systems
analog synaptic devices
backpropagation
hardware-based neural networks
on-chip training
title Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks
title_full Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks
title_fullStr Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks
title_full_unstemmed Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks
title_short Direct Gradient Calculation: Simple and Variation‐Tolerant On‐Chip Training Method for Neural Networks
title_sort direct gradient calculation simple and variation tolerant on chip training method for neural networks
topic analog synaptic devices
backpropagation
hardware-based neural networks
on-chip training
url https://doi.org/10.1002/aisy.202100064
work_keys_str_mv AT hyungyokim directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT joonhwang directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT dongseokkwon directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT jangsaengkim directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT minkyupark directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT jiseongim directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT byunggookpark directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks
AT jongholee directgradientcalculationsimpleandvariationtolerantonchiptrainingmethodforneuralnetworks