Large-Scale Optical Hardware for Neural Network Inference Acceleration

Artificial deep neural networks (DNNs) have revolutionized tasks such as automated classification and natural language processing. To boost accuracy and handle more complex workloads, DNN model sizes have grown exponentially over the last decade, outpacing improvements in digital electronic micropro...

Full description

Bibliographic Details
Main Author: Bernstein, Liane
Other Authors: Englund, Dirk R.
Format: Thesis
Published: Massachusetts Institute of Technology 2024
Online Access:https://hdl.handle.net/1721.1/153830
_version_ 1811073465684328448
author Bernstein, Liane
author2 Englund, Dirk R.
author_facet Englund, Dirk R.
Bernstein, Liane
author_sort Bernstein, Liane
collection MIT
description Artificial deep neural networks (DNNs) have revolutionized tasks such as automated classification and natural language processing. To boost accuracy and handle more complex workloads, DNN model sizes have grown exponentially over the last decade, outpacing improvements in digital electronic microprocessor efficiency. This mismatch limits DNN performance and contributes to soaring data center energy costs. Optical hardware for deep learning (optical neural networks, or ONNs) can theoretically increase DNN processing efficiency; however, the feasibility of large-scale, fully programmable and reconfigurable ONNs has not yet been comprehensively shown in experiments. This thesis reports our demonstrations of ONNs that classify ~1000-element input vectors using standard DNN layers in inference without hardware modeling or retraining. In a first project, we used digital optical links to replace copper wires for transmitting and copying data to electronic multipliers. Our experimental implementation showed an MNIST classification accuracy within <0.6% of the digital electronic ground truth. We estimated that this 'digital ONN' could reduce energy consumption for long data transfer lengths, but not in tightly packed electronic multiplier arrays. Therefore, in a second project, we expanded upon this work by performing reconfigurable optical multicast and analog optoelectronic weighting to compute DNN layer outputs in a single shot. Our proof-of-concept system yielded an MNIST classification accuracy of 96.7% (boosted to 97.3% with weight fine-tuning) with respect to the ground-truth accuracy of 97.9%. We calculated that a near-term optimized version of this system could lower energy consumption and latency by 1-2 orders of magnitude compared to a state-of-the-art digital electronic systolic array. These findings suggest a paradigm shift towards optoelectronic DNN accelerators with lower resource utilization that could enable the next generation of deep learning.
first_indexed 2024-09-23T09:33:33Z
format Thesis
id mit-1721.1/153830
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T09:33:33Z
publishDate 2024
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1538302024-03-22T04:00:12Z Large-Scale Optical Hardware for Neural Network Inference Acceleration Bernstein, Liane Englund, Dirk R. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Artificial deep neural networks (DNNs) have revolutionized tasks such as automated classification and natural language processing. To boost accuracy and handle more complex workloads, DNN model sizes have grown exponentially over the last decade, outpacing improvements in digital electronic microprocessor efficiency. This mismatch limits DNN performance and contributes to soaring data center energy costs. Optical hardware for deep learning (optical neural networks, or ONNs) can theoretically increase DNN processing efficiency; however, the feasibility of large-scale, fully programmable and reconfigurable ONNs has not yet been comprehensively shown in experiments. This thesis reports our demonstrations of ONNs that classify ~1000-element input vectors using standard DNN layers in inference without hardware modeling or retraining. In a first project, we used digital optical links to replace copper wires for transmitting and copying data to electronic multipliers. Our experimental implementation showed an MNIST classification accuracy within <0.6% of the digital electronic ground truth. We estimated that this 'digital ONN' could reduce energy consumption for long data transfer lengths, but not in tightly packed electronic multiplier arrays. Therefore, in a second project, we expanded upon this work by performing reconfigurable optical multicast and analog optoelectronic weighting to compute DNN layer outputs in a single shot. Our proof-of-concept system yielded an MNIST classification accuracy of 96.7% (boosted to 97.3% with weight fine-tuning) with respect to the ground-truth accuracy of 97.9%. We calculated that a near-term optimized version of this system could lower energy consumption and latency by 1-2 orders of magnitude compared to a state-of-the-art digital electronic systolic array. These findings suggest a paradigm shift towards optoelectronic DNN accelerators with lower resource utilization that could enable the next generation of deep learning. Ph.D. 2024-03-21T19:08:44Z 2024-03-21T19:08:44Z 2024-02 2024-02-21T17:18:39.392Z Thesis https://hdl.handle.net/1721.1/153830 In Copyright - Educational Use Permitted Copyright retained by author(s) https://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Bernstein, Liane
Large-Scale Optical Hardware for Neural Network Inference Acceleration
title Large-Scale Optical Hardware for Neural Network Inference Acceleration
title_full Large-Scale Optical Hardware for Neural Network Inference Acceleration
title_fullStr Large-Scale Optical Hardware for Neural Network Inference Acceleration
title_full_unstemmed Large-Scale Optical Hardware for Neural Network Inference Acceleration
title_short Large-Scale Optical Hardware for Neural Network Inference Acceleration
title_sort large scale optical hardware for neural network inference acceleration
url https://hdl.handle.net/1721.1/153830
work_keys_str_mv AT bernsteinliane largescaleopticalhardwareforneuralnetworkinferenceacceleration