Large-Scale Optical Hardware for Neural Network Inference Acceleration
Artificial deep neural networks (DNNs) have revolutionized tasks such as automated classification and natural language processing. To boost accuracy and handle more complex workloads, DNN model sizes have grown exponentially over the last decade, outpacing improvements in digital electronic micropro...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
Massachusetts Institute of Technology
2024
|
Online Access: | https://hdl.handle.net/1721.1/153830 |
_version_ | 1811073465684328448 |
---|---|
author | Bernstein, Liane |
author2 | Englund, Dirk R. |
author_facet | Englund, Dirk R. Bernstein, Liane |
author_sort | Bernstein, Liane |
collection | MIT |
description | Artificial deep neural networks (DNNs) have revolutionized tasks such as automated classification and natural language processing. To boost accuracy and handle more complex workloads, DNN model sizes have grown exponentially over the last decade, outpacing improvements in digital electronic microprocessor efficiency. This mismatch limits DNN performance and contributes to soaring data center energy costs. Optical hardware for deep learning (optical neural networks, or ONNs) can theoretically increase DNN processing efficiency; however, the feasibility of large-scale, fully programmable and reconfigurable ONNs has not yet been comprehensively shown in experiments.
This thesis reports our demonstrations of ONNs that classify ~1000-element input vectors using standard DNN layers in inference without hardware modeling or retraining. In a first project, we used digital optical links to replace copper wires for transmitting and copying data to electronic multipliers. Our experimental implementation showed an MNIST classification accuracy within <0.6% of the digital electronic ground truth. We estimated that this 'digital ONN' could reduce energy consumption for long data transfer lengths, but not in tightly packed electronic multiplier arrays. Therefore, in a second project, we expanded upon this work by performing reconfigurable optical multicast and analog optoelectronic weighting to compute DNN layer outputs in a single shot. Our proof-of-concept system yielded an MNIST classification accuracy of 96.7% (boosted to 97.3% with weight fine-tuning) with respect to the ground-truth accuracy of 97.9%. We calculated that a near-term optimized version of this system could lower energy consumption and latency by 1-2 orders of magnitude compared to a state-of-the-art digital electronic systolic array. These findings suggest a paradigm shift towards optoelectronic DNN accelerators with lower resource utilization that could enable the next generation of deep learning. |
first_indexed | 2024-09-23T09:33:33Z |
format | Thesis |
id | mit-1721.1/153830 |
institution | Massachusetts Institute of Technology |
last_indexed | 2024-09-23T09:33:33Z |
publishDate | 2024 |
publisher | Massachusetts Institute of Technology |
record_format | dspace |
spelling | mit-1721.1/1538302024-03-22T04:00:12Z Large-Scale Optical Hardware for Neural Network Inference Acceleration Bernstein, Liane Englund, Dirk R. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Artificial deep neural networks (DNNs) have revolutionized tasks such as automated classification and natural language processing. To boost accuracy and handle more complex workloads, DNN model sizes have grown exponentially over the last decade, outpacing improvements in digital electronic microprocessor efficiency. This mismatch limits DNN performance and contributes to soaring data center energy costs. Optical hardware for deep learning (optical neural networks, or ONNs) can theoretically increase DNN processing efficiency; however, the feasibility of large-scale, fully programmable and reconfigurable ONNs has not yet been comprehensively shown in experiments. This thesis reports our demonstrations of ONNs that classify ~1000-element input vectors using standard DNN layers in inference without hardware modeling or retraining. In a first project, we used digital optical links to replace copper wires for transmitting and copying data to electronic multipliers. Our experimental implementation showed an MNIST classification accuracy within <0.6% of the digital electronic ground truth. We estimated that this 'digital ONN' could reduce energy consumption for long data transfer lengths, but not in tightly packed electronic multiplier arrays. Therefore, in a second project, we expanded upon this work by performing reconfigurable optical multicast and analog optoelectronic weighting to compute DNN layer outputs in a single shot. Our proof-of-concept system yielded an MNIST classification accuracy of 96.7% (boosted to 97.3% with weight fine-tuning) with respect to the ground-truth accuracy of 97.9%. We calculated that a near-term optimized version of this system could lower energy consumption and latency by 1-2 orders of magnitude compared to a state-of-the-art digital electronic systolic array. These findings suggest a paradigm shift towards optoelectronic DNN accelerators with lower resource utilization that could enable the next generation of deep learning. Ph.D. 2024-03-21T19:08:44Z 2024-03-21T19:08:44Z 2024-02 2024-02-21T17:18:39.392Z Thesis https://hdl.handle.net/1721.1/153830 In Copyright - Educational Use Permitted Copyright retained by author(s) https://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology |
spellingShingle | Bernstein, Liane Large-Scale Optical Hardware for Neural Network Inference Acceleration |
title | Large-Scale Optical Hardware for Neural Network Inference Acceleration |
title_full | Large-Scale Optical Hardware for Neural Network Inference Acceleration |
title_fullStr | Large-Scale Optical Hardware for Neural Network Inference Acceleration |
title_full_unstemmed | Large-Scale Optical Hardware for Neural Network Inference Acceleration |
title_short | Large-Scale Optical Hardware for Neural Network Inference Acceleration |
title_sort | large scale optical hardware for neural network inference acceleration |
url | https://hdl.handle.net/1721.1/153830 |
work_keys_str_mv | AT bernsteinliane largescaleopticalhardwareforneuralnetworkinferenceacceleration |