Fast Multistage Compilation of Machine Learning Computation Graphs

Machine learning applications are increasingly requiring fast and more computational power. Many applications like language models have become so large that they are run on distributed systems in parallel. However, getting into the details of optimally scheduling or even just running machine learnin...

Full description

Bibliographic Details
Main Author: Dighe, Kaustubh
Other Authors: Amarasinghe, Saman
Format: Thesis
Published: Massachusetts Institute of Technology 2024
Online Access:https://hdl.handle.net/1721.1/156774
https://orcid.org/0009-0006-9656-1946
_version_ 1826189035810324480
author Dighe, Kaustubh
author2 Amarasinghe, Saman
author_facet Amarasinghe, Saman
Dighe, Kaustubh
author_sort Dighe, Kaustubh
collection MIT
description Machine learning applications are increasingly requiring fast and more computational power. Many applications like language models have become so large that they are run on distributed systems in parallel. However, getting into the details of optimally scheduling or even just running machine learning models on distributed systems can be a distraction for researchers ideating models. Hence there has been development of abstractions to facilitate running machine learning models in parallel on distributed systems. We present a compiler for the StreamIt language- a language made for abstract signal processing and multicore programming. We use that abstraction as a way to distribute the computation of machine learning models programmed in PyTorch.
first_indexed 2024-09-23T08:08:50Z
format Thesis
id mit-1721.1/156774
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T08:08:50Z
publishDate 2024
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1567742024-09-17T04:07:02Z Fast Multistage Compilation of Machine Learning Computation Graphs Dighe, Kaustubh Amarasinghe, Saman Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Machine learning applications are increasingly requiring fast and more computational power. Many applications like language models have become so large that they are run on distributed systems in parallel. However, getting into the details of optimally scheduling or even just running machine learning models on distributed systems can be a distraction for researchers ideating models. Hence there has been development of abstractions to facilitate running machine learning models in parallel on distributed systems. We present a compiler for the StreamIt language- a language made for abstract signal processing and multicore programming. We use that abstraction as a way to distribute the computation of machine learning models programmed in PyTorch. M.Eng. 2024-09-16T13:48:18Z 2024-09-16T13:48:18Z 2024-05 2024-07-11T14:37:05.421Z Thesis https://hdl.handle.net/1721.1/156774 https://orcid.org/0009-0006-9656-1946 In Copyright - Educational Use Permitted Copyright retained by author(s) https://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Dighe, Kaustubh
Fast Multistage Compilation of Machine Learning Computation Graphs
title Fast Multistage Compilation of Machine Learning Computation Graphs
title_full Fast Multistage Compilation of Machine Learning Computation Graphs
title_fullStr Fast Multistage Compilation of Machine Learning Computation Graphs
title_full_unstemmed Fast Multistage Compilation of Machine Learning Computation Graphs
title_short Fast Multistage Compilation of Machine Learning Computation Graphs
title_sort fast multistage compilation of machine learning computation graphs
url https://hdl.handle.net/1721.1/156774
https://orcid.org/0009-0006-9656-1946
work_keys_str_mv AT dighekaustubh fastmultistagecompilationofmachinelearningcomputationgraphs