A unified schedule policy of distributed machine learning framework for CPU-GPU cluster

With the widespread using of GPU hardware facilities, more and more distributed machine learning applications have begun to use CPU-GPU hybrid cluster resources to improve the efficiency of algorithms. However, the existing distributed machine learning scheduling framework either only considers task...

ver descrição completa

Detalhes bibliográficos
Formato: Artigo
Idioma:zho
Publicado em: EDP Sciences 2021-06-01
coleção:Xibei Gongye Daxue Xuebao
Assuntos:
Acesso em linha:https://www.jnwpu.org/articles/jnwpu/full_html/2021/03/jnwpu2021393p529/jnwpu2021393p529.html
_version_ 1827610168100126720
collection DOAJ
description With the widespread using of GPU hardware facilities, more and more distributed machine learning applications have begun to use CPU-GPU hybrid cluster resources to improve the efficiency of algorithms. However, the existing distributed machine learning scheduling framework either only considers task scheduling on CPU resources or only considers task scheduling on GPU resources. Even considering the difference between CPU and GPU resources, it is difficult to improve the resource usage of the entire system. In other words, the key challenge in using CPU-GPU clusters for distributed machine learning jobs is how to efficiently schedule tasks in the job. In the full paper, we propose a CPU-GPU hybrid cluster schedule framework in detail. First, according to the different characteristics of the computing power of the CPU and the computing power of the GPU, the data is divided into data fragments of different sizes to adapt to CPU and GPU computing resources. Second, the paper introduces the task scheduling method under the CPU-GPU hybrid. Finally, the proposed method is verified at the end of the paper. After our verification for K-Means, using the CPU-GPU hybrid computing framework can increase the performance of K-Means by about 1.5 times. As the number of GPUs increases, the performance of K-Means can be significantly improved.
first_indexed 2024-03-09T07:46:24Z
format Article
id doaj.art-0f11b0cc4a2f47c2ba34ab7e1afd91e3
institution Directory Open Access Journal
issn 1000-2758
2609-7125
language zho
last_indexed 2024-03-09T07:46:24Z
publishDate 2021-06-01
publisher EDP Sciences
record_format Article
series Xibei Gongye Daxue Xuebao
spelling doaj.art-0f11b0cc4a2f47c2ba34ab7e1afd91e32023-12-03T03:14:26ZzhoEDP SciencesXibei Gongye Daxue Xuebao1000-27582609-71252021-06-0139352953810.1051/jnwpu/20213930529jnwpu2021393p529A unified schedule policy of distributed machine learning framework for CPU-GPU cluster012School of Computer Science, Northwestern Polytechnical UniversitySchool of Computer Science, Northwestern Polytechnical UniversitySchool of Computer Science, Northwestern Polytechnical UniversityWith the widespread using of GPU hardware facilities, more and more distributed machine learning applications have begun to use CPU-GPU hybrid cluster resources to improve the efficiency of algorithms. However, the existing distributed machine learning scheduling framework either only considers task scheduling on CPU resources or only considers task scheduling on GPU resources. Even considering the difference between CPU and GPU resources, it is difficult to improve the resource usage of the entire system. In other words, the key challenge in using CPU-GPU clusters for distributed machine learning jobs is how to efficiently schedule tasks in the job. In the full paper, we propose a CPU-GPU hybrid cluster schedule framework in detail. First, according to the different characteristics of the computing power of the CPU and the computing power of the GPU, the data is divided into data fragments of different sizes to adapt to CPU and GPU computing resources. Second, the paper introduces the task scheduling method under the CPU-GPU hybrid. Finally, the proposed method is verified at the end of the paper. After our verification for K-Means, using the CPU-GPU hybrid computing framework can increase the performance of K-Means by about 1.5 times. As the number of GPUs increases, the performance of K-Means can be significantly improved.https://www.jnwpu.org/articles/jnwpu/full_html/2021/03/jnwpu2021393p529/jnwpu2021393p529.htmlcpu-gpu tasksunified schedulerclustering algorithmdistribution
spellingShingle A unified schedule policy of distributed machine learning framework for CPU-GPU cluster
Xibei Gongye Daxue Xuebao
cpu-gpu tasks
unified scheduler
clustering algorithm
distribution
title A unified schedule policy of distributed machine learning framework for CPU-GPU cluster
title_full A unified schedule policy of distributed machine learning framework for CPU-GPU cluster
title_fullStr A unified schedule policy of distributed machine learning framework for CPU-GPU cluster
title_full_unstemmed A unified schedule policy of distributed machine learning framework for CPU-GPU cluster
title_short A unified schedule policy of distributed machine learning framework for CPU-GPU cluster
title_sort unified schedule policy of distributed machine learning framework for cpu gpu cluster
topic cpu-gpu tasks
unified scheduler
clustering algorithm
distribution
url https://www.jnwpu.org/articles/jnwpu/full_html/2021/03/jnwpu2021393p529/jnwpu2021393p529.html