Dynamic Replication Policy on HDFS Based on Machine Learning Clustering
Data growth in recent years has been swift, leading to the emergence of big data science. Distributed File Systems (DFS) are commonly used to handle big data, like Google File System (GFS), Hadoop Distributed File System (HDFS), and others. The DFS should provide the availability of data and reliabi...
Main Authors: | Motaz A. Ahmed, Mohamed H. Khafagy, Masoud E. Shaheen, Mostafa R. Kaseb |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2023-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10049393/ |
Similar Items
-
BlockHDFS: Blockchain-integrated Hadoop distributed file system for secure provenance traceability
by: Viraaji Mothukuri, et al.
Published: (2021-12-01) -
HaRD: a heterogeneity-aware replica deletion for HDFS
by: Hilmi Egemen Ciritoglu, et al.
Published: (2019-10-01) -
A comparison of HDFS compact data formats: Avro versus Parquet / HDFS glaustųjų duomenų formatų palyginimas: Avro prieš Parquet
by: Daiga Plase, et al.
Published: (2017-07-01) -
Big Data Architectures and Concepts
by: Audrey Tembo Welo, et al.
Published: (2023-12-01) -
A Survey on Job Scheduling in Big Data
by: Senthilkumar M., et al.
Published: (2016-09-01)