A dynamic data placement policy for heterogeneous Hadoop cluster

Santa Maria Shithil, Tushar Kanti Saha, Tanusree Sharma

Research output: Chapter in Book/Report/Conference proceedingConference contribution

8 Scopus citations

Abstract

Hadoop Distributed File System (HDFS) is a file system for storing and managing big data. The current HDFS block placement policy works well for the homogeneous cluster. However, it can not evenly and fairly distribute blocks across the heterogeneous cluster and results in an unbalanced cluster. An unbalanced cluster also reduces MapReduce performance. Hadoop relies on load balancer tool to balance replica distributions that in turn degrade the overall performance of hadoop. In this paper, we proposed a dynamic block placement policy that distributes blocks across a heterogeneous cluster more evenly than the current HDFS block placement policy and also improves the performance of MapReduce applications. The solution had been implemented and test had been conducted to evaluate its contribution to Hadoop.

Original languageEnglish (US)
Title of host publication4th International Conference on Advances in Electrical Engineering, ICAEE 2017
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages302-307
Number of pages6
ISBN (Electronic)9781538608692
DOIs
StatePublished - Jul 1 2017
Event4th International Conference on Advances in Electrical Engineering, ICAEE 2017 - Dhaka, Bangladesh
Duration: Sep 28 2017Sep 30 2017

Publication series

Name4th International Conference on Advances in Electrical Engineering, ICAEE 2017
Volume2018-January

Conference

Conference4th International Conference on Advances in Electrical Engineering, ICAEE 2017
Country/TerritoryBangladesh
CityDhaka
Period9/28/179/30/17

All Science Journal Classification (ASJC) codes

  • Computer Networks and Communications
  • Computer Science Applications
  • Biomedical Engineering
  • Electrical and Electronic Engineering
  • Instrumentation

Fingerprint

Dive into the research topics of 'A dynamic data placement policy for heterogeneous Hadoop cluster'. Together they form a unique fingerprint.

Cite this