A security framework in G-Hadoop for big data computing across distributed Cloud data centres

Jiaqi Zhao, Lizhe Wang*, Jie Tao, Jinjun Chen, Weiye Sun, Rajiv Ranjan, Joanna Kołodziej, Achim Streit, Dimitrios Georgakopoulos

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

103 Citations (Scopus)

Abstract

MapReduce is regarded as an adequate programming model for large-scale data-intensive applications. The Hadoop framework is a well-known MapReduce implementation that runs the MapReduce tasks on a cluster system. G-Hadoop is an extension of the Hadoop MapReduce framework with the functionality of allowing the MapReduce tasks to run on multiple clusters. However, G-Hadoop simply reuses the user authentication and job submission mechanism of Hadoop, which is designed for a single cluster. This work proposes a new security model for G-Hadoop. The security model is based on several security solutions such as public key cryptography and the SSL protocol, and is dedicatedly designed for distributed environments. This security framework simplifies the users authentication and job submission process of the current G-Hadoop implementation with a single-sign-on approach. In addition, the designed security framework provides a number of different security mechanisms to protect the G-Hadoop system from traditional attacks.

Original languageEnglish
Pages (from-to)994-1007
Number of pages14
JournalJournal of Computer and System Sciences
Volume80
Issue number5
DOIs
Publication statusPublished - Aug 2014
Externally publishedYes

Fingerprint

Dive into the research topics of 'A security framework in G-Hadoop for big data computing across distributed Cloud data centres'. Together they form a unique fingerprint.

Cite this