Downloads: 112 | Views: 167
Survey Paper | Computer Science & Engineering | India | Volume 5 Issue 7, July 2016
Survey Paper on Big Data Processing and Hadoop Components
Swati M. Gavali | Supriya Sarkar
Abstract: The term Big Data, refers to data sets whose size (volume), complexity (variability), and rate of growth (velocity) make them difficult to capture, manage, process or analyzed. To analyze this enormous amount of data Hadoop can be used. However, processing is often time-consuming. One way to decrease response time is to executing the job partially, where an approximate, early result becomes available to the user, before completion of job. The implementation of the technique will be on top of Hadoop which will help to sample HDFS blocks uniformly. We will evaluate this technique using real-world datasets and applications and we will try to demonstrate the systems performance in terms of accuracy and time. The objective of the proposed technique is to significantly improve the performance of Hadoop MapReduce for efficient Big Data processing.
Keywords: Big data, Hadoop, MapReduce, RDMS
Edition: Volume 5 Issue 7, July 2016,
Pages: 1333 - 1337