Downloads: 132 | Views: 176
Survey Paper | Computer Science & Engineering | India | Volume 4 Issue 12, December 2015
Big Data Hadoop: Aggregation Techniques
Abstract: The term Big Data, refers to data sets whose size (volume), complexity (variability), and rate of growth (velocity) make them difficult to capture, manage, process or analyzed. To analyze this enormous amount of data Hadoop can be used. However, processing is often time-consuming. One way to decrease response time is to executing the job partially, where an approximate, early result becomes available to the user, before completion of job. The implementation of the technique will be on top of Hadoop which will help to sample HDFS blocks uniformly. We will evaluate this technique using real-world datasets and applications and we will try to demonstrate the systems performance in terms of accuracy and time. The objective of the proposed technique is to significantly improve the performance of Hadoop MapReduce for efficient Big Data processing.
Keywords: privacy preservation, security, e-healthcare systems, data mining, image feature extraction
Edition: Volume 4 Issue 12, December 2015,
Pages: 432 - 435