ISSN: 2319-7064
Iniyan Senthil Kumar.T, Poovaraghan.R
Abstract: MapReduce is the program model and an association implementation for processing and generating large data sets. User specify a map function that process a key/value pair to generate a set of intermediate key/value pairs and a reduce function that merge all intermediate values associated with the same intermediate key. Our Implementation of MapReduce runs on a large cluster of commodity machines and is highly scalable, a typical MapReduce computation processes many terabytes of data of thousand machines
Keywords: MapReduce, Hadoop, Cluster, Large scale indexing, ordering