International Journal of Computer Applications
The data is exceedingly large day-by-day. In some organizations, there is a need to analyze and process the gigantic data. This is a big data problem often faced by these organizations. It is not possible for single machine to handle that data. So the authors have used Apache Hadoop Distributed File System (HDFS) for storage and analysis. This paper shows experimental work done on the MapReduce application on health sector dataset. The result shows the behavior of the MapReduce application framework to map and reduce the big volume of the data.