Hadoop Distributed File System for Load Balancing Using Data Declustering Techniques
The Big-data refers to the large-scale distributed data processing applications that operate on unusually huge amounts of data. Google’s Map Reduce and Apache’s Map Reduce, its open-source implementation, are the defector software systems for large scale data applications. Study of the Map Reduce framework is that the framework produces a large amount of intermediate data. Such existing information is thrown away after the tasks finish, because Map Reduce is not able to utilize them. In this paper, they propose, a data-aware cache framework for large data applications.
Subscribe to the Data Insider Newsletter
Learn the latest news and best practices about data science, big data analytics, artificial intelligence, data security, and more. Delivered Mondays and Thursdays