ABSTRACT
MapReduce was designed by Google for large-scale data analysis on slow but cheap disk-based storage. Nevertheless, memory has declined in price to where cost-effective machines offer ever larger memory capacity. Furthermore, a more diverse data analyst community, with smaller datasets, has emerged. These trends motivate new parallel processing frameworks, like Spark [2], with better support for in-memory data analysis.
- T. Kelly and D. Reeves. Optimal web cache sizing: Scalable methods for exact solutions. In Fifth Int'l Web Caching and Content Delivery Workshop, 2000.Google Scholar
- M. Zaharia et al. Spark: Cluster computing with working sets. In HotCloud '10, 2010. Google ScholarDigital Library
Index Terms
- Recommending just enough memory for analytics
Recommendations
Just can't get enough: Synthesizing Big Data
SIGMOD '15: Proceedings of the 2015 ACM SIGMOD International Conference on Management of DataWith the rapidly decreasing prices for storage and storage systems ever larger data sets become economical. While only few years ago only successful transactions would be recorded in sales systems, today every user interaction will be stored for ever ...
Big Data Analytics Based on In-Memory Infrastructure On Traditional HPC: A Survey
ICTCS '16: Proceedings of the Second International Conference on Information and Communication Technology for Competitive StrategiesAs the capacity of main memory is growing, in-memory based big data analytics is becoming more popular. In-memory technologies support interactive analysis by providing high I/O throughput. On traditional high performance computing (HPC), big data ...
Comments