Loading [MathJax]/extensions/MathMenu.js
Large-scale neural modeling in MapReduce and Giraph | IEEE Conference Publication | IEEE Xplore

Large-scale neural modeling in MapReduce and Giraph


Abstract:

One of the most crucial challenges in scientific computing is scalability. Hadoop, an open-source implementation of the MapReduce parallel programming model developed by ...Show More

Abstract:

One of the most crucial challenges in scientific computing is scalability. Hadoop, an open-source implementation of the MapReduce parallel programming model developed by Google, has emerged as a powerful platform for performing large-scale scientific computing at very low costs. In this paper, we explore the use of Hadoop to model large-scale neural networks. A neural network is most naturally modeled by a graph structure with iterative processing. In this paper, we first present an improved graph algorithm design pattern in MapReduce called Mapper-side Schimmy. Experiments show that the application of our design pattern, combined with the current best practices, can reduce the running time of the neural network simulation on a neural network with 100,000 neurons and 2.3 billion edges by 64%. MapReduce, however, is inherently not efficient for iterative graph processing. To address the limitation of the MapReduce model, we then explore the use of Giraph, an open source large-scale graph processing framework that sits on top of Hadoop to implement graph algorithms with a vertex-centric approach. We show that our Giraph implementation boosted performance by 91% compared to a basic MapReduce implementation and by 60% compared to our improved Mapper-side Schimmy algorithm.
Date of Conference: 05-07 June 2014
Date Added to IEEE Xplore: 07 August 2014
Electronic ISBN:978-1-4799-4774-4

ISSN Information:

Conference Location: Milwaukee, WI, USA

Contact IEEE to Subscribe

References

References is not available for this document.