skip to main content
10.1145/3627535.3638568acmconferencesArticle/Chapter ViewAbstractPublication PagesppoppConference Proceedingsconference-collections
keynote

Sparsity in Deep Neural Nets (Keynote)

Published:20 February 2024Publication History

ABSTRACT

Our brain executes very sparse computation, allowing for great speed and energy savings. Deep neural networks can also be made to exhibit high levels of sparsity without significant accuracy loss. As their size grows, it is becoming imperative that we use sparsity to improve their efficiency. This is a challenging task because the memory systems and SIMD operations that dominate todays CPUs and GPUs do not lend themselves easily to the irregular data patterns sparsity introduces. This talk will survey the role of sparsity in neural network computation, and the parallel algorithms and hardware features that nevertheless allow us to make effective use of it.

Index Terms

  1. Sparsity in Deep Neural Nets (Keynote)
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          PPoPP '24: Proceedings of the 29th ACM SIGPLAN Annual Symposium on Principles and Practice of Parallel Programming
          March 2024
          498 pages
          ISBN:9798400704352
          DOI:10.1145/3627535

          Copyright © 2024 Owner/Author(s)

          Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s).

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 20 February 2024

          Check for updates

          Qualifiers

          • keynote

          Acceptance Rates

          Overall Acceptance Rate230of1,014submissions,23%
        • Article Metrics

          • Downloads (Last 12 months)169
          • Downloads (Last 6 weeks)50

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader