Abstract:
Transformers are increasingly used to process time series data. Their deployment in mobile devices is however challenging due to their large computational requirements. H...Show MoreMetadata
Abstract:
Transformers are increasingly used to process time series data. Their deployment in mobile devices is however challenging due to their large computational requirements. Hardware acceleration through custom neural network processors can reduce the resulting latency and energy footprint of the network. Yet, the large variety of different layer types in mobile transformers troubles the selection of the best accelerator architecture and hardware mapping. Specifically, the layer performance strongly depends on the spatial unrolling dimensions, i.e. the layer loop dimensions along which hardware parallelization is enabled. This paper will therefore research the best datapath organization, and required datapath flexibility to efficiently support mobile transformers. The Mobile ViTS network is selected as the reference network for its wide variety in layer types. Results are explored across a wide range of accelerator area (datapath dimensions, memory size) and bandwidth constraints.
Published in: 2022 IEEE 4th International Conference on Artificial Intelligence Circuits and Systems (AICAS)
Date of Conference: 13-15 June 2022
Date Added to IEEE Xplore: 05 September 2022
ISBN Information: