Model compression and simplification pipelines for fast deep neural network inference in FPGAs in HEP

Abstract Resource utilization plays a crucial role for successful implementation of fast real-time inference for deep neural networks (DNNs) and convolutional neural networks (CNNs) on latest generation of hardware accelerators (FPGAs, SoCs, ACAPs, GPUs). To fulfil the needs of the triggers that are...

Full description

Bibliographic Details
Main Authors: Simone Francescato, Stefano Giagu, Federica Riti, Graziella Russo, Luigi Sabetta, Federico Tortonesi
Format: Article
Language:English
Published: SpringerOpen 2021-11-01
Series:European Physical Journal C: Particles and Fields
Online Access:https://doi.org/10.1140/epjc/s10052-021-09770-w