Model compression and simplification pipelines for fast deep neural network inference in FPGAs in HEP
Abstract Resource utilization plays a crucial role for successful implementation of fast real-time inference for deep neural networks (DNNs) and convolutional neural networks (CNNs) on latest generation of hardware accelerators (FPGAs, SoCs, ACAPs, GPUs). To fulfil the needs of the triggers that are...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
SpringerOpen
2021-11-01
|
Series: | European Physical Journal C: Particles and Fields |
Online Access: | https://doi.org/10.1140/epjc/s10052-021-09770-w |