Přístupnostní navigace
E-přihláška
Vyhledávání Vyhledat Zavřít
Detail publikace
BRAVENEC, T. FRÝZA, T.
Originální název
Reducing Memory Requirements of Convolutional Neural Networks for Inference at the Edge
Typ
článek ve sborníku ve WoS nebo Scopus
Jazyk
angličtina
Originální abstrakt
The main focus of this paper is to use post training quantization to analyse the influence of using lower precision data types in neural networks, while avoiding the process of retraining the networks in question. The main idea is to enable usage of high accuracy neural networks in devices other than high performance servers or super computers and bring the neural network compute closer to the device collecting the data. There are two main issues with using neural networks on edge devices, the memory constraint and the computational performance. Both of these issues could be diminished if the usage of lower precision data types does not considerably reduce the accuracy of the networks in question.
Klíčová slova
deep learning; neural networks; computer vision; machine learning; parallel computing; inference optimization; inference at the edge; reduced precision computing
Autoři
BRAVENEC, T.; FRÝZA, T.
Vydáno
20. 4. 2021
Nakladatel
Vysoké učení technické v Brně
Místo
Brno
ISBN
978-0-7381-4436-8
Kniha
International Conference Radioelektronika 2021
Strany od
1
Strany do
6
Strany počet
URL
https://ieeexplore.ieee.org/document/9420214
BibTex
@inproceedings{BUT171248, author="Tomáš {Bravenec} and Tomáš {Frýza}", title="Reducing Memory Requirements of Convolutional Neural Networks for Inference at the Edge", booktitle="International Conference Radioelektronika 2021", year="2021", pages="1--6", publisher="Vysoké učení technické v Brně", address="Brno", doi="10.1109/RADIOELEKTRONIKA52220.2021.9420214", isbn="978-0-7381-4436-8", url="https://ieeexplore.ieee.org/document/9420214" }