@inproceedings{ed415928685040aaaafc6e076fb0ffd8,
title = "Fast prototyping of Quantized neural networks on an FPGA edge computing device with Brevitas and FINN",
abstract = "In this paper, we propose a solution for fast proto-typing of Deep learning neural network models on edge computing devices like FPGA for researchers with limited knowledge of high level languages like VHDL. We use Xilinx' Brevitas tool for Quantization and FINN framework for deployment/inference on Pynq-Z2 board. The paper will also share presently available methods for FPGA prototyping and how tools like Brevitas and FINN can be used for more efficient inference of DNN on small scale edge computers like FPGA by levaraging their 1. Quantization Aware Training(QAT) and Post Training Quanti-zation(PTQ) 2. Streamlining networks and transformations 3. Dataflow partitioning of the NN model using FINN compiler 4. DMA, FIFO and IP generation for HW build and 5. Inference on FPGA using PYNQ python Driver. The weights and activations of a custom model were quantised from floating points to 8, 4 and 2 bit for which an accuracy drop of 0.1 %, 0.8% and 7.6% was observed respectively.",
keywords = "Brevitas, CNN, DNN, FINN, FPGA, Neural Network, Pynq, Quantization",
author = "Devansh Chawda and Benaoumeur Senouci",
note = "Publisher Copyright: {\textcopyright} 2024 IEEE.; 15th International Conference on Ubiquitous and Future Networks, ICUFN 2024 ; Conference date: 02-07-2024 Through 05-07-2024",
year = "2024",
doi = "10.1109/ICUFN61752.2024.10625618",
language = "English",
series = "International Conference on Ubiquitous and Future Networks, ICUFN",
publisher = "IEEE Press",
pages = "238--240",
booktitle = "2024 Fifteenth International Conference on Ubiquitous and Future Networks (ICUFN)",
}