Very Low Power Neural Network FPGA Accelerators for Tag-Less Remote Person Identification Using Capacitive Sensors
Autor: | Abdelkader Mami, Marwen Roukhami, Francesco Gregoretti, Younes Lahbib, Mihai Teodor Lazarescu |
---|---|
Rok vydání: | 2019 |
Předmět: |
General Computer Science
Computer science Capacitive sensing Acceleration Capacitance Capacitive sensors 02 engineering and technology Embedded design optimization hardware acceleration Frequency measurement 0202 electrical engineering electronic engineering information engineering ultra-low power FPGAs General Materials Science Field-programmable gate array capacitive sensing Artificial neural networks Artificial neural network business.industry Field programmable gate arrays Capacitance measurement Indoor person identification Neural networks Hardware design Ultra-low power FPGAs Hardware acceleration 020208 electrical & electronic engineering General Engineering AC power neural networks Microcontroller hardware design 020201 artificial intelligence & image processing lcsh:Electrical engineering. Electronics. Nuclear engineering business lcsh:TK1-9971 Computer hardware Data transmission |
Zdroj: | IEEE Access, Vol 7, Pp 102217-102231 (2019) |
ISSN: | 2169-3536 |
DOI: | 10.1109/access.2019.2931392 |
Popis: | Human detection, identification, and monitoring are essential for many applications aiming to make smarter the indoor environments, where most people spend much of their time (like home, office, transportation, or public spaces). The capacitive sensors can meet stringent privacy, power, cost, and unobtrusiveness requirements, they do not rely on wearables or specific human interactions, but they may need significant on-board data processing to increase their performance. We comparatively analyze in terms of overall processing time and energy several data processing implementations of multilayer perceptron neural networks (NNs) on board capacitive sensors. The NN architecture, optimized using augmented experimental data, consists of six 17-bit inputs, two hidden layers with eight neurons each, and one four-bit output. For the software (SW) NN implementation, we use two STMicroelectronics STM32low-power ARM microcontrollers (MCUs): one MCU optimized for power and one for performance. For hardware (HW) implementations, we use four ultralow-power field-programmable gate arrays (FPGAs), with different sizes, dedicated computation blocks, and data communication interfaces (one FPGA from the Lattice iCE40 family and three FPGAs from the Microsemi IGLOO family). Our shortest SW implementation latency is 54.4 μs and the lowest energy per inference is 990 nJ, while the shortest HW implementation latency is 1.99 μs and the lowest energy is 39 nJ (including the data transfer between MCU and FPGA). The FPGAs active power ranges between 6.24 and 34.7 mW, while their static power is between 79 and 277 μW. They compare very favorably with the static power consumption of Xilinx and Altera low-power device families, which is around 40 mW. The experimental results show that NN inferences offloaded to external FPGAs have lower latency and energy than SW ones (even when using HW multipliers), and the FPGAs with dedicated computational blocks (multiply-accumulate) perform best. |
Databáze: | OpenAIRE |
Externí odkaz: |