LlamaF: An Efficient Llama2 Architecture Accelerator on Embedded FPGAs

Autor: Xu, Han, Li, Yutong, Ji, Shihao
Rok vydání: 2024
Předmět:
Druh dokumentu: Working Paper
Popis: Large language models (LLMs) have demonstrated remarkable abilities in natural language processing. However, their deployment on resource-constrained embedded devices remains difficult due to memory and computational demands. In this paper, we present an FPGA-based accelerator designed to improve LLM inference performance on embedded FPGAs. We employ post-training quantization to reduce model size and optimize for off-chip memory bandwidth. Our design features asynchronous computation and a fully pipelined accelerator for matrix-vector multiplication. Experiments of the TinyLlama 1.1B model on a Xilinx ZCU102 platform show a 14.3-15.8x speedup and a 6.1x power efficiency improvement over running exclusively on ZCU102 processing system (PS).
Databáze: arXiv