MAHASIM: Machine-Learning Hardware Acceleration Using a Software-Defined Intelligent Memory System

Autor: Sudhakar Yalamanchili, Saibal Mukhopadhyay, Bahar Asgari
Rok vydání: 2020
Předmět:
Zdroj: Journal of Signal Processing Systems. 93:659-675
ISSN: 1939-8115
1939-8018
DOI: 10.1007/s11265-019-01505-1
Popis: As computations in machine-learning applications are increasing simultaneously along the size of datasets, the energy and performance costs of data movement dominate that of compute. This issue is more pronounced in embedded systems with limited resources and energy. Although near-data-processing (NDP) is pursued as an architectural solution, comparatively less attention has been focused on how to scale NDP for larger-scale embedded machine learning applications (e.g., speech and motion processing). We propose machine-learning hardware acceleration using a software-defined intelligent memory system (Mahasim). Mahasim is a scalable NDP-based memory system, in which application performance scales with the size of data. The building blocks of Mahasim are the programable memory slices, supported by data partitioning, compute-aware memory allocation, and an independent in-memory execution model. For recurrent neural networks, Mahasim shows up to 537.95 GFLOPS/W energy efficiency and 3.9x speedup, when the size of the system increases from 2 to 256 memory slices, which indicates that Mahasim favors larger problems.
Databáze: OpenAIRE