energy efficient inference accelerator for memory augmented neural networks on an fpga
FOS: Computer and information sciences
Computer Science - Machine Learning
Machine Learning (stat.ML)
02 engineering and technology
01 natural sciences
7. Clean energy
Machine Learning (cs.LG)
Statistics - Machine Learning
0103 physical sciences
0202 electrical engineering, electronic engineering, information engineering
DOI:
10.48550/arxiv.1805.07978
Publication Date:
2019-03-01
AUTHORS (4)
ABSTRACT
Memory-augmented neural networks (MANNs) are designed for question-answering tasks. It is difficult to run a MANN effectively on accelerators designed for other neural networks (NNs), in particular on mobile devices, because MANNs require recurrent data paths and various types of operations related to external memory access. We implement an accelerator for MANNs on a field-programmable gate array (FPGA) based on a data flow architecture. Inference times are also reduced by inference thresholding, which is a data-based maximum inner-product search specialized for natural language tasks. Measurements on the bAbI data show that the energy efficiency of the accelerator (FLOPS/kJ) was higher than that of an NVIDIA TITAN V GPU by a factor of about 125, increasing to 140 with inference thresholding<br/>Accepted to DATE 2019<br/>
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....