Article ID | Journal | Published Year | Pages | File Type |
---|---|---|---|---|
4947125 | Neurocomputing | 2017 | 31 Pages |
Abstract
We demonstrate the use of a low-power and compact hardware implementation of Random Feature Extractor (RFE) core in image recognition applications. We show that weight distributions with zero mean are necessary for good performance in this application. We also show the performance in this application of a recently proposed weight reuse technique that virtually expands the number of random features available from RFE core. Further, we propose a method of reducing computation cost by pruning “incognizant” or redundant random feature neurons. For proof of concept, we validated our approach by using our RFE core as the first stage of Extreme Learning Machine (ELM) - a two layer neural network - and using 12, 800 hidden neurons were able to achieve upto 97.8% accuracy on MNIST database of handwritten digits. The RFE ASIC occupies 5 mm à 5 mm area in 0.35âµm CMOS process. By using the proposed pruning method, an accuracy of 97% is obtained by using only 4749 effective hidden neurons and consuming 6.1 µJ/classify. Input pre-processing and ELM second stage requiring just addition operations can be implemented using digital adders with estimated power consumption of 24.9 nJ/classify. With a total energy consumption of only 6.12 µJ/classify, this low-power mixed signal ASIC can act as a co-processor in portable electronic gadgets with cameras.
Related Topics
Physical Sciences and Engineering
Computer Science
Artificial Intelligence
Authors
Aakash Patil, Shanlan Shen, Enyi Yao, Arindam Basu,