Dear all: There has been a lot of work on hardware acceleration of machine learning algorithms with FPGA, ASIC or GPU, especially for neural networks. I wonder if it makes sense, or if there is any prior work, to build custom hardware to accelerate HTM/NUPIC, in order to achieve real-time performance in an embedded environment?
Thanks! Sam Gu
