Masami Akamine, Jitendra Ajmera
IEICE Trans Inf Syst
Kernel functions are vital ingredients of several machine learning (ML) algorithms but often incur substantial memory and computational costs. We introduce an approach to kernel approximation in ML algorithms suitable for mixed-signal analogue in-memory computing (AIMC) architectures. Analogue in-memory kernel approximation addresses the performance bottlenecks of conventional kernel-based methods by executing most operations in approximate kernel methods directly in memory. The IBM HERMES project chip, a state-of-the-art phase-change memory-based AIMC chip, is utilized for the hardware demonstration of kernel approximation. Experimental results show that our method maintains high accuracy, with less than a 1% drop in kernel-based ridge classification benchmarks and within 1% accuracy on the long-range arena benchmark for kernelized attention in transformer neural networks. Compared to traditional digital accelerators, our approach is estimated to deliver superior energy efficiency and lower power consumption. These findings highlight the potential of heterogeneous AIMC architectures to enhance the efficiency and scalability of ML applications.
Masami Akamine, Jitendra Ajmera
IEICE Trans Inf Syst
Hironori Takeuchi, Tetsuya Nasukawa, et al.
Transactions of the Japanese Society for Artificial Intelligence
S. Winograd
Journal of the ACM
Hagen Soltau, Lidia Mangu, et al.
ASRU 2011