4.3 Article Proceedings Paper

Reconfigurable VLSI implementation for learning vector quantization with on-chip learning circuit

期刊

出版社

IOP PUBLISHING LTD
DOI: 10.7567/JJAP.55.04EF02

关键词

-

资金

  1. Grants-in-Aid for Scientific Research [25420332] Funding Source: KAKEN

向作者/读者索取更多资源

As an alternative to conventional single-instruction-multiple-data (SIMD) mode solutions with massive parallelism for self-organizing-map (SOM) neural network models, this paper reports a memory-based proposal for the learning vector quantization (LVQ), which is a variant of SOM. A dual-mode LVQ system, enabling both on-chip learning and classification, is implemented by using a reconfigurable pipeline with parallel p-word input (R-PPPI) architecture. As a consequence of the reuse of R-PPPI for solving the most severe computational demands in both modes, power dissipation and Si-area consumption can be dramatically reduced in comparison to previous LVQ implementations. In addition, the designed LVQ ASIC has high flexibility with respect to feature-vector dimensionality and reference-vector number, allowing the execution of many different machine-learning applications. The fabricated test chip in 180nm CMOS with parallel 8-word inputs and 102 K-bit on-chip memory achieves low power consumption of 66.38mW (at 75MHz and 1.8 V) and high learning speed of (R + 1) x [d/8] + 10 clock cycles per d-dimensional sample vector where R is the reference-vector number. (C) 2016 The Japan Society of Applied Physics

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据