4.8 Article

Scalar Quantization as Sparse Least Square Optimization

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2019.2952096

关键词

Scalar quantization; l(0) least square; l(1) least square; clustering; approximation

资金

  1. National Natural Science Foundation of China [61876114]
  2. Research Project of State Key Laboratory of Southwest Jiaotong University [TPL1502]
  3. University-Enterprise Cooperation Project [17H1199, 19H0355]

向作者/读者索取更多资源

This paper investigates the application of scalar quantization based on sparse least square optimization, proposing multiple quantization algorithms based on different regularization criteria, and comparing and testing them through iterative methods and clustering-based methods.
Quantization aims to form new vectors or matrices with shared values close to the original. In recent years, the popularity of scalar quantization has been soaring as it is found huge utilities in reducing the resource cost of neural networks. Popular clustering-based techniques suffers substantially from the problems of dependency on the seed, empty or out-of-the-range clusters, and high time complexity. To overcome the problems, in this paper, scalar quantization is examined from a new perspective, namely sparse least square optimization. Specifically, several quantization algorithms based on l(1) least square are proposed and implemented. In addition, similar schemes with l(1) + l(2) and l(0) regularization are proposed. Furthermore, to compute quantization results with given amount of values/clusters, this paper proposes an iterative method and a clustering-based method, and both of them are built on sparse least square optimization. The algorithms proposed are tested under three data scenarios and their computational performance, including information loss, time consumption, and distribution of values of sparse vectors are compared. The paper offers a new perspective to probe the area of quantization, and the algorithms proposed are superior especially under bit-width reduction scenarios, where the required post-quantization resolution (the number of values) is not significantly lower than the original scalar.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据