4.6 Article

QUANTILE REGRESSION UNDER MEMORY CONSTRAINT

期刊

ANNALS OF STATISTICS
卷 47, 期 6, 页码 3244-3273

出版社

INST MATHEMATICAL STATISTICS
DOI: 10.1214/18-AOS1777

关键词

Quantile regression; sample quantile; divide-and-conquer; distributed inference; streaming data

资金

  1. Adobe
  2. Bloomberg
  3. NSFC [11825104, 11431006, 11690013]
  4. Program for Professor of Special Appointment (Eastern Scholar) at Shanghai Institutions of Higher Learning, Youth Talent Support Program, 973 Program [2015CB856004]
  5. Australian Research Council
  6. Alibaba innovation

向作者/读者索取更多资源

This paper studies the inference problem in quantile regression (QR) for a large sample size n but under a limited memory constraint, where the memory can only store a small batch of data of size m. A natural method is the naive divide-and-conquer approach, which splits data into batches of size m, computes the local QR estimator for each batch and then aggregates the estimators via averaging. However, this method only works when n = o(m(2)) and is computationally expensive. This paper proposes a computationally efficient method, which only requires an initial QR estimator on a small batch of data and then successively refines the estimator via multiple rounds of aggregations. Theoretically, as long as n grows polynomially in m, we establish the asymptotic normality for the obtained estimator and show that our estimator with only a few rounds of aggregations achieves the same efficiency as the QR estimator computed on all the data. Moreover, our result allows the case that the dimensionality p goes to infinity. The proposed method can also be applied to address the QR problem under distributed computing environment (e.g., in a large-scale sensor network) or for real-time streaming data.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据