3.8 Proceedings Paper

A Data-Centric Optimization Framework for Machine Learning

出版社

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3524059.3532364

关键词

deep neural network compilers; machine learning; training optimization

资金

  1. European Research Council under the European Union [101002047]
  2. EuroHPC-JU [955513, 955606]
  3. Horizon 2020 programme
  4. Swiss National Science Foundation [185778]
  5. ETH Postdoctoral Fellowship
  6. PASC program (Platform for Advanced Scientific Computing)
  7. Swiss National Supercomputing Center (CSCS)

向作者/读者索取更多资源

This article introduces a flexible and customizable deep learning model training optimization process based on data movement minimization, which helps researchers optimize training of arbitrary deep neural networks. By gradually transforming standard networks, it provides four levels of general-purpose transformations, from internal operator optimization to global data movement reduction.
Rapid progress in deep learning is leading to a diverse set of quickly changing models, with a dramatically growing demand for compute. However, as frameworks specialize performance optimization to patterns in popular networks, they implicitly constrain novel and diverse models that drive progress in research. We empower deep learning researchers by defining a flexible and user-customizable pipeline for optimizing training of arbitrary deep neural networks, based on data movement minimization. The pipeline begins with standard networks in PyTorch or ONNX and transforms computation through progressive lowering. We define four levels of general-purpose transformations, from local intra-operator optimizations to global data movement reduction. These operate on a data-centric graph intermediate representation that expresses computation and data movement at all levels of abstraction, including expanding basic operators such as convolutions to their underlying computations. Central to the design is the interactive and introspectable nature of the pipeline. Every part is extensible through a Python API, and can be tuned interactively using a GUI. We demonstrate competitive performance or speedups on ten different networks, with interactive optimizations discovering new opportunities in EfficientNet.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据