4.1 Article

Reinforcement Learning Enabled Autonomous Manufacturing Using Transfer Learning and Probabilistic Reward Modeling

Journal

IEEE CONTROL SYSTEMS LETTERS
Volume 7, Issue -, Pages 508-513

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LCSYS.2022.3188014

Keywords

Task analysis; Decision making; Transfer learning; Probabilistic logic; Reinforcement learning; Costs; Gaussian processes; Reinforcement learning; transfer learning; Gaussian process; autonomous manufacturing

Ask authors/readers for more resources

This paper proposes a reinforcement learning enabled physical autonomous manufacturing system that can autonomously fabricate complex-geometry artifacts by learning manufacturing process parameters. To improve sample efficiency, the authors use first-principles based source task for training, transfer effective representations, and learn a probabilistic model of the target reward function. The effectiveness of the method is demonstrated through experiments on a custom AMS machine.
Here we propose a reinforcement learning enabled physical autonomous manufacturing system (AMS) that is capable of learning the manufacturing process parameters to autonomously fabricate a complex-geometry artifact with desired performance characteristics. The poor sample efficiency of traditional RL algorithms challenges real-world manufacturing decision making due to a high variable cost from raw material, machine utilization, and labor costs. To make decision making sample efficient, we propose to leverage a first-principles based source task for training, transfer effective representations from trained knowledge, and then use these representations to interact with the physical system to learn a probabilistic model of the target reward function. We deploy this idea to a novel dataset obtained from a custom physical AMS machine that can autonomously manufacture phononic crystals, a complex geometry artifact with spectral response as performance characteristic. We demonstrate that our method uses as low as 25 artifacts to model the interesting part of the target reward function and find an artifact with high reward. This task typically requires manual design of phononic crystals and extensive empirical iterations on the order of hundreds.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.1
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available