4.5 Article Proceedings Paper

Accelerated Deep Reinforcement Learning for Fast Feedback of Beam Dynamics at KARA

Journal

IEEE TRANSACTIONS ON NUCLEAR SCIENCE
Volume 68, Issue 8, Pages 1794-1800

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNS.2021.3084515

Keywords

Hardware; Radio frequency; Real-time systems; Detectors; Control systems; Reinforcement learning; Perturbation methods; Artificial intelligence; beam diagnostics; control systems; machine learning; particle accelerators; real-time control; reinforcement learning (RL)

Ask authors/readers for more resources

Coherent synchrotron radiation (CSR) is generated when the electron bunch length is in the order of the magnitude of the wavelength of the emitted radiation. To stabilize the CSR emission, a real-time feedback control loop based on reinforcement learning (RL) is proposed. The hardware implementation of the RL controller shows similar training performance to the software solution, but outperforms the Keras-RL implementation by an order of magnitude, indicating its potential as an essential platform for developing intelligent CSR control systems.
Coherent synchrotron radiation (CSR) is generated when the electron bunch length is in the order of the magnitude of the wavelength of the emitted radiation. The self-interaction of short electron bunches with their own electromagnetic fields changes the longitudinal beam dynamics significantly. Above a certain current threshold, the micro-bunching instability develops, characterized by the appearance of distinguishable substructures in the longitudinal phase space of the bunch. To stabilize the CSR emission, a real-time feedback control loop based on reinforcement learning (RL) is proposed. Informed by the available THz diagnostics, the feedback is designed to act on the radio frequency (RF) system of the storage ring to mitigate the micro-bunching dynamics. To satisfy low-latency requirements given by the longitudinal beam dynamics, the RL controller has been implemented on hardware (FPGA). In this article, a real-time feedback loop architecture and its performance is presented and compared with a software implementation using Keras-RL on CPU/GPU. The results obtained with the CSR simulation Inovesa demonstrate that the functionality of both platforms is equivalent. The training performance of the hardware implementation is similar to software solution, while it outperforms the Keras-RL implementation by an order of magnitude. The presented RL hardware controller is considered as an essential platform for the development of intelligent CSR control systems.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available