Physical Review Accelerators and Beams (Dec 2020)

Sample-efficient reinforcement learning for CERN accelerator control

  • Verena Kain,
  • Simon Hirlander,
  • Brennan Goddard,
  • Francesco Maria Velotti,
  • Giovanni Zevi Della Porta,
  • Niky Bruchon,
  • Gianluca Valentino

DOI
https://doi.org/10.1103/PhysRevAccelBeams.23.124801
Journal volume & issue
Vol. 23, no. 12
p. 124801

Abstract

Read online Read online

Numerical optimization algorithms are already established tools to increase and stabilize the performance of particle accelerators. These algorithms have many advantages, are available out of the box, and can be adapted to a wide range of optimization problems in accelerator operation. The next boost in efficiency is expected to come from reinforcement learning algorithms that learn the optimal policy for a certain control problem and hence, once trained, can do without the time-consuming exploration phase needed for numerical optimizers. To investigate this approach, continuous model-free reinforcement learning with up to 16 degrees of freedom was developed and successfully tested at various facilities at CERN. The approach and algorithms used are discussed and the results obtained for trajectory steering at the AWAKE electron line and LINAC4 are presented. The necessary next steps, such as uncertainty aware model-based approaches, and the potential for future applications at particle accelerators are addressed.