Complex & Intelligent Systems (Jan 2025)

RL4CEP: reinforcement learning for updating CEP rules

  • Afef Mdhaffar,
  • Ghassen Baklouti,
  • Yassine Rebai,
  • Mohamed Jmaiel,
  • Bernd Freisleben

DOI
https://doi.org/10.1007/s40747-024-01742-3
Journal volume & issue
Vol. 11, no. 2
pp. 1 – 15

Abstract

Read online

Abstract This paper presents RL4CEP, a reinforcement learning (RL) approach to dynamically update complex event processing (CEP) rules. RL4CEP uses Double Deep Q-Networks to update the threshold values used by CEP rules. It is implemented using Apache Flink as a CEP engine and Apache Kafka for message distribution. RL4CEP is a generic approach for scenarios in which CEP rules need to be updated dynamically. In this paper, we use RL4CEP in a financial trading use case. Our experimental results based on three financial trading rules and eight financial datasets demonstrate the merits of RL4CEP in improving the overall profit, when compared to baseline and state-of-the-art approaches, with a reasonable consumption of resources, i.e., RAM and CPU. Finally, our experiments indicate that RL4CEP is executed quite fast compared to traditional CEP engines processing static rules.

Keywords