Mathematics (May 2025)

Overcoming Stagnation in Metaheuristic Algorithms with MsMA’s Adaptive Meta-Level Partitioning

  • Matej Črepinšek,
  • Marjan Mernik,
  • Miloš Beković,
  • Matej Pintarič,
  • Matej Moravec,
  • Miha Ravber

DOI
https://doi.org/10.3390/math13111803
Journal volume & issue
Vol. 13, no. 11
p. 1803

Abstract

Read online

Stagnation remains a persistent challenge in optimization with metaheuristic algorithms (MAs), often leading to premature convergence and inefficient use of the remaining evaluation budget. This study introduces MsMA, a novel meta-level strategy that externally monitors MAs to detect stagnation and adaptively partitions computational resources. When stagnation occurs, MsMA divides the optimization run into partitions, restarting the MA for each partition with function evaluations guided by solution history, enhancing efficiency without modifying the MA’s internal logic, unlike algorithm-specific stagnation controls. The experimental results on the CEC’24 benchmark suite, which includes 29 diverse test functions, and on a real-world Load Flow Analysis (LFA) optimization problem demonstrate that MsMA consistently enhances the performance of all tested algorithms. In particular, Self-Adapting Differential Evolution (jDE), Manta Ray Foraging Optimization (MRFO), and the Coral Reefs Optimization Algorithm (CRO) showed significant improvements when paired with MsMA. Although MRFO originally performed poorly on the CEC’24 suite, it achieved the best performance on the LFA problem when used with MsMA. Additionally, the combination of MsMA with Long-Term Memory Assistance (LTMA), a lookup-based approach that eliminates redundant evaluations, resulted in further performance gains and highlighted the potential of layered meta-strategies. This meta-level strategy pairing provides a versatile foundation for the development of stagnation-aware optimization techniques.

Keywords