ML-Methods for Identifying Features of Glob. Optimization
ML-Methods for Identifying Features of Glob. Optimization
Bilaterale Ausschreibung: Russland
Disciplines
Computer Sciences (70%); Mathematics (30%)
Keywords
-
Evolutionary Compuation,
Automatic Adaptation,
Meteheuristic Optimization
Evolutionary algorithms are nature-inspired learning, search and optimization methods that take the natural evolutionary process of a species as a model to adapt in the best possible way to environmental conditions. Methodical developments in combination with constantly increasing computing resources lead to the fact that more and more complex and higher-dimensional tasks can be solved with evolutionary algorithms. However, almost all of the currently considered optimization and machine learning tasks are stationary which means that the optimization or modeling goal does not change during an algorithm run. This project deals with different methodological approaches to move into the non-stationary domain. Although so far applied almost exclusively to stationary tasks, evolutionary approaches are an ideal starting point, since natural evolution itself is highly non-stationary. A species that loses its adaptability to new environmental conditions by adapting too greedily to currently prevailing conditions would become extinct just as well at this would happen if environmental conditions change too rapidly for a species to adapt. A sufficiently high presence of genetic diversity is therefore just as important in nature in terms of adaptability as it is for non-stationary optimization and modeling. Involving biological expertise, the Austrian project group will research implicit methods such as self- adaptive process extensions that maintain adaptability through a constant support of new genetic diversity without having to abandon what has already been learned. Age-layered population structures represent a promising methodological starting point here. So far, approaches of this kind have been mainly used to reduce premature convergence by a constant supply of new genetic diversity via the young age layers. In the course of the present project, these approaches will be extended in the direction of permanent adaptation to changing goals and framework conditions, and concepts will be explored as to how and which strategic method information can be transferred from the older to the younger age layers. The cooperating group from the Siberian State University of Science and Technology will go in the same direction with explicit approaches that change algorithm parameters or switch between algorithms at runtime based on fitness indicators.
This project explored how optimization algorithms can learn to adapt to changing conditions over time, much like how evolution helps species survive in shifting environments. Traditional optimization methods assume that a problem, such as planning delivery routes or scheduling production, stays fixed until solved. In reality, conditions change constantly: new jobs arrive, routes take longer than expected, or machines break down. The project focused on how evolutionary algorithms, which imitate natural selection and survival of the fittest, can be extended to handle such dynamic situations. A central theme was how to prevent a kind of dynamic collapse that can occur when an optimizer becomes too narrow or greedy for short-term gains. The team explored several ways to build resilience. Diversity-preserving measures like layered population structures let an optimizer maintain a steady mix of new and old solutions and avoid overfocusing. For problems where solutions vary in complexity (such as symbolic regression models), continuous pruning kept mathematical expressions compact and interpretable. And in practical tests like dynamic warehouse stacking, the researchers showed that short-sighted strategies could cause the system to grind to a halt. This highlighted the need to balance short-term efficiency with long-term stability and to include warning systems that alert human operators to trends the optimizer cannot avoid. As the project progressed, attention turned to the interaction between optimization algorithms and the machine learning models used to estimate solution quality. When an algorithm repeatedly avoids certain operating conditions, the predictive models may learn that these conditions are poor choices, even if they are not. This creates a feedback loop that can reinforce bias and limit performance. The team developed methods to detect, monitor, and mitigate such effects, ensuring that learning-based evaluations remain accurate and balanced over time. These findings also pointed to a broader challenge: even when evaluations are correct, optimization often involves trade-offs between competing objectives. Finally, the project addressed these trade-offs directly. Improving one measure, such as cost or speed, often worsens another, such as quality or sustainability. The researchers developed mathematical tools for comparing and predicting the progress of algorithms under such conflicting demands, including refined acquisition functions (selection criteria that weigh predicted qualities against uncertainties) and probability models that estimate the chances of further improvement. In summary, the project combined principles from evolutionary optimization, mathematical modelling, and artificial intelligence to improve how optimization systems respond to change. The developed methods help algorithms analyse the dynamics of problems and adapt their behaviour accordingly.
- FH Oberösterreich - 100%
- Eugene S. Semenkin, Reshetnev Siberian State University of Science and Technology - Russia
Research Output
- 7 Citations
- 14 Publications
- 6 Datasets & models
-
2024
Title A Functional Analysis Approach to Symbolic Regression DOI 10.1145/3638529.3654079 Type Conference Proceeding Abstract Author Antonov K Pages 859-867 -
2024
Title Probability Distribution of Hypervolume Improvement in Bi-objective Bayesian Optimization Type Conference Proceeding Abstract Author Wang H. Conference 41st International Conference on Machine Learning Link Publication -
2025
Title Age-Layer-Population-Structure withSelf-adaptation inOptimization; In: Computer Aided Systems Theory - EUROCAST 2024 - 19th International Conference, Las Palmas de Gran Canaria, Spain, February 25 - March 1, 2024, Revised Selected Papers, Part III DOI 10.1007/978-3-031-83885-9_1 Type Book Chapter Publisher Springer Nature Switzerland -
2025
Title Diversity Management inEvolutionary Dynamic Optimization; In: Computer Aided Systems Theory - EUROCAST 2024 - 19th International Conference, Las Palmas de Gran Canaria, Spain, February 25 - March 1, 2024, Revised Selected Papers, Part I DOI 10.1007/978-3-031-82949-9_13 Type Book Chapter Publisher Springer Nature Switzerland -
2025
Title Continuous Pruning for Symbolic Regression DOI 10.1145/3712255.3734287 Type Conference Proceeding Abstract Author Affenzeller M Pages 2572-2579 -
2024
Title Efficient Global Optimization for Dynamic Problems DOI 10.46354/i3m.2024.emss.018 Type Conference Proceeding Abstract -
2022
Title Dynamic Vehicle Routing with Time-Linkage: From Problem States to Algorithm Performance DOI 10.1007/978-3-031-25312-6_8 Type Book Chapter Author Werth B Publisher Springer Nature Pages 69-77 -
2023
Title A New Acquisition Function for Multi-objective Bayesian Optimization: Correlated Probability of Improvement DOI 10.1145/3583133.3596374 Type Conference Proceeding Abstract Author Chen K Pages 2308-2317 -
2023
Title Walking through the Quadratic Assignment-Instance Space: Algorithm Performance and Landscape Measures DOI 10.1145/3583133.3596325 Type Conference Proceeding Abstract Author Karder J Pages 2108-2114 -
2023
Title Surrogate-assisted Multi-objective Optimization viaGenetic Programming Based Symbolic Regression; In: Evolutionary Multi-Criterion Optimization - 12th International Conference, EMO 2023, Leiden, The Netherlands, March 20-24, 2023, Proceedings DOI 10.1007/978-3-031-27250-9_13 Type Book Chapter Publisher Springer Nature Switzerland -
2023
Title Gradients of Acquisition Functions for Bi-objective Bayesian optimization DOI 10.1109/icnc-fskd59587.2023.10280812 Type Conference Proceeding Abstract Author Liu S Pages 1-9 -
2023
Title Population diversity and inheritance in genetic programming for symbolic regression DOI 10.1007/s11047-022-09934-x Type Journal Article Author Burlacu B Journal Natural Computing -
2023
Title Applying Learning and Self-Adaptation to Dynamic Scheduling DOI 10.3390/app14010049 Type Journal Article Author Karder J Journal Applied Sciences -
2022
Title A parallel technique for multi-objective Bayesian global optimization: Using a batch selection of probability of improvement DOI 10.1016/j.swevo.2022.101183 Type Journal Article Author Yang K Journal Swarm and Evolutionary Computation Pages 101183 Link Publication
-
2025
Link
Title Werth et al. MDPI 2023 DOI 10.5281/zenodo.17531080 Type Database/Collection of data Public Access Link Link -
2025
Link
Title Werth et al. GECCO 2025 DOI 10.5281/zenodo.17530979 Type Database/Collection of data Public Access Link Link -
2025
Link
Title Werth et al. EMSS 2024 DOI 10.5281/zenodo.17530841 Type Database/Collection of data Public Access Link Link -
2025
Link
Title Werth et al. Eurocast 2022 DOI 10.5281/zenodo.17530900 Type Database/Collection of data Public Access Link Link -
2025
Link
Title Werth et al. Eurocast 2024 DOI 10.5281/zenodo.17530929 Type Database/Collection of data Public Access Link Link -
2025
Link
Title Werth et al. Gecco 2023 DOI 10.5281/zenodo.17530949 Type Database/Collection of data Public Access Link Link