Algorithm Configuration is still an intricate problem especially in the continuous black box optimization domain. This paper empirically investigates the relationship between continuous problem features (measuring different problem characteristics) and the best parameter configuration of a given stochastic algorithm over a bench of test functions — namely here, the original version of Differential Evolution over the BBOB test bench. This is achieved by learning an empirical performance model from the problem features and the algorithm parameters. This performance model can then be used to compute an empirical optimal parameter configuration from features values. The results show that reasonable performance models can indeed be learned, resulting in a better parameter configuration than a static parameter setting optimized for robustness over the test bench.
CITATION STYLE
Belkhir, N., Dréo, J., Savéant, P., & Schoenauer, M. (2016). Feature based algorithm configuration: A case study with differential evolution. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9921 LNCS, pp. 156–166). Springer Verlag. https://doi.org/10.1007/978-3-319-45823-6_15
Mendeley helps you to discover research relevant for your work.