Research Repository

Adaptive Operator Selection With Bandits for a Multiobjective Evolutionary Algorithm Based on Decomposition

Li, Ke and Fialho, Alvaro and Kwong, Sam and Zhang, Qingfu (2014) 'Adaptive Operator Selection With Bandits for a Multiobjective Evolutionary Algorithm Based on Decomposition.' IEEE Transactions on Evolutionary Computation, 18 (1). pp. 114-130. ISSN 1089-778X

Full text not available from this repository.


Adaptive operator selection (AOS) is used to determine the application rates of different operators in an online manner based on their recent performances within an optimization process. This paper proposes a bandit-based AOS method, fitness-rate-rank-based multiarmed bandit (FRRMAB). In order to track the dynamics of the search process, it uses a sliding window to record the recent fitness improvement rates achieved by the operators, while employing a decaying mechanism to increase the selection probability of the best operator. Not much work has been done on AOS in multiobjective evolutionary computation since it is very difficult to measure the fitness improvements quantitatively in most Pareto-dominance-based multiobjective evolutionary algorithms. Multiobjective evolutionary algorithm based on decomposition (MOEA/D) decomposes a multiobjective optimization problem into a number of scalar optimization subproblems and optimizes them simultaneously. Thus, it is natural and feasible to use AOS in MOEA/D. We investigate several important issues in using FRRMAB in MOEA/D. Our experimental results demonstrate that FRRMAB is robust and its operator selection is reasonable. Comparison experiments also indicate that FRRMAB can significantly improve the performance of MOEA/D. © 2013 IEEE.

Item Type: Article
Subjects: Q Science > QA Mathematics
Q Science > QA Mathematics > QA75 Electronic computers. Computer science
Divisions: Faculty of Science and Health
Faculty of Science and Health > Computer Science and Electronic Engineering, School of
SWORD Depositor: Elements
Depositing User: Elements
Date Deposited: 12 Nov 2014 20:41
Last Modified: 15 Jan 2022 00:47

Actions (login required)

View Item View Item