Call for Papers


Special Issue on “Benchmarking Multi and Many Objective Evolutionary Algorithms on Challenging Test Problems”


Swarm and Evolutionary Computation Journal, Elsevier

(2017 Impact Factors: 3.89 for 2 years and 7.7 for 5 years)


Multi-objective optimization problems (MOPs) are commonly encountered in real-world applications. Multi-objective evolutionary algorithms (MOEAs) are effective in solving MOPs with a few objectives. In recent years, it was observed that MOEAs face difficulties in solving MOPs with four or more objectives. These problems are known as Many-objective Optimization Problems (MaOPs). Challenges faced by population-based algorithms when solving MaOPs include the inability of dominance based MOEAs to converge to the Pareto front with good diversity, high computational complexity in the computation of performance indicators, and the difficulties in decision making, visualization, and understanding the relationships between objectives and articulated preferences. To tackle these issues, numerous many objective evolutionary algorithms (MaOEAs) have been developed and evaluated on standard benchmark problems.


The objective of this special issue is to evaluate MOEAs as well as the recently developed MaOEAs on newly designed challenging MaOPs presented in the following technical report:


Hui Li, Kalyanmoy Deb, Qingfu Zhang and P N Suganthan, “Challenging Novel Many and Multi-Objective Bound Constrained Benchmark Problems, Technical Report, 2017. (TR updated on 14th May 2018. Codes updated on 14th May 2018. You can do test runs and give us feedback, if you find any problem)


It is expected that to solve these challenging problems effectively, the state of the art algorithms will have to be improved. Hence, while including the novel problems also in their evaluation studies, researchers are invited to present their original works on the following multi and many objective optimization related issues (but not limited to):


Algorithm design issues such as selection rules, reproduction, mating restriction, and so on.

Performance indicators

Objective reduction

Visualization techniques

Preference Articulation

Decision making methods

Hybridized algorithms

Development of further challenging Benchmark problems

Many-objective real-world optimization problems

Model learning

Estimating knee, nadir points

Constraint handling methods

EAs for MCDM



The manuscripts should be prepared according to the “Guide for Authors” section of the journal found at: and submission should be done through the journal’s submission website: by selecting “MOEAs” and also clearly indicating the full title of this special issue “Benchmarking Multi and Many Objective Evolutionary Algorithms on Challenging Test Problems” in comments to the Editor-in-Chief. Each submitted paper will be reviewed by expert reviewers. Submission of a paper will imply that it contains original unpublished work and is not being submitted for publication elsewhere.


Important dates (tentative)

Initial Submission: 1st July 2018

First Notification: 1st November 2018

Resubmission: 31st December 2018

Second Notification: 1st March 2019

Final Submission: 1st April 2019

Final Notification: 30th April 2019

Publication: 2019



Guest Editors:


Hui Li

Xian Jiatong University, China.

Kalyanmoy Deb

Michigan State University, East Lansing, MI 48824, USA

Qingfu Zhang

City University, Hong Kong