a recursive decomposition method for large scale ... · sun et al.: recursive decomposition method...

15
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018 647 A Recursive Decomposition Method for Large Scale Continuous Optimization Yuan Sun , Michael Kirley, and Saman K. Halgamuge Abstract—Cooperative co-evolution (CC) is an evolutionary computation framework that can be used to solve high- dimensional optimization problems via a “divide-and-conquer” mechanism. However, the main challenge when using this frame- work lies in problem decomposition. That is, deciding how to allocate decision variables to a particular subproblem, especially interacting decision variables. Existing decomposition methods are typically computationally expensive. In this paper, we propose a new decomposition method, which we call recursive differential grouping (RDG), by considering the interaction between deci- sion variables based on nonlinearity detection. RDG recursively examines the interaction between a selected decision variable and the remaining variables, placing all interacting decision variables into the same subproblem. We use analytical methods to show that RDG can be used to efficiently decompose a problem, with- out explicitly examining all pairwise variable interactions. We evaluated the efficacy of the RDG method using large scale bench- mark optimization problems. Numerical simulation experiments showed that RDG greatly improved the efficiency of problem decomposition in terms of time complexity. Significantly, when RDG was embedded in a CC framework, the optimization results were better than results from seven other decomposition methods. Index Terms—Continuous optimization problem, cooperative co-evolution (CC), decomposition method, large scale global optimization (LSGO). I. I NTRODUCTION L ARGE-SCALE (high-dimensional) optimization problems are ubiquitous in the real-world, occur- ring in domains spanning the sciences, engineering, and multidisciplinary design problems [1]–[3]. Such problems are very difficult to solve when using evolutionary algorithms (EAs), and in many cases cannot be solved when using tradi- tional mathematical approaches. This in part may be attributed Manuscript received August 15, 2016; revised November 30, 2016, March 20, 2017, and July 29, 2017; accepted November 20, 2017. Date of pub- lication November 28, 2017; date of current version September 28, 2018. (Corresponding author: Yuan Sun.) Y. Sun is with the Department of Mechanical Engineering, University of Melbourne, Parkville, VIC 3010, Australia (e-mail: [email protected]). M. Kirley is with the Department of Computing and Information Systems, University of Melbourne, Parkville, VIC 3010, Australia (e-mail: [email protected]). S. K. Halgamuge is with the Research School of Engineering, Australian National University, Canberra, ACT 2601, Australia (e-mail: [email protected]). This paper has supplementary downloadable multimedia material available at http://ieeexplore.ieee.org provided by the authors. Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TEVC.2017.2778089 to the fact that: 1) the search space of an optimization problem grows exponentially as the dimensionality increases [4]; 2) the complexity of an optimization problem usually grows as the dimensionality increases [5]; and 3) the computational cost of using some EAs (e.g., estimation of distribution algorithms) when solving very high-dimensional problems is extremely high [6]. There has been significant recent interest within the evo- lutionary computation community focussed specifically on tackling large scale global optimization (LSGO) problems. This is best illustrated by the introduction of special sessions held at the leading conferences and the special issues published in related journals. The review papers by Mahdavi et al. [7] and LaTorre et al. [8] highlight recent developments in this exciting field. Cooperative co-evolution (CC) [9] has been used with some success when “scaling up” EAs to tackle very high- dimensional search and optimization problems. For example, CC has been applied to large scale continuous [10], com- binatorial [11], constrained [12], multiobjective [13], and dynamic [14] optimization problems. The CC framework divides the LSGO problem into a number of subcomponents, and uses an (several) EA(s) to solve each subcomponent cooperatively. When optimizing each subcomponent, represen- tatives (typically the best subsolutions found) from the other subcomponents are combined with individuals in the opti- mized subcomponent, to form complete candidate solutions that can be evaluated. A number of studies have shown that the problem decomposition can have a significant impact on the performance of a CC framework (e.g., [10] and [15]–[18]). The existing decomposition methods can be classified into two very different approaches (see Section II-B for a brief review): in the manual decomposition method, the structure of the subcomponents is manually designed (e.g., random group- ing [19]). This method does not take the underlying structure of variable interactions (see Section II-B1 for formal defi- nition) into consideration. In the second method, automatic decomposition, the structure of the subcomponents is deter- mined by the identified decision variable interactions (e.g., differential grouping [10]). However, this approach can be computationally expensive—decomposing an n-dimensional problem typically consumes O(n 2 ) function evaluations (FEs). This high computational complexity results in an inappropri- ate allocation of computational resources to the decomposition stage rather than the optimization stage. In this paper, we propose a recursive differential group- ing (RDG) method, which can decompose an n-dimensional 1089-778X c 2017 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Upload: others

Post on 03-Jul-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018 647

A Recursive Decomposition Method for LargeScale Continuous Optimization

Yuan Sun , Michael Kirley, and Saman K. Halgamuge

Abstract—Cooperative co-evolution (CC) is an evolutionarycomputation framework that can be used to solve high-dimensional optimization problems via a “divide-and-conquer”mechanism. However, the main challenge when using this frame-work lies in problem decomposition. That is, deciding how toallocate decision variables to a particular subproblem, especiallyinteracting decision variables. Existing decomposition methodsare typically computationally expensive. In this paper, we proposea new decomposition method, which we call recursive differentialgrouping (RDG), by considering the interaction between deci-sion variables based on nonlinearity detection. RDG recursivelyexamines the interaction between a selected decision variable andthe remaining variables, placing all interacting decision variablesinto the same subproblem. We use analytical methods to showthat RDG can be used to efficiently decompose a problem, with-out explicitly examining all pairwise variable interactions. Weevaluated the efficacy of the RDG method using large scale bench-mark optimization problems. Numerical simulation experimentsshowed that RDG greatly improved the efficiency of problemdecomposition in terms of time complexity. Significantly, whenRDG was embedded in a CC framework, the optimization resultswere better than results from seven other decomposition methods.

Index Terms—Continuous optimization problem, cooperativeco-evolution (CC), decomposition method, large scale globaloptimization (LSGO).

I. INTRODUCTION

LARGE-SCALE (high-dimensional) optimizationproblems are ubiquitous in the real-world, occur-

ring in domains spanning the sciences, engineering, andmultidisciplinary design problems [1]–[3]. Such problems arevery difficult to solve when using evolutionary algorithms(EAs), and in many cases cannot be solved when using tradi-tional mathematical approaches. This in part may be attributed

Manuscript received August 15, 2016; revised November 30, 2016, March20, 2017, and July 29, 2017; accepted November 20, 2017. Date of pub-lication November 28, 2017; date of current version September 28, 2018.(Corresponding author: Yuan Sun.)

Y. Sun is with the Department of Mechanical Engineering,University of Melbourne, Parkville, VIC 3010, Australia (e-mail:[email protected]).

M. Kirley is with the Department of Computing and InformationSystems, University of Melbourne, Parkville, VIC 3010, Australia (e-mail:[email protected]).

S. K. Halgamuge is with the Research School of Engineering,Australian National University, Canberra, ACT 2601, Australia (e-mail:[email protected]).

This paper has supplementary downloadable multimedia material availableat http://ieeexplore.ieee.org provided by the authors.

Color versions of one or more of the figures in this paper are availableonline at http://ieeexplore.ieee.org.

Digital Object Identifier 10.1109/TEVC.2017.2778089

to the fact that: 1) the search space of an optimization problemgrows exponentially as the dimensionality increases [4]; 2) thecomplexity of an optimization problem usually grows as thedimensionality increases [5]; and 3) the computational cost ofusing some EAs (e.g., estimation of distribution algorithms)when solving very high-dimensional problems is extremelyhigh [6].

There has been significant recent interest within the evo-lutionary computation community focussed specifically ontackling large scale global optimization (LSGO) problems.This is best illustrated by the introduction of special sessionsheld at the leading conferences and the special issues publishedin related journals. The review papers by Mahdavi et al. [7]and LaTorre et al. [8] highlight recent developments in thisexciting field.

Cooperative co-evolution (CC) [9] has been used withsome success when “scaling up” EAs to tackle very high-dimensional search and optimization problems. For example,CC has been applied to large scale continuous [10], com-binatorial [11], constrained [12], multiobjective [13], anddynamic [14] optimization problems. The CC frameworkdivides the LSGO problem into a number of subcomponents,and uses an (several) EA(s) to solve each subcomponentcooperatively. When optimizing each subcomponent, represen-tatives (typically the best subsolutions found) from the othersubcomponents are combined with individuals in the opti-mized subcomponent, to form complete candidate solutionsthat can be evaluated. A number of studies have shown thatthe problem decomposition can have a significant impact onthe performance of a CC framework (e.g., [10] and [15]–[18]).

The existing decomposition methods can be classified intotwo very different approaches (see Section II-B for a briefreview): in the manual decomposition method, the structure ofthe subcomponents is manually designed (e.g., random group-ing [19]). This method does not take the underlying structureof variable interactions (see Section II-B1 for formal defi-nition) into consideration. In the second method, automaticdecomposition, the structure of the subcomponents is deter-mined by the identified decision variable interactions (e.g.,differential grouping [10]). However, this approach can becomputationally expensive—decomposing an n-dimensionalproblem typically consumes O(n2) function evaluations (FEs).This high computational complexity results in an inappropri-ate allocation of computational resources to the decompositionstage rather than the optimization stage.

In this paper, we propose a recursive differential group-ing (RDG) method, which can decompose an n-dimensional

1089-778X c© 2017 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Page 2: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

648 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

problem using O(n log(n)) FEs. The RDG method examinesthe interaction between a selected decision variable xi and theremaining decision variables based on nonlinearity detection.If any interaction is identified, the remaining decision vari-ables will be divided into two equally sized groups, and theinteraction between xi and each group is checked. This processis carried out recursively until all of the individual decisionvariables interacting with xi have been identified and have beenplaced into the same subcomponent as xi.

We have evaluated the efficacy of the RDG method usingbenchmark LSGO problems (problems from the special ses-sions on LSGO at CEC’2010 [20] and CEC’2013 [21]).Comprehensive numerical simulations showed that the RDGmethod can decompose the benchmark problems efficientlyin terms of time complexity. We verified that the subcom-ponent groupings dictated by RDG were in fact useful foroptimization. When embedded into a CC framework, theoptimization results generated using the RDG method werestatistically better in terms of solution quality when com-pared against seven other decomposition methods across thebenchmark suite.

The remainder of this paper is organized as follows.Section II describes the state-of-the-art algorithms and decom-position methods within the context of LSGO. Section IIIdescribes the proposed RDG method in detail. Section IVdescribes experiments to evaluate the proposed RDG method.Section V presents and analyzes the experimental results.Section VI concludes this paper and shows future directions.

II. RELATED WORK

A. Algorithms for LSGO

In this section, we briefly describe the state-of-the-arttechniques that can be used to “scale up” EAs for LSGOproblems.

1) Cooperative Co-Evolution: The CC [9] framework tack-les an LSGO problem using a divide-and-conquer strategy.It divides the problem into a number of low-dimensionalsubcomponents that are solved cooperatively. A standardCC algorithm consists of two stages: 1) decomposition and2) optimization.

In the decomposition stage, an optimization problem isdecomposed into several subcomponents. For example, adecomposition of a 6-D optimization problem ( f : R6 → R)could possibly be {(x1, x2), (x3, x4), (x5, x6)}, as shown inFig. 1. When the structure of the underlying decision variableinteractions are considered, this allocation to subcomponentsmay in fact be different. Recent studies have shown that theperformance of a CC algorithm relies heavily on the way theoptimization problem is decomposed [7], [10], [15], [16], [18].See Sections II-B and III for further details.

In the optimization stage, an EA can be used to optimizeeach subcomponent based on a context vector. The contextvector is a complete candidate solution, typically consistingof the best subsolutions from each subcomponent. When opti-mizing the ith subcomponent, the context vector (excludingthe ith subsolution) is used to combine with the individ-uals in the ith subcomponent, to form complete candidate

Fig. 1. Decomposition and optimization of a 6-D problem using a CCalgorithm. The problem has been decomposed into three subcomponents, eachwith two decision variables. When optimizing the second subcomponent, thecontext vector (excluding the second subsolution) is used to combine with theindividuals in the second subcomponent, to form complete candidate solutionsthat can be evaluated.

solutions that can be evaluated, as shown in Fig. 1. It hasbeen found that using only one context vector may be toogreedy [22]. Therefore, the adaptive multicontext CC [22]framework is proposed, which employs more than one con-text vector to co-evolve subcomponents. The original CCframework [9] optimizes each subcomponent in a round-robinfashion. Recently, a contribution-based CC framework [23]has been proposed to efficiently allocate the computationalresources. In each cycle, it selects and optimizes the subcom-ponent that makes the greatest historical contribution to thefitness improvement.

2) Other Techniques: In addition to CC, there are othertechniques that can be used to address the additional chal-lenges inherent in LSGO problems. Representative techniquesinclude the model complexity control [6] and random projec-tion [24] methods for estimation of distribution algorithms;the multiple strategies [25] and generalized opposition-basedlearning [26] methods for differential evolution; the sociallearning [27] and pairwise competition [28] methods for parti-cle swarm optimization; and the multiple trajectory search [29]as well as multiple offspring sampling [30] methods for algo-rithm hybridization. Due to page limits, we cannot describethese techniques in detail.

B. Decomposition Methods

1) Interacting Decision Variables: In this section, we startby presenting a formal definition of “variable interactions,” assome of the decomposition methods discussed below rely onvariable interactions.

In an optimization problem, two decision variables interactif they cannot be optimized independently to find the globaloptimum. It is important to note that the interaction betweengiven decision variables may be complicated. Take the follow-ing objective function as an example:

f (x) := x21 + (x2 − x3)

2 + (x3 − x4)2, x ∈ [−1, 1]4 (1)

where “:=” denotes “defined as.” Decision variable x2 interactswith x3, and x3 interacts with x4. Therefore, x2 and x4 arelinked by x3. If the optimal value of x3 is known, x2 and x4

Page 3: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649

are independent, as they can be optimized separately. However,if x3 needs to be optimized, x2 and x4 will influence eachother. Therefore, x2 conditionally interacts with x4.1 Notethat conditional interaction only exists in overlapping prob-lems e.g., Rosenbrock’s function [21]. The formal definitionsof interaction and conditional interaction are consistent withthe definitions of direct interaction and indirect interactiondescribed in [15] and [31].

Definition 1: Let f : Rn → R be a differentiable function.Decision variables xi and xj interact if a candidate solution x∗exists, such that

∂2 f (x∗)∂xi∂xj

�= 0 (2)

denoted by xi ↔ xj. Decision variables xi and xj conditionallyinteract if for any candidate solution x∗

∂2 f (x∗)∂xi∂xj

= 0 (3)

and a set of decision variables {xk1, . . . , xkt} ⊂ X exists, suchthat xi ↔ xk1 ↔ . . . ↔ xkt ↔ xj. Decision variables xi and xj

are independent if for any candidate solution x∗, (3) holds anda set of decision variables {xk1, . . . , xkt} ⊂ X does not exist,such that xi ↔ xk1 ↔ . . . ↔ xkt ↔ xj.

2) Manual Decomposition: In these methods, the numberof subcomponents and the size of each subcomponent aremanually designed. These methods work well when combinedwith algorithms to solve fully separable problems. However,the performance deteriorates quickly when applied on par-tially separable problems or fully nonseparable problems. Themain reason is that it does not take the underlying structureof variable interactions into consideration.

Probably the first and simplest decomposition is theuni-variable grouping [9] method, which decomposes an n-dimensional problem into n 1-D subcomponents. The uni-variable grouping method improved the performance of agenetic algorithm (GA) when solving benchmark separa-ble problems, however it degraded the GA’s performancewhen solving a benchmark nonseparable problem—theRosenbrock’s function [9]. This performance difference maybe attributed to the fact that the uni-variable grouping methoddecomposes an optimization problem without considering theinteraction between decision variables.

The Sk grouping [32] method is more flexible than theuni-variable grouping method when used to decomposean optimization problem. It decomposes an n-dimensionalproblem into k s-dimensional subcomponents, s < n. TheSk grouping method has been shown to be able to improvethe performance of a particle swarm optimization [32] algo-rithm and a biogeograph-based optimization [33] algorithm.However, like the uni-variable grouping method, the Sk

grouping method does not take variable interactions intoconsideration.

The random grouping (RG) method is proposed within thecontext of a differential evolution CC (DECC) framework [19].

1In [4], the relationships between {x2, x4} and {x1, x4} are both definedas additively separable. In this paper, we use conditional interaction todifferentiate these two relationships.

It randomly assigns decision variables to predetermined num-ber of subcomponents before each evolutionary cycle. TheRG method has been successfully applied to improve theperformance of a particle swarm optimization [34] algorithmand an artificial bee colony [35] algorithm. However, it hasbeen shown that the probability of assigning more than twointeracting decision variables into one subcomponent usingthe RG method is low [36]. Another limitation of RG is therequirement of setting an appropriate subcomponent size. Toaddress this issue, the multilevel CC [37] algorithm takes thesubcomponent size as a parameter, and selects an appropriatesubcomponent size according to the historical performance.

A more sophisticated method—delta grouping [38] identi-fies variable interactions based on the averaged difference in acertain decision variable across the whole population. It gener-ally outperforms the RG method when incorporated with theDECC framework to solve the CEC’2010 benchmark prob-lems [38]. However on benchmark problems with more thanone nonseparable subcomponent, the performance of the deltagrouping method is low [10].

The k-means grouping [39] method uses a k-means clus-tering algorithm to construct decision variable groups. Thedecision variables with similar effects on the fitness valueare placed into the same subcomponent. The subcomponentwith the greatest contribution to the fitness value will beoptimized with more iterations (FEs). The idea is similar tothe contribution-based CC [23], [40] framework. Unlike mostdecomposition methods which group decision variables basedon variable interactions, the k-means grouping method groupsdecision variables based on their contribution to the fitnessimprovement. Therefore, it is specifically tailored for problemswith unbalanced subcomponents [4], [39].

3) Automatic Decomposition: In these methods, theinteracting decision variables are identified and automaticallyplaced into the same subcomponent. It is important to note thatautomatic decomposition considers the underlying variableinteraction structure encapsulated within the search landscape.The decomposition method proposed in Section III falls intothis category.

A representative automatic decomposition method—CCwith variable interaction learning (CCVIL) [41]—identifiesthe pairwise interaction between decision variables by thenonmonotonicity detection. If the monotonicity of the fitnessfunction with respect to xi does not change for different valueof xj, xi and xj are independent. Otherwise, decision variablesxi and xj interact. The rationale behind the CCVIL method isconsistent with the linkage identification by nonmonotonicitydetection [42] method. The CCVIL method is more accuratethan most of the manual decomposition methods when iden-tifying variable interactions. However, it still cannot obtainacceptable results when used to decompose some benchmarkproblems [10].

The statistical variable interdependence learning (SL) [43]method identifies variable interactions based on the nonmono-tonicity detection as well. Unlike the CCVIL method, the SLmethod detects the monotonicity relationship between xi andxj multitimes. The probability (pij) of the observation of non-monotonicity is calculated. If the probability pij is greater

Page 4: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

650 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

than a given threshold, decision variables xi and xj interact.The main issue of the SL method is the high computationalcomplexity. The number of FEs needed to decompose an n-dimensional optimization problem is 4mn2, where m is thenumber of nonmonotonicity detection conducted for each pairof decision variables.

To address the high computational complexity of the SLmethod, a fast variable interdependence searching [44] methodis proposed. It identifies interactions between two subsetsof decision variables, instead of two decision variables, bynonmonotonicity detection. Therefore, it speeds up the decom-position process. The computational cost to decompose ann-dimensional problem can be reduced to 4mn log(n) in theworst case.

The differential grouping (DG) [10] method identifies vari-able interactions by detecting the fitness changes when per-turbing the decision variables. If the fitness change inducedby perturbing decision variable xi varies for different value ofxj, xi and xj interact. The rationale behind the DG methodis consistent with the linkage identification by nonlinear-ity check [45] method. The DG method outperformed theCCVIL method when used to decompose the CEC’2010benchmark problems [10]. However, it has been shown not tobe able to completely identify interacting decision variables inoverlapping problems [15], [46].

Subsequently, the extended DG (XDG) [15] method wasproposed to address this issue, by placing all the linked(interacting and conditionally interacting) decision variablesinto one subcomponent. It employs the same technique as DGto identify interacting decision variables. Then the overlappingbetween subcomponents are checked to identify conditionalinteractions. However, the number of FEs used by XDGto decompose an n-dimensional problem is usually aroundn2. The high complexity of the XDG method results inan inappropriate allocation of computational budget betweendecomposition and optimization, and prevents it from beingapplied to solve even higher dimensional problems.

The computational cost in the decomposition stage can bereduced to (n2+3n+2)/2 by using the global DG (GDG) [46]method. It employs the same technique as DG to identifythe pairwise interactions between decision variables. The vari-able interaction matrix is calculated, which is regarded asthe adjacency matrix of a graph. Then the depth-first searchor breadth-first search can be used to identify the connectedcomponents. Note that both the interacting and conditionallyinteracting decision variables will also be placed into oneconnected component (subcomponent).

Recently, it has been shown that the minimal number ofFEs used to identify the complete variable interaction matrixbased on DG is (n2 +n+2)/2 [47]. However, it may not needthe entire variable interaction matrix to identify the connectedcomponents (subcomponents). For example, if decision vari-able x1 interacts with x2 and x3, the interaction between x2and x3 needs not to be checked, as they belong to the sameconnected component.

The fast interdependency identification (FII) [48] methodcan further improve the efficiency of problem decomposi-tion by avoiding the need to identify the complete variable

interaction matrix. FII first identifies the separable decisionvariables by examining the interaction between one deci-sion variable and the other variables. Then the interactionbetween nonseparable decision variables is examined, and allthe linked (connected) decision variables are placed into thesame subcomponent. The FII method is efficient when used todecompose benchmark problems with a large portion of sep-arable decision variables. However on benchmark problemswith conditional variable interactions, the number of FEs usedby FII may still be in the magnitude of n2 (�(n2)). In the nextsection, we will propose an efficient and robust method thatcan decompose any n-dimensional problem using less than6n log2(n) FEs.

III. RECURSIVE DIFFERENTIAL GROUPING

In this section, the proposed decomposition method—RDG—is described in detail. Then, the computational com-plexity of the RDG method is presented.

Notation: Let X be the set of decision variables {x1, . . . , xn}and UX be the set of unit vectors in the decision space Rn.Let X1 be a subset of decision variables X1 ⊂ X and UX1 be asubset of UX such that any unit vector u = (u1, . . . , un) ∈ UX1 ,we have

ui = 0, if xi /∈ X1. (4)

Directional Derivative: Let f : Rn → R be a differentiablefunction, and u = (u1, . . . , un) be a vector from UX . Thedirectional derivative of f in the direction u, denoted Du f (x),is given by

Du f (x) =n∑

i=1

∂ f (x)

∂xiui. (5)

Proposition 1: Let f : Rn → R be a differentiable function;X1 ⊂ X and X2 ⊂ X be two mutually exclusive subsets ofdecision variables: X1 ∩X2 = ∅. If there exist two unit vectorsu1 ∈ UX1 and u2 ∈ UX2 , and a candidate solution x∗ in thedecision space such that

Du1Du2 f(x∗) �= 0 (6)

there is some interaction between decision variables in X1and X2.

Proof: Without loss of generality, we assume that X1 ={x1,1, . . . , x1,p}, X2 = {x2,1, . . . , x2,q}, where p and q are thenumber of decision variables in X1 and X2, respectively; u1 =(u1

1, . . . , u1n) and u2 = (u2

1, . . . , u2n). According to directional

derivative

Du1Du2 f (x) =n∑

i=1

n∑

j=1

∂2 f (x)

∂xi∂xju1

i u2j . (7)

As u1 and u2 are two unit vectors from UX1 and UX2 ,respectively, we can obtain that

u1i = 0, if xi /∈ X1 (8)

u2j = 0, if xj /∈ X2. (9)

Page 5: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 651

Therefore,

Du1Du2 f (x) =p∑

i=1

q∑

j=1

∂2 f (x)

∂x1,i∂x2,ju1

1,iu22,j. (10)

If (6) holdsp∑

i=1

q∑

j=1

∂2 f (x∗)∂x1,i∂x2,j

u11,iu

22,j �= 0. (11)

Therefore, there exists at least one pair of (i, j), such that

∂2 f (x∗)∂x1,i∂x2,j

�= 0. (12)

Based on Definition 1, at least one pair of decision variablesx1,i ∈ X1 and x2,j ∈ X2 interact.

Corollary 1: Let f : Rn → R be an objective function;

X1 ⊂ X and X2 ⊂ X be two mutually exclusive subsets ofdecision variables: X1 ∩X2 = ∅. If there exist two unit vectorsu1 ∈ UX1 and u2 ∈ UX2 , two real numbers l1, l2 > 0, and acandidate solution x∗ in the decision space, such that

f(x∗ + l1u1 + l2u2

) − f(x∗ + l2u2

) �= f(x∗ + l1u1

) − f(x∗)

(13)

there is some interaction between decision variables in X1and X2.

Proof: With Proposition 1, we only need to prove thefollowing statement.

Statement 1: If there exist two unit vectors u1 ∈ UX1 andu2 ∈ UX2 , two real numbers l1, l2 > 0, and a candidate solu-tion x∗ in the decision space, such that (13) holds, then (6) istrue.

It is equivalent to prove its contraposition.Statement 2: If for any two unit vectors u1 ∈ UX1 and u2 ∈

UX2 , and for any candidate solution x∗ in the decision space,the following condition holds:

Du1Du2 f(x∗) = 0 (14)

then

f(x∗ + l1u1 + l2u2

) − f(x∗ + l2u2

) = f(x∗ + l1u1

) − f(x∗)

(15)

for any l1, l2 > 0.In order to prove Statement 2, we first introduce line

integral.Line Integral: Let L be a curve with end points A and B in

the decision space Rn and the arc length of L be l. Let C beany point on L and the coordinate of C (x) can be uniquelydetermined by the length of the arc AC (s) : x = x(s), s ∈ [0, l].The integral of a function g : Rn → R along the curve L isgiven by

Lg(x)ds =

∫ l

0g(x(s))ds. (16)

Let A2 (x∗) be any point in Rn, and B2 be x∗ + l2u2, whereu2 is any vector in UX2 and l2 is any positive real number. LetC2 be any point on the segment A2B2. Therefore, the lengthof the segment A2B2 is l2, and the coordinate of C2 (x) can be

uniquely determined by the length of the segment A2C2 (s2):x(s2) = x∗ + s2u2, s2 ∈ [0, l2]. If (14) holds for any candidatesolution in the decision space, then

Du1Du2 f (x) = 0. (17)

As Du1Du2 f (x) = Du2Du1 f (x), by integrating both sidesof (17) along the segment A2B2, we can obtain that

∫ l2

0Du1Du2 f (x)ds2 =

∫ l2

0Du2Du1 f (x)ds2 = 0. (18)

As∫ l2

0Du2

(Du1 f (x(s2)))

ds2 = Du1 f (x(s2))

∣∣∣s2=l2s2=0 (19)

thus,

Du1 f (x(s2))

∣∣∣s2=l2s2=0 = 0 (20)

and

Du1 f(x∗ + l2u2

) − Du1 f(x∗) = 0. (21)

As A2 (x∗) is any point in Rn, therefore

Du1 f (x + l2u2) = Du1 f (x). (22)

Let A1 (x∗) be any point in Rn, and B1 be x∗ + l1u1, where u1is any vector in UX1 and l1 is any positive real number. LetC1 be any point on the segment A1B1. Therefore, the lengthof the segment A1B1 is l1, and the coordinate of C1 (x) canbe uniquely determined by the length of the segment A1C1(s1) : x(s1) = x∗ + s1u1, s1 ∈ [0, l1]. Similarly, by integratingboth sides of (22) along the segment A1B1, we can obtain

∫ l1

0Du1 f (x(s1) + l2u2) ds1 =

∫ l1

0Du1 f (x(s1)) ds1. (23)

Therefore,

f(x∗ + l1u1 + l2u2

)−f(x∗ + l2u2

) = f(x∗ + l1u1

) − f(x∗).

(24)

Thus, Statement 2 is proved, and Statement 1 and Corollary 1are true.

With Corollary 1, the interaction between two subsets ofdecision variables (X1 and X2) can be identified using thefollowing procedures.

1) Set all the decision variables to the lower bounds (lb)of the search space (xl,l).

2) Perturb the decision variables X1 of xl,l from the lowerbounds to the upper bounds (ub), denoted by xu,l.

3) Calculate the fitness difference (δ1) between xl,l and xu,l.4) Perturb the decision variables X2 of xl,l and xu,l from

the lower bounds to the middle between the lowerbounds and upper bounds, denoted by xl,m and xu,m,respectively.

5) Calculate the fitness difference (δ2) between xl,m andxu,m.

6) If the difference between δ1 and δ2 is greater than athreshold ε, there is some interaction between X1 and X2.

The two subscripts of x denote the values of X1 and X2, respec-tively: “l” means lower bounds, “u” means upper bounds, and

Page 6: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

652 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

Fig. 2. Variable interaction structure and the RDG decomposition of theobjective function given in (26). The notation xi ↔ xj denotes that decisionvariable xi interacts with xj.

“m” means the middle between the lower bounds and upperbounds. The threshold ε is estimated based on the magnitudeof the objective space [46]

ε = α · min{| f (x1)|, . . . , | f (xk)|} (25)

where x1, . . . , xk are k randomly generated candidate solu-tions, and α is the control coefficient [46].

Based on Corollary 1, we propose the RDG (Algorithm 1)method to efficiently decompose an optimization problem. Thedecomposition by the RDG method considers the underly-ing structure of variable interactions. Taking the followingobjective function as an example:

f (x) := x21 + (x2 − x3)

2 + (x3 − x4)2

+ (x5 − x6)2, x ∈ [−1, 1]6 (26)

the decision variables (x2, x3, x4) interact, as well as (x5, x6).Therefore, the decomposition by the RDG method is{(x1), (x2, x3, x4), (x5, x6)}, as shown in Fig. 2.

The inputs to the RDG method are the fitness function (f ),the upper bounds (ub), the lower bounds (lb), and the thresh-old (ε) which is estimated using (25). The outputs are theseparable variable group (seps) and the nonseparable variablegroups (nonseps). The seps contains one group of all sepa-rable decision variables. The nonseps contains several groupsof nonseparable decision variables. Each group of decisionvariables will form a subcomponent.

The RDG method begins by identifying the interactionbetween the first decision variable x1 and the remainingdecision variables. If no interaction is detected, x1 will beplaced in the separable decision variable group, and the algo-rithm will move on to the next decision variable x2. If anyinteraction is detected, the remaining decision variables willbe divided into two (nearly) equally sized groups G1 and G2.Then the interaction between x1 and G1, x1 and G2 will beidentified, respectively. This process is recursively conducteduntil all the individual decision variables that interact withx1 are identified and placed in the decision variable subsetX1 with x1.

Then, the RDG method examines the interaction between X1and the remaining decision variables (excluding the decisionvariables in X1) to identify the individual decision variablesthat conditionally interact with x1 (linked by other deci-sion variables). If any interaction is identified, the interactingdecision variables will be placed into X1. This process isrepeated until no interaction can be further detected between

Algorithm 1 RDGRequire: f , ub, lb, ε

1: Initialize seps and nonseps as empty groups2: Set all decision variables to the lower bounds: xl,l = lb3: Calculate the fitness: yl,l = f (xl,l)

4: Assign the first variable x1 to the variable subset X15: Assign the rest of variables to the variable subset X26: while X2 is not empty do7: [X∗

1 ] = INTERACT(X1, X2, xl,l, yl,l, ε)8: if X∗

1 is the same with X1 then9: if X1 contains one decision variable then

10: Add X1 to seps11: else12: Add X1 to nonseps13: end if14: Empty X1 and X∗

115: Assign the first variable of X2 to X116: Delete the first variable in X217: else18: X1 = X∗

119: Delete the variables of X1 from X220: end if21: end while22: return seps and nonseps

1: function INTERACT(X1, X2, xl,l, yl,l, ε)2: xu,l=xl,l; xu,l(X1) = ub(X1) //Set X1 to the ub3: Calculate the fitness change: δ1 = yl,l − f (xu,l)4: xl,m = xl,l; xl,m(X2) = (

lb(X2) + ub(X2))/2

5: xu,m = xu,l; xu,m(X2) = (lb(X2) + ub(X2)

)/2

6: Calculate the fitness change: δ2 = f (xl,m) − f (xu,m)

7: if |δ1 − δ2| > ε then8: if X2 contains one variable then9: X1 = X1 ∪ X2

10: else11: Divide X2 into equally-sized groups G1, G212: [X1

1]=INTERACT(X1, G1, xl,l, yl,l, ε)13: [X2

1]=INTERACT(X1, G2, xl,l, yl,l, ε)14: [X1]= X1

1 ∪ X21

15: end if16: end if17: return X118: end function

X1 and the remaining decision variables (exclusive X1). Thedecision variables in X1 will be placed in a nonseparablegroup.

The RDG method moves on to the next decision vari-able that has not been grouped (xi). The interaction betweenxi and the remaining decision variables will be exam-ined, and both the interacting and conditionally interacting(linked) decision variables will be placed into one groupwith xi. This process is repeated until all of the decisionvariables are grouped. It returns the separable (seps) andthe nonseparable (nonseps) decision variable groups as theoutputs.

Page 7: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 653

The computational complexity of the RDG method whenused to decompose an n-dimensional problem is O(n log(n)),which is analyzed as follows.

1) When decomposing an n-dimensional fully separableproblem, the computational complexity of the RDGmethod is �(n) in terms of the number of FEs. For eachdecision variable, three FEs are used to determine itsseparability. Therefore, totally about 3n FEs are needed.

2) When decomposing an n-dimensional fully nonsepara-ble problem with one subcomponent, the computationalcomplexity of the RDG method is �(n). When group-ing the n interacting decision variables, the function“INTERACT” is executed about

∑ki=0(n/2i) times, where

k = log2(n)

k∑

i=0

n

2i= n

(2 −

(1

2

)k)

< 2n. (27)

It consumes three FEs each time the function INTERACT

is executed. Therefore, totally about 6n FEs are used.3) When decomposing an n-dimensional partially separable

problem with n/m subcomponents, the computationalcomplexity of the RDG method is �(n log(n)), wheren is the dimensionality and m is the number of deci-sion variables in each subcomponent. When groupingm interacting decision variables into one subcompo-nent, the function INTERACT is executed less than2m×log2(n) times, each time consuming three FEs. Thenumber of subcomponents is n/m. Therefore, totally lessthan 3×2m× log2(n)×n/m = 6n log2(n) FEs are used.

4) When decomposing an n-dimensional partially separa-ble problem with an m-dimensional nonseparable sub-component, the computational complexity of the RDGmethod is O(max{n, m log(n)}). The RDG algorithmconsumes about 3(n−m) FEs to identify the n−m sep-arable decision variables, and consumes less than 6m ×log2(n) FEs to identify the m interacting decision vari-ables. Therefore, totally less than 3(n−m)+6m×log2(n)

FEs are used.5) When decomposing an n-dimensional overlapping

problem (e.g., Rosenbrock’s function [21]), the compu-tational complexity of the RDG method is �(n log(n)).Starting from x1, it consumes about 3 × 2 × 2 log2(n) =12 log2(n) FEs to identify the two decision variables (xp

and xq) that interact with x1. Then it also consumes about12 log2(n) FEs to identify the two decision variablesthat interact with (xp, x1, xq). Therefore, totally aboutn/2 × 12 log2(n) = 6n log2(n) FEs are used.

IV. EXPERIMENTAL METHODOLOGY

In this section, comprehensive numerical experiments aredesigned to evaluate the proposed RDG method. Two researchquestions guide the experimental design to evaluate the effi-cacy of the proposed RDG method.Q1: Can the proposed RDG method decompose the

CEC’2010 and CEC’2013 benchmark problems moreefficiently when compared against other well-knowndecomposition methods.

TABLE IPARAMETER SETTINGS FOR ALL THE DECOMPOSITION METHODS USED

IN THE EXPERIMENTS. DG2 IS A PARAMETER-FREE METHOD

Q2: Can the proposed RDG method outperform other well-known decomposition methods when embedded into aCC framework to solve the CEC’2010 and CEC’2013benchmark problems.

To answer Q1, the proposed RDG method was used todecompose the CEC’2010 [20] and CEC’2013 [21] bench-mark problems.2 Two metrics were employed to evaluate theperformance of a decomposition method: 1) the number ofFEs used to decompose the problem and 2) the percentageof interacting decision variables that are correctly grouped,defined as follows.

Decomposition Accuracy: Let G = {g1, . . . , gm} denote thegroups of interacting decision variables in a problem f , andG = {g1, . . . , gn} denote the groups of interacting decisionvariables that are identified by a decomposition method. LetGi = {gi,1, . . . , gi,m} denote the ith permutation of G, where1 ≤ i ≤ m!, and Gj = {gj,1, . . . , gj,n} denote the jth permu-tation of G, where 1 ≤ j ≤ n!. The decomposition accuracy(DA) of the decomposition method on f is defined as

DA =max

i,j

{min{m,n}∑

k=1|gi,k ∩ gj,k|

}

m∑i=1

|gi|(28)

where |gi| denotes the number of decision variables in gi.The performance of the RDG method was then compared

to the GDG [46], XDG [15], DG [10], as well as two recentlypublished methods—DG2 [47] and FII [48]. The parame-ter settings for all the decomposition methods used in theexperiments are shown in Table I. The threshold values (ε)estimated by the RDG (or GDG) method for each problemwere recorded. Note that the RDG and GDG methods use thesame approach (25) to estimate the threshold values.

To answer Q2, the proposed RDG method was embeddedinto the DECC [19]/CMAESCC [46] framework to solve theCEC’2010 and CEC’2013 benchmark problems. The DECC isthe most widely used CC framework, which employs a variantof differential evolution—SaNSDE [49]—to solve each sub-component cooperatively. The CMAESCC framework uses thewell-known CMA-ES [50] algorithm to solve each subcom-ponent. It performs well when used to solve the CEC’2010and CEC’2013 benchmark problems. The parameter settings

2The MATLAB implementation of the RDG method can be accessed fromthe following link: https://bitbucket.org/yuans/rdg.

Page 8: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

654 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

for the DECC and CMAESCC frameworks were consistentwith the original papers. The maximum number of FEs wasset to 3 × 106, divided between the decomposition stage andoptimization stage. For each benchmark problem, the median,mean, and standard deviation of the best solutions found by theDECC/CMAESCC-RDG algorithm based on 25 independentruns were recorded. The performance of the RDG method wascompared against the performance of the XDG, GDG, DG,DG2, FII, as well as two manual decomposition methods—D(delta grouping [38]) and RG [19] methods, when embeddedin each CC framework.

The performance of DECC/CMAESCC-RDG was alsocompared to the performance of two state-of-the-art hybridalgorithms—MOS [30] and MA-SW-Chains [51] with defaultparameter settings. The MOS algorithm evaluates the con-stituent algorithms in each generation, and the better per-formed constituent algorithm will be used to generate moreoffspring. MOS achieved the best performance in the 2011 spe-cial issue of the Soft Computing journal. The MA-SW-Chainsalgorithm assigns to each individual a local search intensitythat depends on its features, by chaining different local searchapplications. MA-SW-Chains achieved the best performancein the CEC 2010 special session and competition on LSGO.

The Kruskal–Wallis nonparametric one-way ANOVAtest [52] with 95% confidence interval was used to determinewhether the performance of at least one algorithm was signif-icantly different from the others. Then a series of Wilcoxonrank-sum tests (significance level α = 0.05) with Holm p-value correction [52] was conducted in a pairwise fashion tofind the better performing algorithm.

V. EXPERIMENTAL RESULTS

Comprehensive experimental results are presented and dis-cussed in this section. Section V-A presents the decompositioncomparison between the RDG method and five other meth-ods, thus addressing Q1. Section V-B presents the optimizationcomparison between the RDG method and seven other meth-ods when embedded into the DECC/CMAESCC framework tosolve the benchmark problems, thus addressing Q2.

A. Decomposition Comparison

Table II lists the decomposition results of the RDG, GDG,XDG, and DG methods on the CEC’2010 and CEC’2013benchmark problems. The parameter settings for the fourdecomposition methods are consistent with Table I. In Table II,“DA” represents the decomposition accuracy—the percentageof interacting decision variables that correctly grouped; “FEs”represents the number of FEs used in the decomposition stage;and “ε” represents the threshold used to identify interactionsbetween decision variables. Note that the threshold values usedby RDG are the same as those used by GDG. The entrieswith the best DA achieved using the smallest number of FEsare highlighted in bold. Different categories of benchmarkproblems are divided by lines.

The RDG and GDG methods obtain nearly the same DAacross all the benchmark problems. The reason for this is thatRDG uses the same approach (25) as GDG to estimate the

threshold values. However, RDG is much more efficient thanGDG in terms of FEs used. Note that the number of FEs usedby GDG to decompose an n-dimensional problem is fixed:(n2 + 3n + 2)/2.

The first three problems (f1 − f3) from each benchmark suiteare fully separable. Therefore, DA is not applicable to theseproblems. On f1 and f2 (CEC’2010 and CEC’2013), the RDGmethod successfully identifies all the decision variables as sep-arable, using a small number of FEs. However on f3, the RDGand GDG methods identify all the separable variables as non-separable, and place them into one subcomponent. The reasonfor this is that the threshold value is underestimated by theRDG and GDG methods on f3.

The CEC’2013 f13 and f14 are benchmark problems withoverlapping (conforming or conflicting) subcomponents. It isnot clear yet what is the best approach to decompose theseproblems [4], [21]. The RDG, GDG, and XDG methods placeall the overlapped subcomponents into one group. On the otherbenchmark problems, where the subcomponents are indepen-dent with each other, the “ideal” decomposition can possiblybe achieved (see [4], [10], [20], [21] for more information).Note that the 100% DA in Table II corresponds to the idealdecomposition.

On the CEC’2010 partially separable problems (f4 − f18),the RDG method consistently achieves the best results whencompared against GDG, XDG, and DG. The RDG, GDG, andXDG methods achieve the ideal decomposition on all of thesebenchmark problems. However, the number of FEs used by theGDG and XDG methods is usually several magnitude largerthan that used by the RDG method. The DG method performswell on the benchmark problems without conditional variableinteractions. On f19, the DG method uses the smallest numberof FEs to decompose the problem. However on problems withconditional variable interactions (overlapping problems, e.g.,CEC’2010 f13, f18, and f20), the DA of the DG method is low.The reason for this is that DG is unable to completely identifyvariable interactions in overlapping problems [15], [46].

On the CEC’2013 partially separable problems (f4−f11), theRDG method achieves the best results on 4 out of 8 bench-mark problems. We observe that it is generally more difficultto identify the variable interactions in the CEC’2013 than theCEC’2010 benchmark problems. None of the four methodscan perfectly decompose the CEC’2013 f8, f10, and f11 prob-lems. On CEC’2013 f11, the threshold value is underestimatedby the RDG and GDG methods (25), resulting in placingall the decision variables into a single nonseparable group.While on CEC’2013 f8 and f10, the threshold value is overes-timated, resulting in some omissions of variable interactionsbeing identified.

The threshold values estimated by the GDG and RDG meth-ods vary significantly across the CEC’2010 and CEC’2013benchmark problems. Therefore, it is very difficult to find asingle threshold value to accurately identify variable interac-tions across all the problems. Although the threshold valueε = 10−1 works well for the XDG method on the CEC’2010benchmark problems, it fails on some of the CEC’2013 bench-mark problems, e.g., f5 − f8. On f5, the XDG algorithm (withε = 10−1) identifies all the interacting decision variables as

Page 9: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 655

TABLE IIEXPERIMENTAL RESULTS OF THE PROPOSED RDG METHOD WHEN USED TO DECOMPOSE THE CEC’2010 AND CEC’2013 BENCHMARK PROBLEMS.

DA IS THE DECOMPOSITION ACCURACY; FES IS THE FES USED; ε IS THE THRESHOLD. NOTE THAT THE THRESHOLD VALUES USED BY RDG ARE THE

SAME AS THOSE USED BY GDG. THE PERFORMANCE OF THE RDG METHOD IS COMPARED WITH THE PERFORMANCES OF THE GDG, XDG,AND DG METHODS. THE ENTRIES WITH THE BEST DA ACHIEVED USING THE LOWEST FES ARE HIGHLIGHTED IN BOLD

separable. The reason for this is that the threshold value 10−1

is too large compared with the computational error, which isequal to 8.03×10−5 estimated by the RDG and GDG methods.When using the estimated computational error as the thresh-old value, the XDG method achieves equal DA with the RDGmethod.

To further show the efficacy of the RDG method, we com-pare the performance of RDG against two recently publishedmethods—DG2 and FII. The average number of FEs used byRDG to decompose the CEC’2010 and CEC’2013 benchmarkproblems is 1.47 × 104, which is less than that used by theDG2 and FII methods: 4.95 × 105 and 4.94 × 104, respec-tively. The detailed decomposition results of the DG2 and FIImethods are presented in the supplementary material.

The DG2 method uses a fixed number of FEs to decom-pose an n-dimensional problem: (n2 + n + 2)/2, which hasbeen shown to be the lower bound of identifying the com-plete variable interaction matrix. With the complete variableinteraction matrix being identified, it is possible to generate

an effective decomposition for the problems with overlappingsubcomponents, e.g., CEC’2013 f13 and f14 [47]. However, theexisting automatic decomposition methods place all the linkeddecision variables into one subcomponent.

DG2 is a parametric-free method, which automatically esti-mates the threshold values in the decomposition process. TheDA of DG2 on the CEC’2010 and CEC’2013 benchmark prob-lems is high. It achieves 100% DA for the CEC’2013 f10 andf11 problems. The variable interactions in these two problemsare difficult for the other decomposition methods to identify.However on CEC’2013 f7 and f8, the DA of DG2 is less thanthat of RDG.

The FII method performs well when used to decompose thebenchmark problems with a large portion of separable decisionvariables. For example on CEC’2010 f4, where there are 950separable and 50 nonseparable decision variables, the numberof FEs used by FII (3.69 × 103) is slightly less than that usedby RDG (4.20 × 103). However, on some benchmark prob-lems especially those with conditional variable interactions,

Page 10: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

656 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

Fig. 3. Benchmark problems from the CEC’2010 test suite on which theRDG and FII methods generate significant different results (the differencebetween the number of FEs used is greater than 104).

TABLE IIIEXTENDED CEC’2010 f12, f15, AND f18 PROBLEMS. FOR EACH PROBLEM,

THE NUMBER OF DECISION VARIABLES IN EACH NONSEPARABLE

SUBCOMPONENT IS FIXED TO 50, WHICH IS CONSISTENT

WITH THE ORIGINAL BENCHMARK SET [20]

e.g., CEC’2010 f18 and f20, RDG is much more efficient thanFII, as shown in Fig. 3.

In fact, the number of FEs used by the FII method to decom-pose the CEC’2010 f18 and f20 problems is in �(n2). It hasbeen shown that FII uses 3n + knn + k FEs when decompos-ing an n-dimensional problem with equally sized nonseparablesubcomponents, where nn is the number of nonseparabledecision variables, and k is the number of nonseparable sub-components [48]. As CEC’2010 f18 and f20 are Rosenbrock’sfunctions, nn is equal to n and each decision variable interactswith at most two other decision variables. Therefore, the totalnumber of FEs used by FII is around 3n+n2/3+n/3 ∈ �(n2).

On CEC’2010 f6, the number of FEs used by RDG (5.00×104) is greater than that used by FII (3.05 × 103). The reasonfor this is that the threshold estimated by the RDG method(ε = 2.13 × 10−5) is too small, resulting in identifying someseparable decision variables as nonseparable. If RDG employsthe same threshold with FII (ε = 0.01), the FEs used by RDGwill decrease to 5.06×103. However, if FII employs the samethreshold with RDG, the FEs used by FII will increase to3.12 × 105.

To test the scalability of the RDG and FII methods, weextend some of the CEC’2010 benchmark problems from 1000to 5000 dimensions (see Table III for details). When tested on

TABLE IVAVERAGE RANKING OF EACH DECOMPOSITION METHOD WHEN

EMBEDDED INTO THE DECC OR CMAESCC FRAMEWORK TO

SOLVE THE CEC’2010 AND CEC’2013 BENCHMARK

PROBLEMS. THE RDG METHOD CONSISTENTLY

ACHIEVES THE SMALLEST AVERAGE RANKING

the extended benchmark problems, we observe that the FEsused by the FII method increases more quickly than that usedby the RDG method as the dimensionality increases, as shownin Fig. 4.

B. Optimization Comparison

The performances of the RDG, GDG, XDG, DG, DG2,FII, D (delta grouping), and RG methods when embedded inthe DECC/CMAESCC framework to solve the CEC’2010 andCEC’2013 benchmark problems are presented in Figs. 5 and 6,respectively. On each benchmark problem, the eight meth-ods are ranked from 1 to 8 (as labeled on the concentriccircles in the radar charts) based on the results from 25independent runs. The average ranking of each decomposi-tion method across all the benchmark problems is presentedin Table IV. The detailed optimization results from DECC-RDG and CMAESCC-RDG are presented in Tables V and VI.The results from the other seven methods are placed in thesupplementary material due to page limits.

When embedded into the DECC/CMAESCC framework,the RDG method achieves the best solution quality whenused to solve 16/15 out of 23 partially separable benchmarkproblems (CEC’2010 f4 − f18 and CEC’2013 f4 − f11). Itobtains the smallest average ranking across all the benchmarkproblems investigated. The RDG method generally uses thesmallest number of FEs in the decomposition stage, assigningmore computational resources to optimize the problems, whencompared against the other automatic decomposition methods.

On the fully separable and some fully nonseparable prob-lems, the RDG method is outperformed by the two manualdecomposition methods—D (delta grouping) and RG, as RDGdoes not actually perform any decomposition for these prob-lems. However, the performance of the D (delta grouping) andRG methods deteriorates quickly on the partially separableproblems.

On some benchmark problems with separable decision vari-ables, e.g., CEC’2010 f10, the GDG method generates bettersolution quality than RDG. The reason for this is that GDGfurther divides the separable decision variables into severalsubcomponents. Interestingly, the GDG method achieves thefirst place when embedded into CMAESCC, however, the lastplace when embedded into DECC, to solve the CEC’2010 f9.

The DG method performs poorly on overlapping bench-mark problems (e.g., CEC’2010 f8, f13, f18, and f20). Thereason for this is that the DG method can not completely iden-tify interaction between decision variables in an overlappingproblem. Once all the variable interactions are identified and

Page 11: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 657

(a) (b) (c)

Fig. 4. Number of FEs used by RDG and FII methods when used to decompose the extended benchmark problems (CEC’2010 f12, f15, and f18) withdimensionality equal to 1000, 2000, 3000, 4000, and 5000.

(a) (b)

Fig. 5. Radar chart of the performance of RDG, GDG, XDG, DG, DG2, FII, D (delta grouping), and RG when embedded in the DECC framework to solvethe CEC’2010 and CEC’2013 benchmark problems. On each benchmark problem, the eight methods are ranked from 1 to 8 (as labeled on the concentriccircles) based on the results from 25 independent runs [Wilcoxon rank-sum tests (α = 0.05) with Holm p-value correction].

all the linked decision variables are placed into one subcom-ponent, the solution quality can be greatly improved by severalmagnitudes.

The DG2 method generates the best solution qualitywhen embedded into the CMAESCC framework to solvethe CEC’2013 f11 problem. The reason for this is thatDG2 achieves 100% accuracy when used to decompose thisproblem, which is higher than RDG. However on the otherbenchmark problems, the RDG method generally obtainsequally well or statistically better solution quality than DG2.

The FII method performs well across the benchmark prob-lems investigated. It achieves the second place according to theaverage ranking. However, on the benchmark problems wherethe decomposition by FII is less efficient, e.g., CEC’2010 f8,the RDG method can generate significantly better solutionquality than FII.

The DECC-D (with delta grouping) algorithm achievesmuch better results than the other DECC-based algorithmswhen used to solve the CEC’2010 f3 and f11 benchmark prob-lems. An interesting observation is that both f3 and f11 areAckley’s functions [21]. Moreover, the DECC-D (with deltagrouping) algorithm also performs well when used to solve theother Ackley’s functions: CEC’2010 f6, f16 and CEC’2013 f3,f6, f10. So far, we do not know the reason why the DECC-D(with delta grouping) algorithm performs well on the Ackley’sfunctions.

The convergence curves of the eight decomposition meth-ods when embedded into the DECC/CMAESCC frameworkto solve the CEC’2010 f18 problem are shown in Fig. 7. TheRDG method uses the smallest number of FEs in the decompo-sition stage, therefore can generate better solution quality whencompared against the other automatic decomposition methods.

Page 12: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

658 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

(a) (b)

Fig. 6. Radar chart of the performance of RDG, GDG, XDG, DG, DG2, FII, D (delta grouping), and RG when embedded in the CMAESCC framework tosolve the CEC’2010 and CEC’2013 benchmark problems. On each benchmark problem, the eight methods are ranked from 1 to 8 (as labeled on the concentriccircles) based on the results from 25 independent runs [Wilcoxon rank-sum tests (α = 0.05) with Holm p-value correction].

(a) (b)

Fitn

ess V

alue

Fitn

ess V

alue

Fig. 7. Convergence curves of the RDG, GDG, XDG, DG, DG2, FII, D (delta grouping), and RG methods when embedded into the DECC and CMAESCCframeworks to solve the CEC’2010 f18 problem. The horizontal axis represents the number of FEs used in the evolutionary process. The vertical axis representsthe median of the best fitness found.

In the next phase of the experimental study, we com-pare the performance of DECC-RDG and CMAESCC-RDGagainst the performance of two state-of-the-art algorithms—MOS and MA-SW-Chains. The experimental results of eachalgorithm when used to solve the CEC’2010 and CEC’2013benchmark problems are presented in Tables V and VI,respectively.

The CMAESCC-RDG algorithm achieves the best solu-tion quality on 22 out of 35 benchmark problems whencompared against DECC-RDG, MOS, and MA-SW-Chains.It does not perform well on the fully separable prob-lems (CEC’2010 f1–f3 and CEC’2013 f1–f3). However on

partially separable and fully nonseparable problems, it gen-erally achieves comparable or statistically better solutionquality than the other algorithms (e.g., on CEC’2010 f4–f9and f11–f19).

The DECC-RDG algorithm achieves the best solution qual-ity when used to solve the CEC’2010 f16 problem. On the other34 benchmark problems, the DECC-RDG is outperformed bythe CMAESCC-RDG. It may indicate that the subcomponentoptimizer used by the CMAESCC framework is more effectivethan that used by the DECC framework.

The MOS algorithm achieves the best results when usedto solve the fully separable problems (CEC’2010 f1–f3, and

Page 13: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 659

TABLE VRESULTS OF THE DECC-RDG, CMAESCC-RDG, MOS, AND

MA-SW-CHAINS ALGORITHMS WHEN USED TO SOLVE THE

CEC’2010 BENCHMARK PROBLEMS. THE BEST

PERFORMANCES ARE HIGHLIGHTED IN BOLD

[WILCOXON RANK-SUM TESTS (α = 0.05)WITH HOLM p-VALUE CORRECTION]

CEC’2013 f1–f3). However, on partially separable problemsand fully nonseparable problems, it is generally outperformedby the CMAESCC-RDG algorithm.

TABLE VIRESULTS OF THE DECC-RDG, CMAESCC-RDG, MOS, AND

MA-SW-CHAINS ALGORITHMS WHEN USED TO SOLVE THE

CEC’2013 BENCHMARK PROBLEMS. THE BEST PERFORMANCES

ARE HIGHLIGHTED IN BOLD [WILCOXON RANK-SUM TESTS

(α = 0.05) WITH HOLM p-VALUE CORRECTION]

The MA-SW-Chains algorithm performs well on theCEC’2013 benchmark problems. It achieves the best resultswhen used to solve the CEC’2013 f5, f6, f9, and f10 problems.However, on the other partially separable and fully nonsepa-rable benchmark problems, it is consistently outperformed bythe CMAESCC-RDG algorithm. In some cases, the best solu-tion found by the CMAESCC-RDG algorithm is much betterthan that found by the MA-SW-Chains algorithm.

Page 14: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

660 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

VI. CONCLUSION

In this paper, we have investigated the influence of problemdecomposition on the performance of CC algorithms whenused to solve LSGO problems. A robust decompositionmethod—RDG—was proposed, which can decompose ann-dimensional problem using O(n log(n)) FEs based on a mea-sure of nonlinearity between decision variables. Significantly,RDG outperformed seven other decomposition methods whenembedded into the DECC/CMAESCC framework and testedacross a suite of benchmark LSGO problems. When com-pared against two other state-of-the-art hybrid algorithms, theCMAESCC-RDG algorithm achieved statistically significantlybetter results.

In future work, we plan to apply the CMAESCC-RDG algo-rithm to solve real-world LSGO problems. Another directionworth pursuing is focused on the modification of the RDGmethod to the combinatorial and multiobjective spaces.

ACKNOWLEDGMENT

The authors would like to thank M. A. Muñoz and W. Wangfor their valuable comments.

REFERENCES

[1] P. Benner, “Solving large-scale control problems,” IEEE Control Syst.,vol. 24, no. 1, pp. 44–59, Feb. 2004.

[2] S. Shan and G. G. Wang, “Survey of modeling and optimization strate-gies to solve high-dimensional design problems with computationally-expensive black-box functions,” Struct. Multidisciplinary Optim.,vol. 41, no. 2, pp. 219–241, 2010.

[3] H. Liu and L. Yu, “Toward integrating feature selection algorithms forclassification and clustering,” IEEE Trans. Knowl. Data Eng., vol. 17,no. 4, pp. 491–502, Apr. 2005.

[4] M. N. Omidvar, X. Li, and K. Tang, “Designing benchmark problemsfor large-scale continuous optimization,” Inf. Sci., vol. 316, pp. 419–436,Sep. 2015.

[5] T. Weise, R. Chiong, and K. Tang, “Evolutionary optimization: Pitfallsand booby traps,” J. Comput. Sci. Technol., vol. 27, no. 5, pp. 907–936,2012.

[6] W. Dong, T. Chen, P. Tino, and X. Yao, “Scaling up estimation ofdistribution algorithms for continuous optimization,” IEEE Trans. Evol.Comput., vol. 17, no. 6, pp. 797–822, Dec. 2013.

[7] S. Mahdavi, M. E. Shiri, and S. Rahnamayan, “Metaheuristics in large-scale global continues optimization: A survey,” Inf. Sci., vol. 295,pp. 407–428, Feb. 2015.

[8] A. LaTorre, S. Muelas, and J.-M. Peña, “A comprehensive compari-son of large scale global optimizers,” Inf. Sci., vol. 316, pp. 517–549,Sep. 2015.

[9] M. A. Potter and K. A. De Jong, “A cooperative coevolutionary approachto function optimization,” in Parallel Problem Solving From NaturePPSN III. Heidelberg, Germany: Springer, 1994, pp. 249–257.

[10] M. N. Omidvar, X. Li, Y. Mei, and X. Yao, “Cooperative co-evolutionwith differential grouping for large scale optimization,” IEEE Trans.Evol. Comput., vol. 18, no. 3, pp. 378–393, Jun. 2014.

[11] Y. Mei, X. Li, and X. Yao, “Cooperative coevolution with route distancegrouping for large-scale capacitated arc routing problems,” IEEE Trans.Evol. Comput., vol. 18, no. 3, pp. 435–449, Jun. 2014.

[12] E. Sayed, D. Essam, R. Sarker, and S. Elsayed, “Decomposition-basedevolutionary algorithm for large scale constrained problems,” Inf. Sci.,vol. 316, pp. 457–486, Sep. 2015.

[13] K. C. Tan, Y. J. Yang, and C. K. Goh, “A distributed cooperative coevo-lutionary algorithm for multiobjective optimization,” IEEE Trans. Evol.Comput., vol. 10, no. 5, pp. 527–549, Oct. 2006.

[14] C.-K. Goh and K. C. Tan, “A competitive-cooperative coevolutionaryparadigm for dynamic multiobjective optimization,” IEEE Trans. Evol.Comput., vol. 13, no. 1, pp. 103–127, Feb. 2009.

[15] Y. Sun, M. Kirley, and S. K. Halgamuge, “Extended differential group-ing for large scale global optimization with direct and indirect variableinteractions,” in Proc. Genet. Evol. Comput. Conf., Madrid, Spain, 2015,pp. 313–320.

[16] W. Chen and K. Tang, “Impact of problem decomposition on cooperativecoevolution,” in Proc. IEEE Congr. Evol. Comput., Cancún, Mexico,2013, pp. 733–740.

[17] H. Liu, Y. Wang, X. Liu, and S. Guan, “Empirical study of effect ofgrouping strategies for large scale optimization,” in Proc. Int. Joint Conf.Neural Netw. (IJCNN), Vancouver, BC, Canada, 2016, pp. 3433–3439.

[18] B. Kazimipour, M. N. Omidvar, X. Li, and A. Qin, “A sensitivity analysisof contribution-based cooperative co-evolutionary algorithms,” in Proc.IEEE Congr. Evol. Comput. (CEC), Sendai, Japan, 2015, pp. 417–424.

[19] Z. Yang, K. Tang, and X. Yao, “Large scale evolutionaryoptimization using cooperative coevolution,” Inf. Sci., vol. 178, no. 15,pp. 2985–2999, 2008.

[20] K. Tang, X. Yao, P. Suganthan, Z. Yang, and T. Weise “Benchmarkfunctions for the CEC’2010 special session and competition on largescale global optimization,” Nat. Inspired Comput. Appl. Lab., USTC,Hefei, China, Tech. Rep., 2010, pp. 1–23.

[21] X. Li, K. Tang, M. N. Omidvar, Z. Yang, and K. Qin, “Benchmarkfunctions for the CEC 2013 special session and competition on large-scale global optimization,” Gene, vol. 7, no. 33, p. 8, 2013.

[22] R.-L. Tang, Z. Wu, and Y.-J. Fang, “Adaptive multi-context coopera-tively coevolving particle swarm optimization for large-scale problems,”Soft Comput., vol. 21, no. 16, pp. 4735–4754, 2016.

[23] M. Yang et al., “Efficient resource allocation in cooperative co-evolutionfor large-scale global optimization,” IEEE Trans. Evol. Comput., vol. 21,no. 4, pp. 493–505, Aug. 2017.

[24] A. Kabán, J. Bootkrajang, and R. J. Durrant, “Toward large-scale contin-uous EDA: A random matrix theory perspective,” Evol. Comput., vol. 24,no. 2, pp. 255–291, Jun. 2016.

[25] J. Brest and M. S. Maucec, “Self-adaptive differential evolution algo-rithm using population size reduction and three strategies,” Soft Comput.,vol. 15, no. 11, pp. 2157–2174, 2011.

[26] H. Wang, Z. Wu, and S. Rahnamayan, “Enhanced opposition-based dif-ferential evolution for solving high-dimensional continuous optimizationproblems,” Soft Comput., vol. 15, no. 11, pp. 2127–2140, 2011.

[27] R. Cheng and Y. Jin, “A social learning particle swarm optimizationalgorithm for scalable optimization,” Inf. Sci., vol. 291, pp. 43–60,Jan. 2015.

[28] R. Cheng and Y. Jin, “A competitive swarm optimizer for largescale optimization,” IEEE Trans. Cybern., vol. 45, no. 2, pp. 191–204,Feb. 2015.

[29] L.-Y. Tseng and C. Chen, “Multiple trajectory search for large scaleglobal optimization,” in Proc. IEEE Congr. Evol. Comput. CEC (IEEEWorld Congr. Comput. Intell.), Hong Kong, 2008, pp. 3052–3059.

[30] A. LaTorre, S. Muelas, and J.-M. Peña, “A MOS-based dynamicmemetic differential evolution algorithm for continuous optimization:A scalability test,” Soft Comput., vol. 15, no. 11, pp. 2187–2199, 2011.

[31] Y. Sun, M. Kirley, and S. K. Halgamuge, “Quantifying vari-able interactions in continuous optimization problems,” IEEETrans. Evol. Comput., vol. 21, no. 2, pp. 249–264, Apr. 2017,doi: 10.1109/TEVC.2016.2599164.

[32] F. Van den Bergh and A. P. Engelbrecht, “A cooperative approach toparticle swarm optimization,” IEEE Trans. Evol. Comput., vol. 8, no. 3,pp. 225–239, Jun. 2004.

[33] X.-W. Zheng, D.-J. Lu, X.-G. Wang, and H. Liu, “A cooperative coevo-lutionary biogeography-based optimizer,” Appl. Intell., vol. 43, no. 1,pp. 95–111, 2015.

[34] X. Li and X. Yao, “Cooperatively coevolving particle swarms forlarge scale optimization,” IEEE Trans. Evol. Comput., vol. 16, no. 2,pp. 210–224, Apr. 2012.

[35] Y. Ren and Y. Wu, “An efficient algorithm for high-dimensional functionoptimization,” Soft Comput., vol. 17, no. 6, pp. 995–1004, 2013.

[36] M. N. Omidvar, X. Li, Z. Yang, and X. Yao, “Cooperative co-evolutionfor large scale optimization through more frequent random grouping,”in Proc. IEEE Congr. Evol. Comput. (CEC), Barcelona, Spain, 2010,pp. 1–8.

[37] Z. Yang, K. Tang, and X. Yao, “Multilevel cooperative coevolution forlarge scale optimization,” in Proc. IEEE Congr. Evol. Comput. CEC(IEEE World Congr. Comput. Intell.), Hong Kong, 2008, pp. 1663–1670.

[38] M. N. Omidvar, X. Li, and X. Yao, “Cooperative co-evolution with deltagrouping for large scale non-separable function optimization,” in Proc.IEEE Congr. Evol. Comput. (CEC), Barcelona, Spain, 2010, pp. 1–8.

Page 15: A Recursive Decomposition Method for Large Scale ... · SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 649 are independent, as they can be optimized

SUN et al.: RECURSIVE DECOMPOSITION METHOD FOR LARGE SCALE CONTINUOUS OPTIMIZATION 661

[39] S. Mahdavi, S. Rahnamayan, and M. E. Shiri, “Multilevel frameworkfor large-scale global optimization,” Soft Comput., vol. 21, no. 14,pp. 4111–4140, 2017.

[40] M. N. Omidvar, X. Li, and X. Yao, “Smart use of computationalresources based on contribution for cooperative co-evolutionary algo-rithms,” in Proc. 13th Annu. Conf. Genet. Evol. Comput., Dublin,Ireland, 2011, pp. 1115–1122.

[41] W. Chen, T. Weise, Z. Yang, and K. Tang, “Large-scale globaloptimization using cooperative coevolution with variable interactionlearning,” in Parallel Problem Solving From Nature, PPSN XI.Heidelberg, Germany: Springer, 2010, pp. 300–309.

[42] M. Munetomo and D. E. Goldberg, “Linkage identification by non-monotonicity detection for overlapping functions,” Evol. Comput., vol. 7,no. 4, pp. 377–398, 1999.

[43] L. Sun, S. Yoshida, X. Cheng, and Y. Liang, “A cooperative particleswarm optimizer with statistical variable interdependence learning,” Inf.Sci., vol. 186, no. 1, pp. 20–39, 2012.

[44] H. Ge, L. Sun, X. Yang, S. Yoshida, and Y. Liang, “Cooperativedifferential evolution with fast variable interdependence learning andcross-cluster mutation,” Appl. Soft Comput., vol. 36, pp. 300–314,Nov. 2015.

[45] M. Tezuka, M. Munetomo, and K. Akama, “Linkage identificationby nonlinearity check for real-coded genetic algorithms,” in Geneticand Evolutionary Computation–GECCO 2004. Heidelberg, Germany:Springer, 2004, pp. 222–233.

[46] Y. Mei, M. N. Omidvar, X. Li, and X. Yao, “A competitivedivide-and-conquer algorithm for unconstrained large-scale black-boxoptimization,” ACM Trans. Math. Softw., vol. 42, no. 2, p. 13,2016.

[47] M. N. Omidvar, M. Yang, Y. Mei, X. Li, and X. Yao, “DG2: Afaster and more accurate differential grouping for large-scale black-boxoptimization,” IEEE Trans. Evol. Comput., vol. 21, no. 6, pp. 929–942,Dec. 2017.

[48] X.-M. Hu, F.-L. He, W.-N. Chen, and J. Zhang, “Cooperationcoevolution with fast interdependency identification for large scaleoptimization,” Inf. Sci., vol. 381, pp. 142–160, Mar. 2017.

[49] Z. Yang, K. Tang, and X. Yao, “Self-adaptive differential evolu-tion with neighborhood search,” in Proc. IEEE Congr. Evol. Comput.CEC (IEEE World Congr. Comput. Intell.), Hong Kong, 2008,pp. 1110–1116.

[50] N. Hansen, “The CMA evolution strategy: A tutorial,” Mach. Learn.Optim. Group, Inria, France, Rep., 2011.

[51] D. Molina, M. Lozano, and F. Herrera, “MA-SW-Chains: Memetic algo-rithm based on local search chains for large scale continuous globaloptimization,” in Proc. IEEE Congr. Evol. Comput. (CEC), Barcelona,Spain, 2010, pp. 1–8.

[52] D. J. Sheskin, Handbook Of Parametric and Nonparametric StatisticalProcedures. Boca Raton, FL, USA: CRC Press, 2003.

Yuan Sun received the B.Sc. degree in theoreti-cal and applied mechanics from Peking University,Beijing, China, in 2013. He is currently pursuingthe Ph.D. degree with the University of Melbourne,Parkville, VIC, Australia.

His doctoral research focuses on the analy-sis of interaction between decision variables inoptimization and classification problems. His cur-rent research interests include exploratory landscapeanalysis, large scale optimization, multiobjectiveoptimization, and feature selection.

Michael Kirley received the B.Ed. degree fromDeakin University, Burwood, VIC, Australia, in1998 and the Ph.D. degree from Charles SturtUniversity, Bathurst, NSW, Australia, in 2003.

He is currently an Associate Professor with theDepartment of Computing and Information Systems,University of Melbourne, Parkville, VIC, Australia.His current research interests include theory andapplication of evolutionary computation, evolution-ary game theory, multiagent system, and complexsystems science. He has published over 100 papersin the above areas.

Saman K. Halgamuge received the B.Sc.Eng.degree in electronic and telecommunicationengineering from the University of Moratuwa,Moratuwa, Sri Lanka, and the Dr.-Ing and Dipl.-Ingdegrees in electrical engineering (data engineering)from TU Darmstadt, Darmstadt, Germany, in 1990and 1995, respectively.

He is a Professor and the Director of theResearch School of Engineering, AustralianNational University, Canberra, ACT, Australia.He held appointments as a Professor with the

Department of Mechanical Engineering and the Associate Dean Internationalwith the Melbourne School of Engineering, University of Melbourne,Parkville, VIC, Australia. He also holds the Professor V. K. SamaranayakeEndowed Visiting Chair with the University of Colombo, Colombo, Sri Lanka.His current research interests include data engineering, optimization, smartgrids, mechatronics, and bioinformatics. He has published over 250 papersand graduated 32 Ph.D. students in the above areas and obtained substantialresearch grants from the ARC, NHMRS, and industry.