Loading [MathJax]/jax/element/mml/optable/GeneralPunctuation.js
A journal of IEEE and CAA , publishes high-quality papers in English on original theoretical/experimental research and development in all areas of automation
Zhiming Lv, Linqing Wang, Zhongyang Han, Jun Zhao and Wei Wang, "Surrogate-Assisted Particle Swarm Optimization Algorithm With Pareto Active Learning for Expensive Multi-Objective Optimization," IEEE/CAA J. Autom. Sinica, vol. 6, no. 3, pp. 838-849, May 2019. doi: 10.1109/JAS.2019.1911450
Citation: Zhiming Lv, Linqing Wang, Zhongyang Han, Jun Zhao and Wei Wang, "Surrogate-Assisted Particle Swarm Optimization Algorithm With Pareto Active Learning for Expensive Multi-Objective Optimization," IEEE/CAA J. Autom. Sinica, vol. 6, no. 3, pp. 838-849, May 2019. doi: 10.1109/JAS.2019.1911450

Surrogate-Assisted Particle Swarm Optimization Algorithm With Pareto Active Learning for Expensive Multi-Objective Optimization

doi: 10.1109/JAS.2019.1911450
Funds:  This work was supported by the National Natural Sciences Foundation of China (61603069, 61533005, 61522304, U1560102) and the National Key Research and Development Program of China (2017YFA0700300)
More Information
  • For multi-objective optimization problems, particle swarm optimization (PSO) algorithm generally needs a large number of fitness evaluations to obtain the Pareto optimal solutions. However, it will become substantially time-consuming when handling computationally expensive fitness functions. In order to save the computational cost, a surrogate-assisted PSO with Pareto active learning is proposed. In real physical space (the objective functions are computationally expensive), PSO is used as an optimizer, and its optimization results are used to construct the surrogate models. In virtual space, objective functions are replaced by the cheaper surrogate models, PSO is viewed as a sampler to produce the candidate solutions. To enhance the quality of candidate solutions, a hybrid mutation sampling method based on the simulated evolution is proposed, which combines the advantage of fast convergence of PSO and implements mutation to increase diversity. Furthermore, ε-Pareto active learning (ε-PAL) method is employed to pre-select candidate solutions to guide PSO in the real physical space. However, little work has considered the method of determining parameter ε. Therefore, a greedy search method is presented to determine the value of ε where the number of active sampling is employed as the evaluation criteria of classification cost. Experimental studies involving application on a number of benchmark test problems and parameter determination for multi-input multi-output least squares support vector machines (MLSSVM) are given, in which the results demonstrate promising performance of the proposed algorithm compared with other representative multi-objective particle swarm optimization (MOPSO) algorithms.

     

  • MULTI-OBJECTIVE optimization problems are co mmonly encountered in real world [1], e.g., drug design [2], aerodynamic configuration optimization [3] and hyper-parameter optimization for machine learning models [4]. However, the analytic objective or constraints functions in the above cases are difficult to express accurately. Thus, it is a challenge to implement the optimization only based on data collected from physical experiments, real events, or complex numerical simulations [5]. In these cases, it is essential to obtain the desired outcome within a small number of evaluations. Generally, the multi-objective optimization problem can be generalized as follows:

    miny=f(x)=(f1(x),f2(x),,fm(x)) (1)
    s.t.xS (2)

    where xRn represents a vector of n decision variables; SRn indicates the feasible region which is typically specified as a set of constraints on the decision variables; fRm is made of m objective functions which need to be jointly minimized; y is an m-dimensional objective vector.

    Considering the high convergence speed and simple implementation, multi-objective particle swarm optimization (MOPSO) attracts great attention which gives rise to a series of variants in the existing literature [6]. Generally, they can be classified into two categories. One is based on the Pareto sorting strategy [7]–[9], and the other employs the decomposition strategy [10], [11]. Although the variants of MOPSO algorithm above have been proved to be with the satisfactory convergence, they require thousands of fitness evaluations to obtain satisfied solutions. As a result, applying MOPSO on computationally expensive problems becomes impractical when a single exact fitness evaluation exhibits a high computational cost. To enhance the performance of most traditional PSOs in solving multi-objective optimization problems, a number of approaches have been proposed [12].

    Surrogate-assisted PSO algorithms (SAPSOs) have be found very effective in handling computationally expensive multi-objective optimization problems [13], [14]. Some employ single or multiple surrogates to approximate the fitness function. For instance, a study presented in [15] took use of a response surface model (RSM) rather than finite-element analysis. The Kriging model is applied to gain error compensation function between the approximated values obtained by RSM and the actual measured values in [16]. While others typically estimates the degree of uncertainty to choose certified candidates for re-evaluation, e.g., the Kriging model was combined with MOPSO method to predict the expected improvements of each particle in [17]. Reference [18] captured the uncertainty with the hypervolume to pre-select promising particles for re-evaluation. All of the above studies have shown the effectiveness of SAPSO on reducing the number of fitness function evaluations.

    Furthermore, as indicated in [5], [14], surrogate model management is critical for the performance of SAPSOs [19]. A typical strategy is to employ the generation-based method, in which the expensive fitness evaluations are replaced by computationally cheap approximating models [20]. But it is not practical for many extremely expensive problems because a large number of expensive fitness evaluations were needed for an accurate surrogate model [14]. Another known as individual-based model management focus on the determination of individuals demanding evaluation for each generation. Three criteria, such as minimizing the predicted fitness [21], minimizing the prediction uncertainty [18] and a combination of the previous two [22] are normally used to select new candidate solutions to be evaluated. However, it is difficult to strike a proper balance between exploration and exploitation [19]. Besides, active learning (AL) is a commonly deployed alternative for model management. In such a case, the new training data are actively queried for efficiently improving the approximation quality of the current model without significantly increasing the size of training dataset. Also, AL terminates when any additional training data have no effect on the information gain [23].

    As already shown in [24], AL-based methods can also be used for solving the multi-objective optimization problem. For example, an active learning method of the Pareto front was reported in [23], in which a surrogate model was constructed to approximate the Pareto front. A PSO assisted by the active learning method based on the committee was presented in [19], but it only considered the single objective problem with the high computational cost. In [25], a Pareto active learning (PAL) method was proposed to predict the Pareto optimal solutions. In comparison to the state-of-the-art algorithm PAL, a ε-Pareto active learning (ε-PAL) was presented in [26] to reduce computational complexity. However, how to select ε for achieving a satisfied tradeoff between prediction accuracy and classification cost is a challenge.

    In this paper, we assume that the optimization algorithm is allowed to evaluate a small number of selected solutions during optimization. It should be noted that we focus on developing a surrogate-assisted PSO with Pareto active leaning so as to effectively accelerate the convergence process. The main contributions of this work can be highlighted as follows:

    1) A mixed mutation sampling method based on simulated evolution is proposed to enhance the quality of the candidate solutions. Based on the advantage of fast convergence of PSO, the evolution of the swarm is simulated with the surrogate model rather than the true fitness function. And the particles which are beyond the Pareto front and locate at the sparse part of the Pareto front are carried out the mutation operation to promote the diversity of the population. Then the positions of the offspring are considered as the candidate solutions.

    2) A classification strategy using the improved ε-PAL is employed to identify candidate solutions that are Pareto-optimal and not Pareto-optimal. It is wellknown that the prediction accuracy and classification cost are controlled by parameter ε. However, the mathematical relationship between parameter ε, prediction accuracy and classification cost are not clear. This poses a challenge for determining the value of ε. Therefore, a greedy search method is suggested to select ε where the number of active sampling is employed as the evaluation criteria of the classification cost.

    3) A surrogate-assisted PSO with Pareto active learning is proposed to solve the computationally expensive optimization problems with multiple objectives. The Pareto-optimal solutions obtained by ε-PAL are used to guide the global search of the swarm, while the history information of PSO is employed to update the surrogate models.

    The rest of this paper is organized as follows. Section II presents a brief introduction to PSO, Gauss process modeling and ε-PAL. Then, a surrogate-assisted PSO with Pareto active learning is proposed in Section III. In Section IV, the results obtained from the numerical studies on eight benchmark test problems and two MLSSVM model optimization problems with the proposed algorithm are discussed. Finally, Section V draws the conclusions for this study.

    PSO [27] adopts experiences from both individuals and swarms to guide the direction of evolution. Assuming the population size is known as NP, the velocity and the location of the ith particle at t generation are updated as follows:

    vi(t+1)=ωvi(t)+c1r1[pi(t)xi(t)]+c2r2[pg(t)xi(t)] (3)
    xi(t+1)=xi(t)+vi(t+1) (4)

    where ω indicates the inertia weight; c1 and c2 are the learning rate; r1 and r2 represent the random numbers which are belong to (0,1); pi(t) and pg(t) stand for the personal best and the global best at t generation, respectively.

    Gauss process (GP) [28] surrogate model is employed to approximate each objective. For the jth objective, the response function is expressed as y(xi)=f(xi)+ζ, where f(xi) indicates the mapping function between the input x1:N and the output y1:N; ζN(0,δ2y) denotes the Gaussian noise. With the GP prior, the response function of the jth objective is formulated as yN(K(x1:N,x1:N)+δ2yI). Given the new point x, the mean μj(x) and variance δ2j(x) are expressed as follows:

    μj(x)=k(x1:N,x)T(K(x1:N,x1:N)+δ2yI)1y1:N (5)
    δ2j(x)=k(x,x)k(x1:N,x)T(K(x1:N,x1:N)+δ2yI)1k(x1:N,x) (6)

    where μj and δ2j indicate the predicted mean value and variance for the jth objective, respectively; K denotes the covariance matrix using the squared exponential function k(xp,xq)=δ2sexp(12(xpxq)TW(xpxq)), where δ2s represents the signal variance and W is the widths of the kernel.

    The ε-PAL [26] method is a classification method which is utilized to find ε-accurate Pareto front with as few function evaluations as possible. In ε-PAL, GP models are adopted to predict the candidate solutions that have not been evaluated yet. Furthermore, the predictive uncertainty is employed to guide the iterative sampling. Specifically, the sampling strategy is designed to maximize the probability that candidate solutions are Pareto-optimal. With a set of classification rules, the candidate solutions are identified as the Pareto-optimal and non-Pareto-optimal solutions with high probability. At last, ε-PAL terminates when all the candidate solutions are classified. The computing procedures for ε-PAL are summarized as Algorithm 1.

    Algorithm 1 The framework of ε-PAL
    Input: Ui (undecided points), Pi (predicted set), ε (user-controlled
    parameter)
    1: Initialize U1=E, P1= and set i=1.
    2: while Ui do
    3: Construct the surrogate model for each objective using (5) and (6).
    4: Assign the uncertainty region Ri(x) for each point xUiPi.
    5: Discard the points in Ui which are ε-dominated by another point with
    high probability.
    6: Move the points that belong to the ε-Pareto front of E from Ui to Pi.
    7: Sample a point from UiPi.
    8: i=i+1.
    9: end while
    Output: ε-accurate Pareto set ˆP=Pi.
     | Show Table
    DownLoad: CSV

    In this paper, the evaluation of the fitness function is considered to be computationally expensive. In order to obtain Pareto optimal solutions with the limited number of the fitness evaluations, a novel surrogate-assisted PSO with the help of Pareto active learning, termed PAL-SAPSO, is proposed in this section.

    The pseudo code of PAL-SAPSO is given as Algorithm 2, which can be summarized as six procedures as follows:

    1) Initialization (line 1): NP particles are generated using Latin hypercube sampling, which are evaluated using the fitness function. Then the external archive and the knowledge base are initialized.

    2) Swarm update (line 3 to 4): The swarm is updated with the personal best and the global best.

    3) Construction of surrogate model (line 7): The surrogate model is built with data in the knowledge base to approximate the fitness function for each objective.

    4) Sampling of candidate solution set (line 8): The mixed mutation sampling method based on the simulated evolution is implemented to get the candidate solutions.

    5) Classification of candidate solution set (line 9): The improved ε-PAL is employed to categorize the candidate solutions into Pareto-optimal solutions and non-Pareto-optimal ones.

    6) Update of the external archive (line 10): The external archive is updated by pre-selecting results from the improved ε-PAL.

    7) Repeat Step 2) to Step 6) until the maximum number of fitness evaluations is reached.

    In PAL-SAPSO, the sampling method and the classification strategy are crucial for the success of the algorithm. We will discuss them in detail in the following subsections.

    Algorithm 2 The framework of PAL-SAPSO
    Input: NP (population size), v (velocity), x (position), A (external
    archive), Dt (knowledge base with Nt samples at t generation),
    Nmax (maximum number of the fitness evaluations).
    1: Initialize NP, v, x, A, Dt and set Nt=NP.
    2: while NtNmax do
    3: Select the personal best pi(t) and the global best pg(t).
    4: Update v(t) and x(t) using (3) and (4), and calculate the corresponding
    fitness f(x).
    5: Update A (Using Algorithm 5).
    6: Update Dt with {x,f(x)}.
    7: Construct the surrogate model for each objective using (5) and (6).
    8: Get the candidate solution set (Using Algorithm 3).
    9: Get the ε-accurate Pareto optimal solution set (Using Algorithm 4).
    10: Update A (Using Algorithm 5).
    11: Update Dt.
    12: Count the number of the fitness evaluations Nt.
    13: t=t+1.
    14: end while
    Output: Pareto-optimal solutions from A.
     | Show Table
    DownLoad: CSV

    For maintaining convergence and diversity of the candidate solutions, a hybrid mutation sampling method based on simulated evolution is proposed. The main framework of the sampling method is listed in Algorithm 3.

    1) Simulated Evolution of the Particle Swarm: The PSO algorithm tries to find the optimal solution using a population of particles. Based on (3) and (4), it is clear that the trajectories drawn by the particles are semi-random in nature, as they derive from the contribution of systematic attraction towards the personal and global best solutions and the stochastic weighting of two acceleration terms [27]. Furthermore, PSO is known to have a very high convergence speed. Inspired by the above ideas, a sampling method based on the simulated evolution is designed to obtain the candidate solutions.

    Let us assume that the population of the tth generation in the real physical space is migrated into the virtual space where the cheap surrogate models are employed to take place of the computationally expensive fitness functions. Then the information of the initial population is expressed as xi(1)=xi(t), vi(1)=vi(t), ˆpi(1)=pi(t), ˆpg(1)=pg(t) and ˆA(1)=A(t). With the initial information, the velocity and position of each particle are updated as follows:

    vi(k+1)=ωvi(k)+c1r1[ˆpi(k)xi(k)]+c2r2[ˆpg(k)xi(k)] (7)
    xi(k+1)=xi(k)+vi(k+1) (8)

    where vi(k) and xi(k) indicate the velocity and the position of the ith particle in the virtual space at the kth iteration in the virtual space. ˆpi(k) and ˆpg(k) denote the personal best and the global best of the swarm at the kth iteration in the virtual space; ˆA and A are the external archives in the virtual space and the real physical space, respectively.

    In the process of simulated evolution, the surrogate models based on GP are used to replace the real fitness function. Based on the experience knowledge base Dt={xi,yi,j|1iNt,1jm}, m surrogate models are constructed for m objectives. Given any position of particle xi, fitness is estimated by ˆyi=(μ1(xi),,μm(xi)), where μj() represents the estimated value of the jth surrogate model. Then the personal best and the global best are updated as presented in Section III-D.

    2) Mutation on Particles Beyond the Pareto Front: Although particles which float away from the Pareto front are useful for maintaining the diversity of the population, they will slow down the convergence speed. In addition, non-dominant solutions are usually required to be as close as possible to the real Pareto front. Therefore, the mutation operation moving the particles towards to the Pareto front can not only accelerate the convergence speed, but also improve the quality of the solution.

    The particles which are beyond the Pareto front are identified first. The approximate Pareto front corresponding to the non-dominant solutions in the external archive is used as the reference front because of the real Pareto front is usually unknown. In the object space, the Euclidean distances between the fitness values of the particles and the reference front are different. Then, the particles whose Euclidean distance is larger than the average distance are considered to be far from the Pareto front, as expressed in the following formula:

    O={xh|h=find(li>ˉl)} (9)
    li=min( (10)

    where O denotes the set of the particles which are far from the Pareto front; {{x}}^{\prime} indicates the position of the particle which beyond the Pareto front; l_{i} indicates the Euclidean distance between the fitness values of the i\mathrm{th} particle and the closest non-dominant solution; \bar{l} is the mean of l_{i} ; find(\cdot) represents the query function on the index of the element; \widehat{A}(k) stands for the external archive at the k\mathrm{th} iteration in the virtual space; \widehat{{{y}}}_{i} indicates the estimated values of the fitness of i\mathrm{th} particle; {{a}}_{j} denotes the fitness values of the non-dominant solutions on the Pareto front.

    Let {{x}}^{\prime}_{i} is the particle beyond the Pareto front, then the mutation operation is carried out as follows:

    {{x}}^{\prime \prime}_{i}(k) = {{x}}^{\prime}_{i}(k)+r[{{x}}^{\prime}_{i}(k)-{{x}}^{\ast}_{i}(k)] (11)

    where {{x}}^{\prime \prime}_{i}(k) indicates the new position of the i\mathrm{th} particle in the virtual space at the k\mathrm{th} iteration; {{x}}^{\ast}_{i} denotes the closest non-dominant solution to {{x}}^{\prime} ; r\in\{0, 1\} .

    3) Mutation to Particles Locating at the Sparse Part of the Pareto Front: In the process of simulated evolution, a large number of particles will aggregate to the individual non-dominant solution according to the fast convergence of PSO. It causes that parts of Pareto front correspond to more non-dominant solutions, while others correspond to fewer. Then, the diversity distribution of the solutions becomes worse. Therefore, it is an effective method to identify the sparse parts of the Pareto front and generate new non-dominant solutions in the sparse parts.

    First of all, the particles which locate at the sparse part of the Pareto front are identified. The Pareto front constructed by the non-dominant solutions in the external archive is also used as the reference front. For any objective, the fitness values of the non-dominant solutions in the external archive are arranged in ascending order and the Euclidean distance between two adjacent non-dominant solutions is computed. When the Euclidean distance of the two non-dominant solutions is larger than the average distance of all adjacent non-dominant solutions, the part between two non-dominant solutions is considered to be sparse according to the following formula:

    S = \{({{x}}^{\ast}_{h}, {{x}}^{\ast}_{h+1})|h = find(d_{i}>\bar{d})\}, {{x}}^{\ast}_{h}\in \widehat{A}(k) (12)
    d_{j} = \min(\|{{a}}_{j}-{{a}}_{j+1} \|), {{a}}_{j}\in \widehat{A}(k), {{a}}_{j}<{{a}}_{j+1} (13)

    where S denotes the set of the sparse parts of the Pareto front; {{x}}^{\ast}_{h} and {{x}}^{\ast}_{h+1} represent the non-dominant solutions at the edge of the sparse part of the Pareto front; d_{j} indicates the Euclidean distance of the two adjacent non-dominant solutions; \bar{d} is the mean of d_{j} .

    Let {{x}}^{\ast}_{h} and {{x}}^{\ast}_{h+1} denote the non-dominant solutions at the edge of the sparse part of the Pareto front. Then, both particles move towards each other. The mutation operation is formulated as

    \begin{aligned} {{x}}^{\prime \prime \prime}(k) = \begin{cases} {{x}}^{\ast}_{h}(k)+r({{x}}^{\ast}_{h+1}(k)-{{x}}^{\ast}_{h}(k))\\ {{x}}^{\ast}_{h+1}(k)+r({{x}}^{\ast}_{h}(k)-{{x}}^{\ast}_{h+1}(k)) \end{cases} \end{aligned} (14)

    where {{x}}^{\prime \prime \prime}(k) indicates the position of the new particle which locates at the sparse part of the Pareto front in virtual space at the k\mathrm{th} iteration.

    4) A Hybrid Mutation Sampling Strategy Based on the Simulated Evolution: In the process of the simulated evolution, the positions of the particles involving the offspring and the particles which are beyond the Pareto front and locate at the sparse part of the Pareto front are updated at each generation. To get more information on position in the search space, the swarm is simulated \eta generations in virtual space. Then the candidate solutions obtained by mutation operation are determined as

    X_{\rm pss} = \{{{x}}^{\prime}(k), {{x}}^{\prime \prime}(k), {{x}}^{\prime \prime \prime}(k)\}_{k = 1:\eta} (15)

    where X_{\rm pss} denotes the set of candidate solutions; {{x}}^{\prime} , {{x}}^{\prime \prime} and {{x}}^{\prime \prime \prime} indicate the candidate solutions which are obtained from three mutation operations, respectively.

    Algorithm 3 A hybrid mutation sampling method based on the simulated evolution
    Input: {v}^{\prime} (velocity), {x}^{\prime} (position), \widehat{{p}}_{i} (personal best), \widehat{{p}}_{g} (global best), \widehat{A}
    (external archive), \eta (maximum number of generations).
    1: Initialize {{v}}^{\prime}={{v}} , {{x}}^{\prime}={{x}} , \widehat{{{p}}}_{i}(1)={{{p}}}_{i}(t) , \widehat{{{p}}}_{g}(1)={{{p}}}_{g}(t) , \widehat{A}(1)=A(t) .
    2: for k=1 to \eta do
    3: Update {{v}}^{\prime}(k) and {{x}}^{\prime}(k) using (7) and (8).
    4: Evaluate the fitness based on (5).
    5: Determine the particles which are beyond Pareto front using (9) and
    (10), and obtain new particles {{x}}^{\prime \prime}(k) based on (11).
    6: Determine the particles which locate the sparse part of the Pareto front
    using (12) and (13), and obtain the new particles {{x}}^{\prime \prime \prime}(k) based on (14).
    7: Get the set of the candidate solution X_{\rm pss} according to (15).
    8: Update \widehat{A}(k) with \{X_{\rm pss}, \widehat{{{y}}}(X_{\rm pss})\} .
    9: Select the personal best \widehat{{{p}}}_{i}(k) and the global best \widehat{{{p}}}_{g}(k) .
    10: end for
    Output: The candidate solution set X_{\rm pss}.
     | Show Table
    DownLoad: CSV

    In PAL-SAPSO, \varepsilon -PAL is employed to categorize the candidate solutions. Although the Pareto-optimal set obtained by \varepsilon -PAL is likely to be Pareto-optimal at current generation, but the candidate solutions can not cover the whole search space. As a result, the predicted Pareto-optimal set may not belong to the true Pareto front for the problem to be solved. Furthermore, the fitness function is computational expense, the classification cost from the sampling and the evaluations of the predicted Pareto-optimal solutions must be considered. In order to achieve a tradeoff between prediction accuracy and the classification cost, an improved \varepsilon -PAL is proposed, as described in Algorithm 4.

    As suggested in [26], the \varepsilon -PAL automatically stops when the target accuracy \varepsilon is achieved with confidence 1-\varpi . Obviously, once the target accuracy \varepsilon is fixed, there is a theoretical boundary for the sampling number. In turn, the target accuracy \varepsilon can be obtained when the sampling number is given.

    From line 7 in Algorithm 1, it is clear that each point in U_{i}\cup P_{i} have the possibility of being sampled. Let the desired number of evaluations, T , is equal to the number of the points in U_{i}\cup P_{i} . Then, the prediction accuracy \varepsilon is determined if the classification cost T is given. In order to achieve a balance between prediction accuracy \varepsilon and classification cost T , the resulting nonlinear programming problem can be expressed as follows:

    \underset{\varepsilon\in [\varepsilon_{l}, \varepsilon_{u}]}{\min}\; (n_{j}-T)^{2} (16)

    where n_{j} denotes the number of the points in U_{i}\cup P_{i} at the j\mathrm{th} greedy search; \varepsilon_{l} and \varepsilon_{u} represent the upper and lower bound of the target accuracy \varepsilon , respectively.

    Taking (16) into account, it is apparent that satisfactory prediction accuracy can be found under the expected classification cost. As a result, a balance strategy is presented by solving (16) greedily in the following form:

    \varepsilon(s) = \varepsilon(s-1)+\tau \Delta \varepsilon(s-1) (17)
    \Delta =\frac{n_{j}-T}{T} (18)
    \varepsilon_{l}\leq \varepsilon(s)\leq \varepsilon_{u} (19)

    where \tau represents the learning factor, \tau\in\{0, 1\} ; \Delta denotes the speed of searching. Based on the balance strategy, a set of \varepsilon -accurate Pareto optimal solutions \widehat{P}_{T} is obtained with the desired classification cost T .

    Algorithm 4 The framework of improved \varepsilon-PAL
    Input: U_{i} (undecided points), P_{i} (predicted set), T\; (classification cost),
    N_{\rm GS} (maximum number of greedy searches), \tau (learning factor).
    1: Initialize U_{1}=X_{\rm pss} , P_{1}= \emptyset , \varepsilon\in [\varepsilon_{l}, \varepsilon_{u}] , U^{\prime}= \emptyset , P^{\prime}= \emptyset and set i=1
    2: while U_{i}\neq \emptyset do
    3: Assign the uncertainty region R_{i}({{x}) for each point {{x}\in U_{i}\cup P_{i}
    4: Discard the points in U_{i} which are \varepsilon -dominated by another point with
    high probability.
    5: Move the points that belong to the \varepsilon -Pareto front of X_{\rm pss} from U_{i} to P_{i} .
    6: Count the number of the points n_{1} in U_{i}\cup P_{i} .
    7: Initialize U^{\prime}_{1}=U_{i} , P^{\prime}_{1}=P_{i} and set j=1 .
    8: while n_{j}>T or j<N_{\rm GS} do
    9: Assign the uncertainty region R^{\prime}_{j}({{x}^{\prime}) for each point {{x}^{\prime}\in U^{\prime}_{j}\cup P^{\prime}_{j} .
    10: Determine the value of \varepsilon using (17) to (19).
    11: Discard the points in U^{\prime}_{j} which are \varepsilon -dominated byanother point
    with high probability.
    12: Move the points that belong to the \varepsilon -Pareto frontof X_{\rm pss} from to U^{\prime}_{j}
    P^{\prime}_{j} .
    13: j=j+1 .
    14: Count the number of the points n_{j} in U^{\prime}_{j}\cup P^{\prime}_{j} .
    15: Save the value of \varepsilon .
    16: end while
    17: Sample a point from U_{i}\cup P_{i} .
    18: i=i+1 .
    19: end while
    Output: The \varepsilon-accurate Pareto set \widehat{P}_{T}=P_{i}.
     | Show Table
    DownLoad: CSV

    For the multiobjective optimization problems, PSO intends to search for a set of non-dominated solutions to be stored in an external archive with a predefined size N_{\rm rep} . After initializing the PSO population, every particle will fly toward the direction guided by its personal best and global best. However, the PSO-based algorithm may converge to a false Pareto front because of its very high convergence speed. Therefore, it is important to ensure that the particles in the population move toward a more sparse and promising region in the search space.

    At the t\mathrm{th} generation, the personal best of the i\mathrm{th} particle {{p}}_{i}(t) is updated by a new solution {{x}}_{i}(t+1) found by this particle if {{p}}_{i}(t) is dominated by {{x}}_{i}(t+1) . And, the personal best of the i\mathrm{th} particle is selected between {{p}}_{i}(t) and {{x}}_{i}(t+1) randomly if they are non-dominated with respect to each other.

    The global best {{p}}_{g}(t) is selected from the external archive randomly, which is useful in guiding the particle to fly in the different direction. In addition, the external archive is updated with the approximate Pareto solution set \widehat{P}_{T} to enhance the social experience of the population, as described in Algorithm 5. Here, the popular archive update mechanism used in [6] is also adopted, which is based on both Pareto dominance and crowding distance.

    Algorithm 5 Archive update
    Input: \widehat{P}_{T} (\varepsilon-accurate Pareto set), A (external archive), N_{\rm rep} (size of external archive)
    1: for i=1 to |\widehat{P}_{T}| do
    2: if \widehat{P}_{i} is dominated by any member in A then
    3: Discard \widehat{P}_{i} .
    4: else if \widehat{P}_{i} dominates a set of member in A then
    5: A=A \cup \{\widehat{P}_{i}\} .
    6: else if \widehat{P}_{i} and any member of A are non-dominatedwith each other then
    7: A=A \cup \{\widehat{P}_{i}\} .
    8: end if
    9: if |A| > N_{\rm rep} then
    10: Calculate the crowding distance value for each solution in A .
    11: Delete the most crowed one.
    12: end if
    13: end for
    Output: The external archive A.
     | Show Table
    DownLoad: CSV

    The proposed algorithm in this study consists of a hybrid mutation sampling method based on simulated evolution, improved \varepsilon -PAL and PSO. Firstly, in the hybrid mutation sampling method based on simulated evolution, 1) implementation of PSO requires {\rm{O}}(NP\times N^2_{\rm rep}) computations. 2) Construction of the surrogate models require {\rm{O}}(m\times N^3_{t}) computations. 3) Assuming the number of candidate solutions is N_{\rm pss} = (NP+N_{\rm rep})\times \eta (In fact, the number of candidate solutions is smaller than the hypothetical value because not all particles implement mutation). Thus, the estimation of the candidate solutions require {\rm{O}}(N_{\rm pss}\times N^2_{t}) computations. Due to N_{t} increases dynamically, the worst-case complexity of the hybrid mutation sampling method based on simulated evolution is {\rm{O}}(m\times N^3_{t}) . Secondly, in the improved \varepsilon -PAL, 1) determination of the value of \varepsilon requires {\rm{O}}(N_{\rm pss}{\rm log}(N_{\rm pss})\times N_{\rm GS}) computations ( \varepsilon -PAL requires {\rm{O}}(n{\rm log}(n)) computations on every iteration [26], where n indicates the number of the undecided points), where N_{\rm GS} is the number of greedy searches. 2) Identification of Pareto solutions require {\rm{O}}(m\times (N_{t}+T-1)^3+(N_{t}+T-1)^2) computations (In the worst case, the active sampling is carried out T times and 1{\rm log}(1) = 0 ). Thus, considering the worst case, the complexity of the improved \varepsilon -PAL is {\rm{O}}(m\times (N_{t}+T-1)^3+(N_{t}+T-1)^2) . And finally, PSO optimizer requires {\rm{O}}(NP\times N^2_{\rm rep}) computations. Considering all the above computations, the overall worst-case complexity of generation of the proposed algorithm is {\rm{O}}(m\times (N_{t}+T-1)^3+(N_{t}+T-1)^2).

    To verify the performance of the proposed method (PAL-SAPSO), three existing MOPSO algorithms were employed to make a comparison with a set of widely used benchmark test problems. Three state-of-the-art MOPSO algorithms include standard MOPSO [27], decomposition-based multi-objective particle swarm optimization (dMOPSO) [29] and speed-constrained multi-objective PSO (SMPSO) [30]. Both dMOPSO and SMPSO are improved version of standard MOPSO. In addition, dMOPSO is based on decomposition strategy and SMPSO adopts Pareto sorting strategy. The source codes of three MOPSO algorithms are provided in PlatEMO [31]. All parameters are set to the recommended values in the original papers.

    A total of 8 test problems with two objectives are used to evaluate the performance of the algorithms, where ZDT1, ZDT2, ZDT3, UF1, UF2, UF4 and UF7 have 6 decision variables and ZDT6 is 10. The mathematical description of the test problems was given in [32]. The inverted generational distance (IGD) metric [33] is deployed as the index for evaluating the performance of the compared algorithms, and almost 1000 points uniformly sampled on the Pareto front are used in the calculation of IGD for each test problem. The IGD is defined as follows:

    IGD(P_{\rm true}, P_{\rm rep}) = \frac{\sum\limits_{{{x}}\in P_{\rm true}}dis({{x}}, P_{\rm rep})}{|P_{\rm true}|} (20)

    where P_{\rm true} represents the true Pareto front reference points; P_{\rm rep} denotes the non-dominant solution in the external archive; dis({{x}}, P_{\rm rep}) denotes the Euclidean distance between {{x}} and the non-dominant solution in the external archive; |P_{\rm true}| is the number of real Pareto forward reference points. The IGD is a metric for evaluating the quality of obtained solution set in terms of both convergence and distribution. The smaller the IGD value, the better the quality of solution set obtained by the algorithm.

    The parameters of the proposed algorithm PAL-SAPSO are set as follows: the population size is set to NP = 30 ; the inertia weight is set to \omega = 0.5 ; the learning rate is set to c_{1} = 1 and c_{2} = 2 ; the generation of the simulated evolutionary is set to \eta = 20 ; the learning factor is set to \tau = 0.02 ; the prediction accuracy is set to \varepsilon\in \{0, 1\} , the classification cost is set to T = 5 and the number of the greedy searches is set to N_{\rm GS} = 1000 . For each test problem, 20 independent experiments are conducted and the maximum number of the fitness evaluations is set to 500, 1000, 1500 and 2000. All experimental results are obtained on a PC (with an Intel Core i5 6200U CPU) and MATLAB 2017a.

    The comparative results from 4 algorithms (MOPSO, dMOPSO, SMPSO, and PAL-SAPSO) with 500 function evaluation on ZDT and UF test problems are showed in Table I. The statistical results consist of the mean (the corresponding standard deviation) of IGD which are obtained from 20 independent experiments. It can be seen from Table I that the mean of IGD from PAL-SAPSO is smaller than the other algorithms. It exhibits that the Pareto fronts obtained by PAL-SAPSO have the best performance on the convergence and the distribution. And, the standard deviation of the means from PAL-SAPSO are the smallest. It shows that PAL-SAPSO has good stability. To clearly visualize the comparison of results, the Pareto fronts which are obtained by 4 algorithms with 500 function evaluation on ZDT and UF test problems are illustrated in Fig. 1. The results in Fig. 1 correspond to the smallest values of IGD in the 20 independent experiments. It can be concluded that PAL-SAPSO effectively converges to the true front and find the more diverse solutions compared to MOPSO, dMOPSO and SMPSO.

    Table  I.  Comparative Results From 4 MOPSO Algorithms
    Problem MOPSO dMOPSO SMPSO PAL-SAPSO
    ZDT1 3.86 (2.69) 3.39×10−1 (2.18×10−1) 6.87×10−1 (9.07×10−1) 1.21×10−2 (3.38×10−3)
    ZDT2 3.60 (2.43) 5.52×10−1 (2.01×10−1) 6.26×10−1 (9.94×10−1) 4.45×10−2 (1.34×10−1)
    ZDT3 2.70 (1.85) 3.21×10−2 (1.70×10−1) 5.36×10−1 (4.62×10−1) 2.90×10−2 (1.31×10−2)
    ZDT6 1.06×10 (8.96×10−1) 5.04×10−1 (1.37×10−1) 8.14 (2.33) 4.80×10−2 (3.12×10−2)
    UF1 2.79×10−1 (6.93×10−2) 2.93×10−1 (7.80×10−2) 2.70×10−1 (8.44×10−2) 6.68×10−2 (2.41×10−2)
    UF2 2.13×10−1 (2.22×10−2) 1.71×10−1 (1.56×10−2) 2.11×10−1 (1.83×10−2) 1.15×10−1 (4.75×10−2)
    UF4 1.93×10−1 (1.30×10−2) 1.88×10−1 (7.97×10−3) 1.74×10−1 (6.58×10−3) 6.69×10−2 (7.93×10−3)
    UF7 3.87×10−1 (8.55×10−2) 3.39×10−1 (1.35×10−1) 1.74×10−1 (1.27×10−1) 1.34×10−1 (4.81×10−2)
     | Show Table
    DownLoad: CSV
    Figure  1.  The convergence results obtained by algorithms on 8 test problems.

    To further investigate the convergence of PAL-SAPSO, the mean of IGD from MOPSO, dMOPSO, SMPSO, and PAL-SAPSO on ZDT and UF test problems with the 500, 1000, 1500, and 2000 fitness evaluations are shown in Fig. 2. As shown in these figures, the convergence of PAL-SAPSO is better than MOPSO, dMOPSO and SMPSO.

    Figure  2.  The convergence results obtained by algorithms on 8 test problems.

    In summary, compared with MOPSO, dMOPSO and SMPSO, the Pareto fronts obtained by PAL-SAPSO have better performance on convergence and distribution with the same number of the fitness evaluations.

    In order to further verify the effectiveness of the proposed algorithm, two application problems arising from the parameter determination for MLSSVM-based models [34] are employed. In general, the generalization capability of the data-driven model is evaluated by the multifold cross-validation method [35].

    For predicting an output vector {{y}}\in\mathbb{R}^{m} from a given input vector {{x}}\in\mathbb{R}^{n} , the MLSSVM-based model is established by solving a constraint-based optimization problem [34]

    \begin{split} \underset{{{w}}_{0}\in \mathbb{R}^{n_{h}}, {{V}}\in \mathbb{R}^{n_{h}}, {{b}}\in \mathbb{R}^{m}}{\min}{{J}}({{w}}_{0}, {{V}}, \rm{\Xi}) = \;&\frac{1}{2}{{w}}_{0}^{T}{{w}}_{0}+\frac{\lambda}{2m}{\rm tr}({{V}}^{T}{{V}})\\ &+\frac{\gamma}{2}{\rm tr}({\Xi}^{T}{\Xi}) \end{split} (21)
    {\rm{\; s.t.\; }}{{Y}} = {{Z}}^{T}{{W}} = repmat({{b}}^{T}, l, 1)+\rm{\Xi}. (22)

    Once the Lagrangian function is constructed with respect to (21) and (22), a linear system is obtained with Karush-Kuhn-Tucker (KKT) conditions as follows:

    \begin{aligned} \begin{bmatrix} {{0}}_{ml\times m}&{{P}}^{T}\\ {{P}}&{{H}} \end{bmatrix} \begin{bmatrix} {{b}}\\\alpha \end{bmatrix} = \begin{bmatrix} {{0}}_{m}\\ {{y}} \end{bmatrix} \end{aligned} (23)

    where {{P}} = blockdiag({\bf{1}}_{l}, {\bf{1}}_{l}, \dots, {\bf{1}}_{l})\in \mathbb{R}^{ml\times m} , {{H}} = \Omega+\gamma^{-1}{\bf{I}}_{ml}+ (m/ \lambda){{Q}}\in \mathbb{R}^{ml\times ml} is the positive definite matrix. If the solutions ( \alpha^{\ast} and {{b}}^{\ast} ) of (23) are obtained, one can get the corresponding decision function for multiple outputs as follows:s

    f({{x}}) = repmat \left(\sum_{i = 1}^m \sum_{j = 1}^l\alpha^{\ast}_{i, j} k({{x}}, {{x}}_{j}), 1, m \right)\\ +\frac{\lambda}{m}\sum_{j = 1}^l \alpha^{\ast}_{j}k({{x}}, {{x}}_{j})+{{b}}^{\ast {T}} (24)

    where k({ x}, { y}) = {\rm exp}(-\dfrac{1}{2\delta^{2}}\|{ x}-{ y}\|^{2}) . Three parameters (two positive real regularized parameters \lambda , \gamma and kernel parameter \delta ) in MLSSVM model should be determined. In order to obtain a MLSSVM regression model with the satisfactory generalization, the 10-fold cross-validation method is adopted. The optimization problem is described as follows:

    \min\; {{E}} = {{F}}({\theta}) = (f_{1}({\theta}), f_{2}({\theta}), \dots, f_{m}({\theta})) (25)
    \begin{aligned} {\rm{\; s.t.\; }} \begin{cases} f({\theta})_{j} = \dfrac{1}{10} \displaystyle\sum\limits_{i = 1}^{10} L(A_{{\theta}}, D_{\rm train}^{i}, D_{\rm valid}^{i})\\\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\; j = 1, 2, \dots, m\\\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\; {\theta}_{l}\leq {\theta}\leq {\theta}_{u} \end{cases} \end{aligned} (26)

    where {{E}} represents the cross-validation error (the mean square error); f({\theta})_{j} denotes the j\mathrm{th} objective function; A_{{\theta}} indicates the MLSSVM model with the parameters {\theta} = (\lambda, \gamma, \delta) ; L(A_{{\theta}}, D_{\rm train}^{i}, D_{\rm valid}^{i}) represents the cross-validation method; D_{\rm train}^{i} stands for the training data; D_{\rm valid}^{i} is the validation data; {\theta}_{l} and {\theta}_{u} denote the upper and lower boundaries of the parameters {\theta} , respectively.

    To evaluate the effectiveness of the parameter configuration scheme, the predictive performance of the MLSSVM model is evaluated by the mean square error MSE = \dfrac{1}{l}\displaystyle\sum\nolimits_{i = 1}^{l}(y_{i}-\widehat{y}_{i}) , where y_{i} and \widehat{y}_{i} represent the real output and the predicted output, respectively.

    1) Parameter Determination for MLSSVM-based Model Using the Standard Data: The purpose of optimization considered here is to find a MLSSVM regression model with satisfactory generalization. In this experiment, the energy efficiency dataset [36] is employed which is downloaded from UCI machine learning repository. 500 data points are selected to train this model and 200 data points are used as testing data. The MLSSVM model has three parameters to be configured and the energy efficiency dataset contains two output responses. Then, the optimization problem with three decision variables and two objective variables is considered according to (25) and (26).

    The optimal parameter configuration corresponds to the minimum of the mean square error between the predicted and real values. The i\mathrm{th} parameter configuration is expressed as MSE({\theta}_{i}) = MSE_{1}({\theta}_{i})+MSE_{2}({\theta}_{i}) , where MSE_{1}({\theta}_{i}) indicates the mean square error between the predicted values and the real values for the objective 1 and MSE_{2}({\theta}_{i}) represents the mean square error between the predicted values and the real values for objective 2. \min(MSE({\theta}_{i})) corresponds to the MLSSVM model with the best prediction performance. Since the order of magnitude of the mean square error is different, MSE_{1}({\theta}_{i}) and MSE_{2}({\theta}_{i}) need to be normalized. In this experiment, 10 independent experiments are conducted and the maximum number of the fitness evaluations is set to 300 and 500. The statistical results of the experiment are shown in Table II, where MSE1, MSE2 and MSE represent the mean (the corresponding standard deviation) of MSE_{1}({\theta}_{i}) , MSE_{2}({\theta}_{i}) and MSE({\theta}_{i}) , i = 1, 2, \dots, 10 . The parameter setting of the PAL-SAPSO is the same as that in the experiments on benchmark problems. For the MLSSVM model, the parameter ranges are set to \lambda\in [2^{-15}, 2^{8}] , \gamma\in [2^{-10}, 2^{4}] and \delta\in [2^{-15}, 2^{3}] .

    Table  II.  Statistic Results for Parameter Determination of MLSSVM-based Model
    300 fitness evaluations 500 fitness evaluations
    MSE1 MSE2 MSE MSE1 MSE2 MSE
    MOPSO 9.15×10−1 (3.19×10−1) 6.97×10−1 (4.06×10−2) 1.61×10−1 (3.43×10−1) 8.63×10−1 (2.30×10−1) 6.89×10−1 (2.97×10−2) 1.55×10−1 (2.43×10−2)
    dMOPSO 7.92×10−1 (1.10×10−1) 7.22×10−1 (1.43×10−1) 1.51×10−1 (1.36×10−1) 8.25×10−1 (1.15×10−1) 6.82×10−1 (2.28×10−2) 1.50×10−1 (1.06×10−1)
    SMPSO 8.90×10−1 (2.53×10−1) 6.92×10−1 (4.53×10−2) 1.58×10−1 (2.65×10−1) 7.35×10−1 (1.45×10−1) 6.81×10−1 (3.24×10−2) 1.42×10−1 (1.52×10−1)
    ALMOPSO 6.77×10−1 (5.98×10−2) 6.92×10−1 (4.53×10−2) 1.37×10−1 (6.93×10−2) 6.63×10−1 (9.20×10−2) 6.87×10−1 (1.59×10−2) 1.34×10−1 (8.12×10−2)
     | Show Table
    DownLoad: CSV

    With respect to the results obtained with the 300 fitness evaluations in Table II, it can be clearly seen that the values of MSE1 obtained by PAL-SAPSO are the smallest for objective 1; the values of MSE2 obtained by PAL-SAPSO are similar to the ones obtained by MOPSO and SMPSO for objective 2; the values of MSE obtained by PAL-SAPSO are the smallest. These facts indicate that the convergence and the stability of PAL-SAPSO are better than MOPSO, dMOPSO and SMPSO, and the performance of MLSSVM model obtained by PAL-SAPSO is the best. As for the results obtained with the 500 fitness evaluations, it can be seen that MOPSO, dMOPSO and SMPSO can effectively reduce the mean square error by increasing the number of fitness evaluations. A similar parameter configuration scheme is obtained by PAL-SAPSO with the 300 and 500 fitness evaluations. It appears that PAL-SAPSO can find the reasonable parameter configuration more effectively with the same number of fitness evaluations compared with MOPSO, dMOPSO and SMPSO. In order to illustrate the effectiveness of PAL-SAPSO more intuitively, the predicted results with the best parameter configuration scheme are shown in Fig. 3.

    Figure  3.  The predicted results with MLSSVM.

    2) Parameter Determination for MLSSVM Model Using Industrial Data: In this experiment, the goal of optimization is to accurately predict real variation for the gas holder level with the MLSSVM model. The blast furnace gas (BFG) system [37] consists of BFG converters for gas generation, pipeline networks for gas transportation, gas users for consumption and gas holder for temporary storage. As the only storage equipment and buffer unit of the pipe network, the gas holder level reflects the balance state of the generation and consumption of the gas in the pipe network. In order to simplify the problem, the inner relationships between the gas generation users and the gas consumption users are not considered. Then, the total gas generation flow, the total gas consumption flow, the 1# gas holder level at the first moment and the 2# gas holder level at the first moment are considered as the inputs. The 1# gas holder level and the 2# gas holder level are seen as the outputs. Based on samples from the input and output above, the MLSSVM model has three parameters need to be determined. Therefore, the optimization problem with three decision variables and two objective variables is considered according to (25) and (26).

    To establish the MLSSVM model, the gas flow data over 4 weeks are randomly selected. The sampling interval of the time series data is one minute. The number of the training samples is set to 5040 (two weeks data). The process of predicting is triggered every 5 points (5 minutes) to predict the future 30-min gas flow. It takes about 127.50 s to carry out an exact fitness function evaluation (10-fold cross-validation). In order to achieve the optimization of the MLSSVM model within 12 hours, the maximum number of fitness evaluations is set to 300. Furthermore, in order to predict the real variation of the gas holder level, the dynamic optimization of the MLSSVM model (DMLSSVM) is carried out every 12 hours. The training samples of the DMLSSVM model are updated every 12 hours. For verifying the feasibility of the DMLSSVM model, a static MLSSVM model (SMLSSVM) is developed. The training samples of the SMLSSVM model are from the gas flow data of the first two weeks. The parameter setting of the PAL-SAPSO is the same as that in the experiments on benchmark problems. The parameter ranges of the MLSSVM model are set to \lambda\in [2^{-15}, 2^{8}] , \gamma\in [2^{-10}, 2^{4}] and \delta\in [2^{-15}, 2^{3}]

    For verifying the effectiveness of PAL-SAPSO in the optimization problem on MLSSVM model driven by the gas flow data, two types of working conditions are selected as the representatives in this study: 1) 1# gas holder is running, 2# gas holder is turned from off-line to running; 2) 1# gas holder and 2# gas holder are both running. The corresponding simulation results are shown in Fig. 4 respectively. As shown in Fig. 4, DMLSSVM can effectively predict the change trend of 1# gas holder level and 2# gas holder level for two types of working conditions. Although it roughly predicts the changing trend of 1# gas holder level in the first type of working conditions, SMLSSVM fails to estimate the tendency of 2# gas holder level and the 1# and 2# gas holder level in the last type of working condition. In all, based on the dynamic optimization with PAL-SAPSO, the real changing trend of the gas holder level can be accurately forecasted with the parameter optimized MLSSVM model.

    Figure  4.  The simulation results obtained by algorithms on two types of working condition.

    In this paper, a surrogate-assisted PSO with Pareto active learning is proposed to solve the multi-objective optimization problem with high computational cost. In this algorithm, PSO is regarded as a sampler to generate candidate solutions. The performance of the PSO optimizer is improved by pre-selecting results from the candidate solutions with the improved \varepsilon -PAL. To improve the performance of the sampler, a hybrid mutation sampling method based on simulated evolution is proposed, which takes into account fast convergence and the diversity of distribution. In addition, a new method for determining the parameter ε is suggested where the number of the active sampling is used as an evaluation index of classification cost. The effectiveness of the proposed algorithm is verified by eight standard test problems. Compared with MOPSO, dMOPSO and SMPSO, the experimental results demonstrated that the Pareto optimal solutions obtained by the proposed algorithm has better convergence and distribution. In addition, comparative studies are also conducted on two MLSSVM model optimization problems, in which the advantage of determining effective parameter configuration schemes for the proposed method is verified on not only the standard data but also the real-world applications.

    A few parameters (e.g., the value of the classification cost T ) in the proposed algorithm must be predefined by the user, which do affect the performance of algorithms and must be defined properly. Therefore, more attention will be paid to address this issue. In addition, the performance of the proposed algorithm on expensive optimization problems with high dimension also remains to be verified.

  • [1]
    G. R. Zavala, A. J. Nebro, F. Luna, and C. A. C. Coello, " A survey of multi-objective metaheuristics applied to structural optimization,” Struct. Multidiscipl. Optim., vol. 49, no. 4, pp. 537–558, Apr. 2014. doi: 10.1007/s00158-013-0996-4
    [2]
    M. Park, M. Nassar, and H. Vikalo, " Bayesian active learning for drug combinations,” IEEE Trans. Biomed. Eng., vol. 60, no. 11, pp. 3248–3255, Nov. 2013. doi: 10.1109/TBME.2013.2272322
    [3]
    D. Ronco, C. Comis, R. Ponza, and E. Benini, " Aerodynamic shape optimization of aircraft components using an advanced multi-objective evolutionary approach,” Comput. Methods Appl. Mech. Eng., vol. 285, pp. 255–290, Mar. 2015. doi: 10.1016/j.cma.2014.10.024
    [4]
    J. Snoek, I. Larochelle, and R. P. Adans., " Practical Bayesian optimization of machine learning algorithms,” in Proc. 25th Int. Conf. NIPS, Lake Tahoe, Nevada, 2012, pp. 2951–2959.
    [5]
    H. D. Wang, Y. C. Jin, and J. O. Jansen, " Data-driven surrogate-assisted multiobjective evolutionary optimization of a trauma system,” IEEE Trans. E Comput., vol. 20, no. 6, pp. 939–952, Dec. 2016. doi: 10.1109/TEVC.2016.2555315
    [6]
    Q. L. Zhu, Q. Z. Lin, W. N. Chen, K. C. Wong, C. A. Coello Coello, J. Li, J. Chen, and J. Zhang, " An external archive-guided multiobjective particle swarm optimization algorithm,” IEEE Trans. Cybern., vol. 47, no. 9, pp. 2794–2808, Sep. 2017. doi: 10.1109/TCYB.2017.2710133
    [7]
    Q. Z. Lin, S. B. Liu, Q. L. Zhu, C. Y. Tang, R. Z. Song, J. Y. Chen, C. A. Coello Coello, K. C. Wong, and J. Zhang, " Particle swarm optimization with a balanceable fitness estimation for many-objective optimization problems,” IEEE Trans. E Comput., vol. 22, no. 1, pp. 32–46, Feb. 2018. doi: 10.1109/TEVC.2016.2631279
    [8]
    W. J. Kong, T. Y. Chai, J. L. Ding, and Z. W. Wu, " A real-time multiobjective electric energy allocation optimization approach for the smelting process of magnesia,” Acta Autom. Sinica, vol. 40, no. 1, pp. 51–61, Jan. 2014.
    [9]
    W. Hu and G. G. Yen, " Adaptive multiobjective particle swarm optimization based on parallel cell coordinate system,” IEEE Trans. E Comput., vol. 19, no. 1, pp. 1–18, Feb. 2015. doi: 10.1109/TEVC.2013.2296151
    [10]
    A. Trivedi, D. Srinivasan, K. Sanyal, and A. Ghosh, " A survey of multiobjective evolutionary algorithms based on decomposition,” IEEE Trans. E Comput., vol. 21, no. 3, pp. 440–462, Jun. 2017.
    [11]
    Q. Kang, S. W. Feng, M. C. Zhou, A. C. Ammari, and K. Sedraoui, " Optimal load scheduling of plug-in hybrid electric vehicles via weight-aggregation multi-objective evolutionary algorithms,” IEEE Trans. Intell. Transp. Syst., vol. 18, no. 9, pp. 2557–2568, Sep. 2017. doi: 10.1109/TITS.2016.2638898
    [12]
    R. T. Haftka, D. Villanueva, and A. Chaudhuri, " Parallel surrogate-assisted global optimization with expensive functions — a survey,” Struct. Multidiscip. Optim., vol. 54, no. 1, pp. 1–11, Apr. 2016. doi: 10.1007/s00158-016-1491-5
    [13]
    M. Tabatabaei, J. Hakanen, M. Hartikainen, K. Miettinen, and K. Sindhya, " A survey on handling computationally expensive multiobjective optimization problems using surrogates: non-nature inspired methods,” Struct. Multidiscip. Optim., vol. 52, no. 1, pp. 1–25, Jul. 2015. doi: 10.1007/s00158-015-1226-z
    [14]
    Y. Jin, " Surrogate-assisted evolutionary computation: recent advances and future challenges,” Swarm E Comput., vol. 1, no. 2, pp. 61–70, Jun. 2011. doi: 10.1016/j.swevo.2011.05.001
    [15]
    C. Ma and L. Y. Qu, " Multiobjective optimization of switched reluctance motors based on design of experiments and particle swarm optimization,” IEEE Trans. Energy Convers., vol. 30, no. 3, pp. 1144–1153, Sep. 2015. doi: 10.1109/TEC.2015.2411677
    [16]
    X. R. Ye, H. Chen, H. M. Liang, X. J. Chen., and J. X. You, " Multi-objective optimization design for electromagnetic devices with permanent magnet based on approximation model and distributed cooperative particle swarm optimization algorithm,” IEEE Trans. Magn., vol. 54, no. 3, pp. 1–5, Mar. 2018.
    [17]
    H. X. Jie, Y. Z. Wu, J. J. Zhao, J. W. Ding, and L. liang, " An efficient multi-objective PSO algorithm assisted by Kriging metamodel for expensive black-box problems,” J. Glob. Optim., vol. 67, no. 1–2, pp. 399–423, Jan. 2017.
    [18]
    F. Bittner and I. Hahn, " Kriging-assisted multi-objective particle swarm optimization of permanent magnet synchronous machine for hybrid and electric cars,” in Proc. IEEE Int. Elect. Mach. Drives Conf., Chicago, IL, USA, 2013, pp. 15–22.
    [19]
    H. D. Wang, Y. C. Jin, and J. O. Jansen, " Committee-based active learning for surrogate-assisted particle swarm optimization of expensive problems,” IEEE Trans. Cybern., vol. 47, no. 9, pp. 2664–2677, Jun. 2017. doi: 10.1109/TCYB.2017.2710978
    [20]
    M. Parno, T. Hemker, and K. R. Fowler, " Applicability of surrogates to improve efficiency of particle swarm optimization for simulation-based problems,” Eng. Optim., vol. 44, no. 5, pp. 521–535, Sep. 2012. doi: 10.1080/0305215X.2011.598521
    [21]
    Z. C. Cao, C. R. Lin, M. C. Zhou, and R. Huang, " Scheduling semiconductor testing facility by using cuckoo search algorithm with reinforcement learning and surrogate modeling,” IEEE Trans. Autom. Sci. Eng.. 2018. DOI: 10.1109/TASE.2018.2862380.
    [22]
    X. Y. Sun, D. W. Gong, Y. C. Jin, and S. S. Chen, " A new surrogate-assisted interactive genetic algorithm with weighted semisupervised learning,” IEEE Trans. Cybern., vol. 43, no. 2, pp. 685–698, Apr. 2013. doi: 10.1109/TSMCB.2012.2214382
    [23]
    P. Campigotto, A. Passerini, and R. Battiti, " Active learning of Pareto fronts,” IEEE Trans. Neural Netw. Learn. Syst., vol. 25, no. 3, pp. 506–519, Mar. 2014. doi: 10.1109/TNNLS.2013.2275918
    [24]
    D. Reker, P. Schneider, and G. Schneider, " Multi-objective active machine learning rapidly improves structure-activity models and reveals new protein-protein interaction inhibitors,” Chem. Sci., vol. 7, pp. 3919–3727, Mar. 2016. doi: 10.1039/C5SC04272K
    [25]
    M. Zuluaga, G. Sergent, A. Krause, and M. Püschel, " Active learning for multi-objective optimization,” in Proc. Int. Conf. Mach. Learning, Atlanta, GA, USA, 2013, pp. 462–470.
    [26]
    M. Zuluaga, A. Krause, and M. Puschel, " ε-PAL: an active learning approach to the multi-objective optimization problem,” J. Mach. Learn. Res., vol. 17, no. 1, pp. 3619–3650, Aug. 2016.
    [27]
    D. F. Wang, and L. Meng, " Performance analysis and parameter selection of PSO algorithms,” Acta Autom. Sinica, vol. 42, no. 10, pp. 1552–1561, Jan. 2016.
    [28]
    C. E. Rasmussen and H. Nickisch, " Gaussian processes for machine learning (GPML) toolbox,” J. Mach. Learn. Res., vol. 11, pp. 3011–3015, Nov. 2010.
    [29]
    S. Z. Martínez and C. C. Coello, " A multi-objective particle swarm optimizer based on decomposition,” in Proc. Genetic Evol. Comput., Dublin, Ireland, 2011, pp. 69–76.
    [30]
    A. J. Nebro, J. J. Durillo, J. Garcia-Nieto, C. A. coello coeelo, F. Luna, and R. Aeba, " SMPSO: a new PSO-based metaheuristic for multi-objective optimization,” in Proc. IEEE Symp. Comput. Intell. Miulticriteria Decision Making, Nashville, TN, USA, 2009, pp. 66–73.
    [31]
    Y. Tian, R. Cheng, X. Y. Zhang, and Y. C. Jin, " PlatEMO: A MATLAB platform for evolutionary multi-objective optimization,” IEEE Comput. Intell. Mag., vol. 12, pp. 73–87, Nov. 2017.
    [32]
    Q. F. Zhang, A. M. Zhou, S. Z. Zhao, P. Suganthan, W. D. Liu, and S. Tiwari,, " Multiobjective optimization test instances for the CEC-2009 special session and competition,” Nanyang Technol. Univ., Singapore, Tech. Rep., 2008. http://www.ntu.edu.sg/home/epnsugan/.
    [33]
    Q. F. Zhang, A. M. Zhou, and Y. C. Jin, " RM-MEDA: a regularity model-based multiobjective estimation of distribution algorithm,” IEEE Trans. E Comput., vol. 21, no. 1, pp. 41–63, Feb. 2008.
    [34]
    S. Xu, X. An, X. D. Qiao, L. J. Zhu, and L. Li, " Multi-output least-squares support vector regression machines,” Pattern Recognit. Lett., vol. 34, no. 9, pp. 1078–1084, Jul. 2013. doi: 10.1016/j.patrec.2013.01.015
    [35]
    J. Bergstra and Y. Bengio, " Random search for hyper-parameter optimization,” J. Mach. Learn. Res., vol. 13, no. 1, pp. 281–305, Feb. 2012.
    [36]
    S. Chen, " Multi-output regression using a locally regularised orthogonal least square algorithm,” IEEE Proc. Vision Image Signal Process., vol. 149, no. 4, pp. 185–195, Aug. 2002. doi: 10.1049/ip-vis:20020401
    [37]
    Z. Y. Han, Y. Liu, J. Zhao, and W. Wang, " Real time prediction for converter gas tank levels based on multi-output least square support vector regressor,” Control Eng. Pract., vol. 20, no. 12, pp. 1400–1409, Dec. 2012. doi: 10.1016/j.conengprac.2012.08.006
  • Related Articles

    [1]Jianqing Lin, Cheng He, Ye Tian, Linqiang Pan. Variable Reconstruction for Evolutionary Expensive Large-Scale Multiobjective Optimization and Its Application on Aerodynamic Design[J]. IEEE/CAA Journal of Automatica Sinica, 2025, 12(4): 719-733. doi: 10.1109/JAS.2024.124947
    [2]Mengli Wei, Wenwu Yu, Duxin Chen, Mingyu Kang, Guang Cheng. Privacy Distributed Constrained Optimization Over Time-Varying Unbalanced Networks and Its Application in Federated Learning[J]. IEEE/CAA Journal of Automatica Sinica, 2025, 12(2): 335-346. doi: 10.1109/JAS.2024.124869
    [3]Tai-You Chen, Xiao-Min Hu, Qiuzhen Lin, Wei-Neng Chen. Multi-Agent Swarm Optimization With Contribution-Based Cooperation for Distributed Multi-Target Localization and Data Association[J]. IEEE/CAA Journal of Automatica Sinica. doi: 10.1109/JAS.2025.125150
    [4]Qi Deng, Qi Kang, MengChu Zhou, Xiaoling Wang, Shibing Zhao, Siqi Wu, Mohammadhossein Ghahramani. Evolutionary Algorithm Based on Surrogate and Inverse Surrogate Models for Expensive Multiobjective Optimization[J]. IEEE/CAA Journal of Automatica Sinica.
    [5]Ke Chen, Wenjie Wang, Fangfang Zhang, Jing Liang, Kunjie Yu. Correlation-Guided Particle Swarm Optimization Approach for Feature Selection in Fault Diagnosis[J]. IEEE/CAA Journal of Automatica Sinica. doi: 10.1109/JAS.2025.125306
    [6]Fei Ming, Wenyin Gong, Ling Wang, Yaochu Jin. Constrained Multi-Objective Optimization With Deep Reinforcement Learning Assisted Operator Selection[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(4): 919-931. doi: 10.1109/JAS.2023.123687
    [7]Jiufang Chen, Kechen Liu, Xin Luo, Ye Yuan, Khaled Sedraoui, Yusuf Al-Turki, MengChu Zhou. A State-Migration Particle Swarm Optimizer for Adaptive Latent Factor Analysis of High-Dimensional and Incomplete Data[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(11): 2220-2235. doi: 10.1109/JAS.2024.124575
    [8]Sheng Qi, Rui Wang, Tao Zhang, Xu Yang, Ruiqing Sun, Ling Wang. A Two-Layer Encoding Learning Swarm Optimizer Based on Frequent Itemsets for Sparse Large-Scale Multi-Objective Optimization[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(6): 1342-1357. doi: 10.1109/JAS.2024.124341
    [9]Jing Liang, Hongyu Lin, Caitong Yue, Ponnuthurai Nagaratnam Suganthan, Yaonan Wang. Multiobjective Differential Evolution for Higher-Dimensional Multimodal Multiobjective Optimization[J]. IEEE/CAA Journal of Automatica Sinica, 2024, 11(6): 1458-1475. doi: 10.1109/JAS.2024.124377
    [10]Wenhua Li, Xingyi Yao, Kaiwen Li, Rui Wang, Tao Zhang, Ling Wang. Coevolutionary Framework for Generalized Multimodal Multi-Objective Optimization[J]. IEEE/CAA Journal of Automatica Sinica, 2023, 10(7): 1544-1556. doi: 10.1109/JAS.2023.123609
    [11]Guijun Ma, Zidong Wang, Weibo Liu, Jingzhong Fang, Yong Zhang, Han Ding, Ye Yuan. Estimating the State of Health for Lithium-ion Batteries: A Particle Swarm Optimization-Assisted Deep Domain Adaptation Approach[J]. IEEE/CAA Journal of Automatica Sinica, 2023, 10(7): 1530-1543. doi: 10.1109/JAS.2023.123531
    [12]Zhenyu Lei, Shangce Gao, Zhiming Zhang, Haichuan Yang, Haotian Li. A Chaotic Local Search-Based Particle Swarm Optimizer for Large-Scale Complex Wind Farm Layout Optimization[J]. IEEE/CAA Journal of Automatica Sinica, 2023, 10(5): 1168-1180. doi: 10.1109/JAS.2023.123387
    [13]Jun Tang, Gang Liu, Qingtao Pan. A Review on Representative Swarm Intelligence Algorithms for Solving Optimization Problems: Applications and Trends[J]. IEEE/CAA Journal of Automatica Sinica, 2021, 8(10): 1627-1643. doi: 10.1109/JAS.2021.1004129
    [14]Haowei Lin, Bo Zhao, Derong Liu, Cesare Alippi. Data-based Fault Tolerant Control for Affine Nonlinear Systems Through Particle Swarm Optimized Neural Networks[J]. IEEE/CAA Journal of Automatica Sinica, 2020, 7(4): 954-964. doi: 10.1109/JAS.2020.1003225
    [15]Jiahai Wang, Yuyan Sun, Zizhen Zhang, Shangce Gao. Solving Multitrip Pickup and Delivery Problem With Time Windows and Manpower Planning Using Multiobjective Algorithms[J]. IEEE/CAA Journal of Automatica Sinica, 2020, 7(4): 1134-1153. doi: 10.1109/JAS.2020.1003204
    [16]Yong Zhang, Ping Zhou, Guimei Cui. Multi-Model Based PSO Method for Burden Distribution Matrix Optimization With Expected Burden Distribution Output Behaviors[J]. IEEE/CAA Journal of Automatica Sinica, 2019, 6(6): 1506-1512. doi: 10.1109/JAS.2018.7511090
    [17]Jiajun Wang, Tufan Kumbasar. Parameter Optimization of Interval Type-2 Fuzzy Neural Networks Based on PSO and BBBC Methods[J]. IEEE/CAA Journal of Automatica Sinica, 2019, 6(1): 247-257. doi: 10.1109/JAS.2019.1911348
    [18]Pratik Roy, Ghanshaym Singha Mahapatra, Kashi Nath Dey. Forecasting of Software Reliability Using Neighborhood Fuzzy Particle Swarm Optimization Based Novel Neural Network[J]. IEEE/CAA Journal of Automatica Sinica, 2019, 6(6): 1365-1383. doi: 10.1109/JAS.2019.1911753
    [19]Haibin Duan, Pei Li, Yaxiang Yu. A Predator-prey Particle Swarm Optimization Approach to Multiple UCAV Air Combat Modeled by Dynamic Game Theory[J]. IEEE/CAA Journal of Automatica Sinica, 2015, 2(1): 11-18.
    [20]Yanqiong Zhang, Youcheng Lou, Yiguang Hong. An Approximate Gradient Algorithm for Constrained Distributed Convex Optimization[J]. IEEE/CAA Journal of Automatica Sinica, 2014, 1(1): 61-67.

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(4)  / Tables(7)

    Article Metrics

    Article views (1577) PDF downloads(124) Cited by()

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return