1. Introduction
Multi-objective optimization problems (MOPs)—i.e., problems where several objectives have to be optimized concurrently –naturally arise in many applications (e.g., [1,2,3,4]). As an example, in many portfolio problems, one is interested in maximizing the expected return and social responsibility or sustainability while minimizing the risk to a financial portfolio ([5,6]). In multi-objective optimization, we distinguish between the decision space, which contains the vectors of decision variables, and the objective space, which is the k dimensional real vectors and comprises the images of the vector-valued objective function. A typical approach to the solution of MOPs is to compute or approximate the non-dominated (or efficient) set with respect to the Pareto dominance order (the image of which in the objective space is called the Pareto front). One important characteristic of (continuous) MOPs is that in regular cases, the Pareto front is a manifold of dimensions, where k denotes the number of objective functions. In general, it is possible that parts of the Pareto front are of lower dimension, but the Pareto front is never more than dimensions. Since, in the continuous case, the non-dominated set and the Pareto front can contain infinitely many points, it is usually approximated by a finite set of points. In particular, in the area of evolutionary multi-objective optimization (EMO), many performance indicators have been proposed that propagate optimal approximations of the Pareto front (e.g., [7,8,9,10]). While their definitions slightly differ, most have in mind to obtain (more or less) evenly spread solutions along the Pareto front.
Interestingly, with the hypervolume indicator [10], there exists an indicator that does not require the knowledge of the location of the true Pareto front. Still, its maximization leads to well-distributed approximation sets consisting of only non-dominated solutions. In this work, by “well-distributed”, we mean the objective points have good coverage of the Pareto front and are gap-free when the population size is large. At the maximum of the hypervolume indicator, the density of objective points is inversely proportional to the local curvature of the Pareto front [11]. Here is where the idea of set-scalarization comes into play. In set-scalarization methods, rather than focusing on the improvement of single points of the approximation set, the focus is on the optimization of a fixed cardinality set as an entity concerning a set-based indicator, e.g., the hypervolume indicator. The objective function of the set-scalarization method, in our case, the hypervolume indicator, provides a mapping from the set of fixed cardinality sets in the decision space to a scalar that has to be maximized. Due to the properties mentioned above of the hypervolume indicator, the resulting set will provide a well-distributed set of points on the Pareto front.
Multi-objective evolutionary algorithms (MOEAs) have long since adopted the idea of set-scalarization. The so-called indicator-based MOEAs (e.g., [12,13,14]) use performance indicators to guide the search, e.g., by indicator-based selection. In numerical methods, the set scalarization approach was first addressed in gradient-based hypervolume maximization [15,16,17,18,19] and in the maximization of the Averaged Hausdorff Metric [14]). More recently, the approach was generalized to second-order methods with the Hypervolume Newton Method (HVN), a set scalarization-based Newton–Raphson method for the maximization of the hypervolume indicator value of a given MOP (e.g., [20,21]). However, this method has only been discussed for unconstrained and bi-objective optimization problems, which limits its application.
In this paper, we extend the HVN for constrained MOPs with a general number of objectives. To this end, we present the HVN for equality-constrained problems and further discuss a straightforward active set method to handle inequalities. Since the HVN is highly local, we also discuss the hybridization of this method with an MOEA. Finally, we present numerical results indicating the strength of the novel approaches.
The remainder of this paper is organized as follows: in Section 2, we present the necessary background required for understanding the sequel, and we review the related work. In Section 3, we present the HVN for constrained multi-objective optimization problems. Section 4 presents the numerical results of the constrained HVN as a standalone algorithm and a local search strategy within a hybrid evolutionary algorithm. Finally, we conclude and give possible paths for future research in Section 5.
2. Background and Related Work
2.1. Notations
We will always denote a finite Pareto approximate set by . When differentiating a set function, e.g., the hypervolume, over the input set, we often concatenate the points in into a much longer vector, i.e., . To make our discussion less cumbersome, we abuse the notation slightly such that it can be interpreted as a finite set in or an -vector, depending on the context. (See [16] for a detailed formal discussion of the mapping between fixed cardinality sets and vectors.) We will explain the meaning of on the spot whenever it is unclear from the text. We will always denote by ∇ and the gradient/Jacobian and Hessian operators on real-valued functions, respectively, when the domain of such a function is clear from the text. Otherwise, we take the derivative operator . When expressing the Hessian matrix, we will use the numerator layout for matrix calculus notations [22].
2.2. Multi-Objective Optimization
A real-valued multi-objective optimization problem (MOP) involves minimizing multiple objective functions simultaneously, i.e., , . For every and , we say weakly dominates (written as ) iff , . The Pareto order ≺ on is defined: iff. and . A point is called efficient or (Pareto) optimal iff. . The set of all Pareto optimal solutions of a MOP is called the Pareto set, and its image is called the Pareto front. Typically, i.e., under certain (mild) assumptions on the model, one can assume that the Pareto set and front of a given continuous MOP form at least locally an object of dimension ([23]).
The Pareto order can also be extended to the family of sets [10], i.e., we say iff. . The set of all efficient points of is called the efficient set. The image of the efficient set under F is called the Pareto front. Multi-objective optimization algorithms (MOAs) often employ a finite multiset to approximate the efficient set, whose image under F is denoted by . Multi-objective optimization is an active research field that has produced many algorithms for the approximation of the entire Pareto set/front of a given MOP. There exist, for instance, scalarization methods, and mathematical programming techniques that transform the given MOP into an auxiliary scalar optimization problem (SOP) (e.g., [24]). Via solving a clever sequence of such SOPs, one can obtain in many cases suitable Pareto front approximations (e.g., [25,26,27,28]). In [29], a Newton method is proposed for multi-objective optimization. Next to these point-wise iterative local search strategies there exist global set-based algorithms such as cell-to-cell mapping techniques and subdivision techniques ([30,31,32]) as well as specialized evolutionary algorithms ([33,34,35,36]). There exist in particular indicator-based evolutionary algorithms (IBEAs) that aim for Pareto front approximations of a given performance indicator (e.g., [12,13,14]). Widely used performance indicators are the Generational Distance (GD [7]), the Inverted Generational Distance and variants ([8,37,38]), the averaged Hausdorff distance ([9,39,40]), and the Hypervolume indicator, which we will use in this work and briefly review in the next section.
Finally, there exist multi-objective continuation methods that make use of the fact that the solution set forms at least locally a manifold (e.g., [23,41,42,43,44,45,46]).
2.3. Hypervolume Indicator and Its First-Order Derivatives
The hypervolume indicator (HV) [10,47] is defined as the Lebesgue measure of the compact set dominated by a Pareto approximation set and cut from above by a reference point :
where denotes the Lebesgue measure in . HV is Pareto compliant, i.e., for all , , and is extensively used to assess the quality of approximation sets to the Pareto front, e.g., in SMS-MOEA [12] and multi-objective Bayesian optimization [48]. Being a set function, it is cumbersome to define the derivative of HV. (The derivative of a set function is not defined for an arbitrary family of sets. For some special cases, it can be defined directly, e.g., on Jordan-measurable sets [49].) Therefore, we follow the generic set-based approach for MOPs [16], which considers a finite approximation sets of size vectors as a point in , i.e., . Similarly, the image of under F can also be represented by a -vector: . In this sense, the objective function F is also extended as follows:Taking , we can express the hypervolume indicator as a function on :
We will henceforth omit the reference point in for simplicity. It is straightforward to express the gradient of with respect to using the chain rule as reported in our previous works [16,19]: , in which we also discussed the time complexity of computing the hypervolume gradient. It is noted here that an alternative to the computation of the gradient of the entire set, it was also suggested to compute only the gradient of a single point with respect to the hypervolume indicator; this approach is referred to as hypervolume scalarization [50].
2.4. Hypervolume Hessian and Hypervolume Newton Method
Here, we assume F is at least twice continuously differentiable. In general, the Hessian matrix of the hypervolume indicator can be expressed as follows:
(1)
Note that in the above expression, and denote the Hessian matrix of the hypervolume indicator with respect to objective points and of the objective function , respectively. In our previous work [21], we derived the analytical expression of for bi-objective cases and analyzed the structure and properties of the hypervolume Hessian matrix. In addition, we implemented a standalone hypervolume Newton (HVN) algorithm for unconstrained MOPs. Moreover, we have shown that the Hessian is a tridiagonal block matrix in bi-objective cases and provided the non-singularity condition thereof, which states the Hessian is only singular on a null subset of [21], thereby ascertaining the safety of applying the HVN method.
The analytical expression of the Hessian matrix for higher dimensions contains the derivatives . To compute these derivatives analytically, the chain rule can be applied (see [21]). In [21]. However, the Hessian matrix of the second mapping—from the points in the objective space to the hypervolume indicator—was only given analytically for two dimensions. The Hessian matrix of this second mapping can be generalized to k dimensional objective spaces, and it is continuous in regular cases. Here, we will only sketch the construction of this matrix and leave the detailed analysis for future research. It is known that in the N-dimensional case, the first derivatives are given by the -dimensional Lebesgue measure of the dimensional faces of the attainment surface that separates the dominated space from the non-dominated space (see Figure 1, ). These faces themselves have a derivative that is given by the -dimensional Lebesgue measure of the -dimensional segments (or patches) at the boundary of these faces, which are also changing continuously with (see Figure 1, examples and ). Note that points in the objective space need to be in a general position to guarantee differentiability; otherwise, one-sided differentiability applies and one of the two derivatives, i.e., when the derivative with perturbed coordinate falls to the dominated subspace, it is always zero [16].
In this work, however, rather than investigating in detail the analytical and computational properties of the Hessian for more than two objective functions, we compute the second-order derivative with the automatic differentiation (AD) method [51] and focus on solving equality-constrained MOPs using the Hessian matrix of the hypervolume indicator.
3. Hypervolume Newton Method for Constrained MOPs
In this section, we first describe the base method of HVN for the treatment of equality constrained MOPs and will then discuss how to deal with inequalities and with dominated points that may be computed during the run of the Newton method.
3.1. Handling Equalities
Consider a continuous equality-constrained MOP of the form
(2)
where , and , , being the i-th equality constraint. The objective map is defined by , where is the i-the individual objective to be considered in the MOP. The feasible set is given by:(3)
The set (population) based hypervolume optimization problem we are considering in this work is the following one:
(4)
where denotes the value of the hypervolume for a given set of magnitude , where each . Note that the set can be interpreted as a point in (via considering , and hence, problem (4) can be identified by a scalar objective optimization problem of dimension .The feasibility of X (i.e., ) is identical to
(5)
For the related set-based equality constraints, we define for and
(6)
For checking the feasibility of all decision points, we define via
(7)
then its Jacobian is given by(8)
where(9)
The Karush-Kuhn-Tucker (KKT) equations of the problem (4) hence read as
(10)
for a Lagrange multiplier (or the dual variable) which directly leads to the root finding problem(11)
where . The Jacobian of G at is given by(12)
where(13)
Denoting by and , the variables in iteration t, a Newton step for problem (11) is given by
(14)
In our computations, we have omitted M in . A Newton step is hence obtained by solving
(15)
3.2. Handling Inequalities
In order to handle inequalities, we have chosen an active set approach which we will discuss in the following. This approach is straightforward; however, it has led to satisfying results in our computations, in particular when the initial candidate set was computed by the evolutionary algorithm.
Assume problem (2) contains inequalities of the form
(16)
where and , , is the i-th inequality constraint. Analogous to the equality-constrained case, we define the feasibility of by(17)
Define for and
(18)
and by(19)
The active set we have used is as follows: if for an inequality constraint it holds
(20)
for a given tolerance at , then we impose the equality(21)
(i.e., it will be added to the set of equalities) while all other inequalities are disregarded at .3.3. Handling Dominated Points
Since Newton’s method tends to realize relatively longer steps, it often occurs that some decision points are dominated after a Newton step/iteration. Therefore, it is necessary to discuss how the equality-constrained HVN method behaves in this case. For the reason that will become clear during our discussion, we will investigate two scenarios: (1) infeasible and dominated points and (2) feasible but dominated points.
For the first scenario, we consider the simplest case, where and there is only one dominated point. Without loss of generality, we can assume that for an approximation set , is dominated by at least one of the remaining points (as the indices are assigned to arbitrarily). Denoting by the approximation set after removing , we can express the constraint function on as:
Note that we are only considering the special case of one constraint, i.e., . The root finding problem G can re-expressed in the following form, equivalent to Equation (11):
Let and , we express the derivative of G as a block matrix:
Note that the upper left block equals . The inverse of can be obtained by applying the Schur complement recursively (first consider the block partition indicated above and then apply it again to each partition), provided that both and are non-singular.
After simplification, the inverse of admits the following form:
where , andThe first row of blocks is of particular interest to us since it determines the search step of . It is obvious that
(22)
where notation takes rows from 1 to n and columns from 1 to in matrix . Similarly, the search step of the dual variable is:(23)
Now, consider the function , whose first- and second-order derivatives are:
The global minimum/maximum of corresponds to the feasible set, i.e., . Hence, Newton iterations that optimize will equivalently find the feasible set. Computing the Newton direction of , we have:
(24)
Setting in the above equation and comparing it to Equation (22), we notice that the Newton direction of and the hypervolume Newton step only differ by a scalar, which can be neglected in practice since we implement a step-size control to re-scale the search step (see the following sub-section). Therefore, we conclude that for infeasible and dominated points, our HVN method (Equation (15)) only considers the constraint function and moves such decision points to the feasible set rapidly (ideally at quadratic speed when the point is close to the feasible set). This satisfactory property allows for handling infeasible and dominated points without modifying our HVN method.
In addition, due to this nice property, an infeasible point will eventually lie on the feasible set, where it can still be dominated if other feasible points exist. This is precisely the second scenario of our discussion, in which the hypervolume of feasible but dominated points will be zero. To move such points, we propose to employ the famous non-dominated sorting [36] procedure, where we partition all feasible points into “layers” of mutually non-dominated ones (formally, anti-chains of Pareto order) and compute the Newton direction for each layer (using Equation (15)) regardless of other dominating layers. In this manner, the HVN method can move all feasible points along the feasible set for achieving a good distribution.
3.4. The HVN Method for Constrained MOPs
Taking the above considerations regarding the HVN method, in this section, we aim to devise and implement a standalone HVN algorithm, which is outlined in Algorithm 1. First, we check if any decision point is feasible (i.e., for some ), where the feasibility can be tested numerically with a pre-defined small threshold (e.g., used in this work) for the equality constraints. Then, we employ the non-dominated sorting point procedure [36] to partition the feasible points into “layers” of mutually non-dominated ones, where the Newton direction (Equation (15)) is calculated separately on each layer. Taking L for the indices of points in a layer and for the subset indexed by L, we express this partitioning as . Note that the dominance relation for the remaining infeasible and dominated points is not well-defined, considering the equality constraints since they are incomparable to the feasible ones (also among themselves). In this case, we simply merge them into the first layer and compute the Newton direction thereof, which can be justified by the observation in Equations (22) and (24). The resulting search direction of the infeasible and dominated points is a Newton direction of the function . In this treatment, a special case arises when there are no feasible points, usually in the first several iterations of the algorithm.
Algorithm 1: Standalone hypervolume Newton algorithm for equality-constrained MOPs |
Finally, another important aspect is the step-size control for each Newton step. We propose maintaining individual step-sizes for each partition, which is determined using the well-known Armijo’s backtracking line search [52]. In detail, this method starts with an initial step-size and tests whether the Euclidean norm of has sufficiently decreased after applying the Newton step to the primal-dual pair . Since Newton’s direction for equality-constrained problems (Equation (15)) is not necessarily an ascent direction for the hypervolume, we take the Euclidean norm as the convergence measure since (1) the optimality condition is (Equation (11)) and (2) the Newton step is always a descent direction of . Let be the primal-dual variable and , then we have . If the test fails, then we halve the step-size and repeat the test. Notably, for infeasible and dominated points, the test checks if the value of the squared constraint value is sufficiently decreased as the HVN method computes the Newton direction of for those points. In our implementation, we use maximally six iterations of such tests, resulting in a minimal step-size of . As for the initial step-size , the commonly used value often leads to Newton steps that jump out of the decision space when the Newton direction is large or the point is in the vicinity of the decision boundary. Therefore, we set it to the minimum of one and the maximal step-size that the primal vector can take without leaving the decision space, i.e., . The value of can be calculated in a straightforward way when the decision space is a convex and compact subset of , e.g., a hyperbox.
3.5. Computational Cost
The above method requires the knowledge of the Jacobian and the Hessian of both objective and constraint functions. In this work, we have used automatic differentiation (AD) techniques [53]. Note that finite differences can also be utilized when the AD-computation is not applicable. The AD-computation takes maximally four times the used multiply–add operations taken in evaluating the function value [54]. Hence, to make a fair comparison between HVN and MOEA methods, we will take 4 function evaluations (FEs) and FEs to quantify the computational cost of each AD-computed Jacobian and Hessian, respectively. In total, the number of FEs consumed in each iteration comprises:
which amounts to computations of function evaluation, constraint evaluation, Jacobian of the objective function and the constraint, Hessian of the objective function and the constraint, and the backtracking line search of the step-size. Computing the hypervolume Hessian takes time in addition to the AD-computation of derivatives in Equation (1). For solving Equation (15), we use Cholesky decomposition, which has a computational complexity of . It is certainly desired either to have an analytic expression of the HV Hessian or to exploit the block diagonal structure this matrix will certainly have for AD, which we, however, have to leave for future research.We have implemented the standalone algorithm in Python, which is accessible at
4. Numerical Results
In this section, we present some numerical results of the HVN both as standalone algorithms as well as a local search engine within the NSGA-III algorithm.
4.1. HVN as Standalone Algorithm
We showcase the behavior of the proposed Newton method as a standalone method on three example problems:
Importantly, we will use different initializations of the decision points that are specific to each problem in order to investigate the behavior of the standalone HVN with respect to the characteristic of each problem; We do not aim to provide a unified and systematic initialization method for the standalone HVN in this section. Note that problem P3 defines an inequality constraint on the first component of the decision point, where the feasible set is , and the optimum lies on the active set of g, i.e., . This problem is meant to test if the proposed HVN algorithm can manage to solve inequality-constrained problems where the optimum is on the active set of the constraint. To measure the empirical performance of the HVN algorithm, we take the Euclidean norm since the Newton direction is not necessarily an ascent direction for the hypervolume.
Moreover, since it is well-known that the Newton-like method can be affected by choice of initial solutions, we investigate the performance of the HVN algorithm on problem P1 with three different initializations. Specifically, in the two-dimensional decision space, we create initial decision points on the line segment , where we determine the value of by (i) taking evenly spaced points (linear), (ii) logistic-transformed evenly spaced points (which makes the points denser around the tails of the line segment), or (iii) logit-transformed evenly spaced points (higher density of points in the middle). The results are illustrated in Figure 2 and Table 1, which shows a set of well-distributed points on the feasible set in the objective space (the red dashed sphere) for all three initializations. In addition, the empirical convergence rate is quadratic regardless of the choice of initialization methods, as reported in Table 1.
The results on problem P2 are depicted in Figure 3 and Table 2 for three different sizes of the approximation set. The initial decision points are sampled uniformly at random in the convex hull of three points , and . Whereas the final approximation set is well-distributed in the objective space, we observe that empirical convergence of is considerably rugged in the first 20–25 iterations, after which quadratic convergence appears. This is indeed attributed to the fact that decision points often become dominated in the first couple of iterations on this problem, resulting in zero hypervolume gradient thereof and hence quite a large norm of . Nevertheless, the proposed treatment of those dominated points (Algorithm 1), which is based on the non-dominated sorting procedure, is capable of bringing the dominated points to the active set with a quadratic speed. Similarly, the same ruggedness is seen in the convergence chart of problem P3 (shown in Figure 4). On this problem, we again take the setting , and the initial decision points are sampled uniformly at random in the feasible space of . We extend the HVN algorithm slightly for this inequality-constrained problem in the following way: whenever the decision points are feasible, i.e., , and quite distant from the active set (, shown as the red plane in Figure 4), we ignore the constraint function when computing the Newton step. When the feasible decision points are sufficiently close to the active set (the distance is less than in our implementation), we consider an equality constraint and utilize Equation (15) to compute the Newton step.
Moreover, we test the standalone HVN method on large-scale, complicated MOPs. We choose the well-known DTLZ problems with one spherical constraint [55,56] with decision points, resulting in a relatively large Hessian matrix (for an 11-dimensional decision space and one constraint, the object is of size ). In this case, we use sparse matrix operations for computation efficiency, exploiting the sparsity of the Hessian. Since the DTLZ problems are highly multi-modal, the initial approximation set is generated in a local vicinity of the Pareto set, i.e., , where is sampled uniformly at random on the Pareto set. We execute the standalone HVN method for 15 iterations and illustrate the result in Figure 5. In the plot, we observe well-distributed final points (green dots) in contrast to non-uniform initial ones (black crosses), showing the standalone HVN works properly as a local method for large-scale problems.
4.2. HVN within NSGA-III
In this section, we investigate the empirical performance of the HVN algorithm on more complicated, equality-constrained DTLZ (Eq-DTLZ) problems [55,56] and their inverted counterparts (Eq-IDTLZ). As Newton-like algorithms are local methods, running the standalone algorithm (Algorithm 1) will stagnate at local Pareto sets. Therefore, we hybridize the HVN algorithm with an MOEA, in which we first execute the MOEA for a pre-defined budget to overcome the local optimum and get close to the global Pareto set, and then initialize the HVN algorithm from the final approximation set of the MOEA to make local refinements. We summarize this hybrid approach in Algorithm 2. Notably, in line 3, we transfer the whole approximation set (rather than only the non-dominated points) to HVN upon the termination of MOEA since the standalone HVN method is able to move dominated points towards the feasible set at quadratic speed, as proven in Section 3.3.
Algorithm 2: Hybridization of HVN and MOEA |
The following empirical study aims to check whether the hybridization approach can achieve a better final approximation set/front than an MOEA alone under the same computation budget. As for the test problem, a single spherical constraint is imposed on problems DTLZ. The decision space is , the reference point is for HVN, and the approximation set is of size 200. Here, we choose the well-known NSGA-III algorithm [34,35], where the equality constraints are handled using the adaptive -constraint handling technique. We utilize the implementation in the
We first depict one example of the final approximation set (only the non-dominated subset is shown) in Figure 6 for both methods, where we clearly observe that the hybridization achieves much more non-dominated points than NSGA-III. Second, we show, in Table 3, the hypervolume indicator value and the number of final non-dominated points for both algorithms obtained from 15 independent runs. In addition, we compute the above metrics for the hybrid algorithm right before the HVN phase starts (NSGA-III (1000) in the table), showing the progress that HVN manages to make. From the results, we conclude that the hybrid algorithm significantly improves upon the hypervolume metric and outputs substantially more non-dominated points than NSGA-III alone. We conjecture that the observed advantage of the hybrid algorithm is very likely attributed to HVN’s ability to move dominated points to the feasible set with quadratic convergence (see Section 3), which disregards the objective function and thereby its multi-modal landscape.
5. Conclusions
In this paper, we propose a hypervolume Newton method for equality-constrained multi-objective optimization problems (MOPs) under the assumption that both the objective and the constraint functions are twice continuously differentiable. Based on previous works on set-oriented hypervolume Hessian matrix and hypervolume Newton (HVN) method for unconstrained MOPs, we propose, in this paper, the generalization of the HVN for equality-constrained problems and also elaborate a treatment for inequality-constrained based on an active set approach, which regards an inequality function as equality if the constraint values are within some small tolerance. In addition, we devised and tested two resulting algorithms: the standalone HVN method as an efficient local optimizer and a hybridization of the HVN and an MOEA for solving complicated and multi-modal MOPs. Moreover, in detail, we discuss the search direction for dominated points obtained from the set-oriented Newton step in which we prove that for dominated and infeasible points, the computed search step is the Newton step of the squared equality constraint function. Therefore, our HVN method can efficiently steer the non-dominated and dominated decision points.
We first illustrate the empirical behavior of the standalone algorithm on three simple MOPs, where we observe quadratic convergence of the two-norm of the root finding problem G. Then, on highly multi-modal DTLZ problems with one spherical constraint (Eq-DTLZ), we tested the local convergence of the standalone HVN algorithm with a relatively large approximation set () by initializing the approximation set in the neighborhood around the Pareto set, which shows a fast convergence to well-distributed points on the feasible set. Finally, we benchmark the hybrid algorithm against NSGA-III on Eq-DTLZ1-4 and Eq-IDTLZ1-4 problems, in which we observe that with roughly the same computational budget, the hybrid algorithm achieves substantially more non-dominated points in the final population, which leads to significantly higher hypervolume values. We conjecture that such an advantage is attributed to (1) the fast local convergence of the HVN method and (2) HVN’s ability to move infeasible and dominated points.
For future works, we contemplate (1) testing the hybridization of the HVN method with other EMOAs for more than three objectives, e.g., SMS-EMOA, to investigate the benefit of the HVN method in a broader setup; (2) comparing the hybrid HVN method to other state-of-the-art algorithms, e.g., MOEA/D (decomposition-based), EHVI-EGO (Bayesian optimization), or the average Hausdorff distance-based Newton method (mathematical optimization) on complex, or even real-world MOPs with multiple non-linear constraint functions; (3) investigating the analytical expression (as sketched in Figure 1) and computation of the hypervolume Hessian matrix, which can reduce the computation cost of the HVN method; (4) devising generic methodologies to handle inequality constraints for the HVN method, which will make it more applicable in practice; (5) extending the HVN to methods that provide non-zero sub-gradients for dominated points as in [17,18]; and (6) incorporating a surrogate-assisted method for tackling high-dimensional and complex problems, e.g., as in [57].
Conceptualization, O.S., H.W., A.D., M.E. and V.A.S.H.; methodology, O.S. and H.W.; software, M.E. geometrical analysis, and visualization, H.W. and V.A.S.H.; validation, H.W., M.E. and O.S.; formal analysis, H.W., O.S. and M.E.; investigation, H.W. and O.S.; resources, O.S.; data curation, H.W.; writing—original draft preparation, all; writing—review and editing, all; visualization, H.W.; supervision, O.S. and M.E.; project administration, O.S. All authors have read and agreed to the published version of the manuscript.
We have hosted all the data sets of this work on Zenodo:
We dedicate this work to Kalyanmoy Deb for his pioneering, inspiring, and fundamental contributions to the evolutionary multi-objective optimization (EMO) community, and particularly for his famous non-dominated sorting procedure, which plays a crucial role in this work in order to efficiently handle dominated points that can be generated throughout the Newton iteration.
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
The evolution of
Linear | Logistic | Logit | |
---|---|---|---|
1 | 4.23 |
4.55 |
4.20 |
2 | 2.33 |
2.54 |
2.27 |
3 | 8.81 |
1.01 |
8.52 |
4 | 8.19 |
7.82 |
8.30 |
5 | 2.29 |
2.17 |
2.29 |
6 | 1.06 |
8.77 |
1.11 |
7 | 1.91 |
3.48 |
1.93 |
8 | 7.38 |
7.05 |
1.03 |
9 | 1.76 |
1.06 |
1.55 |
10 | 1.62 |
1.79 |
2.33 |
The evolution of
Problem P2 | Problem P3 | |||||
---|---|---|---|---|---|---|
|
|
|
|
|
|
|
1 | 1.365 |
1.055 |
18.036836 | 1.433 |
569.121097 | 438.983791 |
2 | 9.454 |
9.259 |
16.083916 | 9.368 |
541.523806 | 434.703270 |
3 | 1.247 |
8.977 |
16.403643 | 1.197 |
444.774066 | 365.952443 |
4 | 1.589 |
8.628 |
18.052126 | 9.522 |
261.636562 | 362.014326 |
5 | 9.791 |
6.888 |
12.364802 | 6.194 |
212.841570 | 341.897644 |
6 | 8.618 |
1.123 |
3.899254 | 5.232 |
145.076665 | 254.253017 |
7 | 8.024 |
8.779 |
11.323440 | 3.557 |
103.986300 | 240.719767 |
8 | 4.737 |
7.632 |
13.320606 | 2.419 |
57.592159 | 165.603954 |
9 | 9.037 |
7.090 |
2.543622 | 1.511 |
12.628821 | 109.411195 |
10 | 7.393 |
1.816 |
5.984437 | 8.527 |
0.104307 | 70.516402 |
11 | 1.182 |
2.660 |
5.749496 | 3.732 |
0.097777 | 41.699152 |
12 | 4.399 |
2.877 |
0.702964 | 3.248 |
0.097013 | 19.525977 |
13 | 1.535 |
3.232 |
2.240449 | 2.008 |
0.096634 | 0.447690 |
14 | 2.299 |
3.694 |
13.274468 | 1.829 |
0.096256 | 0.257345 |
15 | 7.425 |
7.159 |
15.201915 | 5.425 |
0.005277 | 2.066379 |
16 | 1.572 |
1.378 |
11.273571 | 1.735 |
0.002934 | 2.016149 |
17 | 4.216 |
1.231 |
3.318978 | 6.372 |
0.001602 | 1.019636 |
18 | 1.630 |
5.630 |
2.818340 | 9.702 |
0.001552 | 0.944297 |
19 | 1.674 |
5.454 |
0.400360 | 9.373 |
0.001528 | 4.904926 |
20 | 1.733 |
5.411 |
0.335107 | 2.546 |
0.001522 | 2.937413 |
21 | 1.803 |
4.697 |
0.074058 | 7.243 |
0.001516 | 3.118031 |
22 | 1.761 |
5.901 |
0.081798 | 8.897 |
0.001139 | 0.336917 |
23 | 1.384 |
6.140 |
0.057776 | 6.654 |
0.001072 | 0.004270 |
24 | 1.020 |
4.508 |
0.029809 | 7.210 |
0.001010 | 0.000866 |
25 | 9.765 |
2.759 |
0.001956 | 5.851 |
0.000994 | 0.000489 |
26 | 9.788 |
7.794 |
0.081949 | 5.851 |
0.000990 | 0.000477 |
27 | 1.177 |
7.767 |
0.031275 | 5.851 |
0.000954 | 0.000459 |
28 | 1.176 |
7.688 |
0.000492 | 5.851 |
0.128140 | 0.000460 |
29 | 1.052 |
5.918 |
0.000053 | 5.851 |
0.252721 | 0.000460 |
30 | 1.314 |
6.821 |
0.000002 | 5.851 |
0.022233 | 0.000460 |
On Eq-DTLZ1-4 and Eq-IDTLZ1-4 problems, the sample mean and standard error of the hypervolume (HV) value and the number of final non-dominated (ND) points over 15 independent runs for each algorithm. The hypervolume values are computed with reference point
Eq-DTLZ1 | Eq-DTLZ2 | Eq-DTLZ3 | Eq-DTLZ4 | |||||
---|---|---|---|---|---|---|---|---|
Algorithm | HV | #ND | HV | #ND | HV | #ND | HV | #ND |
NSGA-III (1000) | 0.867 ± 1.4 |
28.4 ± 0.7 | 0.297 ± 1.9 |
32.7 ± 0.9 | 0.292 ± 1.9 |
26.0 ± 1.0 | 8.4 |
12.3 ± 0.8 |
Hybridization | 0.876 ± 2.4 |
80.9 ± 2.0 | 0.324 ± 3.6 |
95.3 ± 1.9 | 0.321 ± 6.6 |
75.2 ± 2.4 | 1.1 |
200.0 ± 0.0 |
NSGA-III (3400) | 0.873 ± 4.5 |
38.5 ± 1.3 | 0.304 ± 9.2 |
32.6 ± 0.9 | 0.301 ± 1.1 |
30.1 ± 0.7 | 9.2 |
14.5 ± 0.6 |
Eq-IDTLZ1 | Eq-IDTLZ2 | Eq-IDTLZ3 | Eq-IDTLZ4 | |||||
Algorithm | HV | #ND | HV | #ND | HV | #ND | HV | #ND |
NSGA-III (1000) | 0.517 ± 1.8 |
23.2 ± 0.5 | 3.224 ± 2.0 |
74.1 ± 1.2 | 1.5 |
81.7 ± 1.6 | 8.4 |
12.3 ± 0.8 |
Hybridization | 0.534 ± 1.5 |
112.1 ± 2.1 | 3.388 ± 1.7 |
198.2 ± 0.4 | 1.6 |
197.1 ± 0.4 | 1.1 |
200.0 ± 0.0 |
NSGA-III (3400) | 0.529 ± 2.9 |
33.4 ± 0.4 | 3.359 ± 4.7 |
88.3 ± 0.4 | 1.5 |
92.1 ± 0.8 | 9.2 |
14.5 ± 0.6 |
References
1. Stewart, T.; Bandte, O.; Braun, H.; Chakraborti, N.; Ehrgott, M.; Göbelt, M.; Jin, Y.; Nakayama, H. Real-World Applications of Multiobjective Optimization. Proceedings of the Multiobjective Optimization, Lecture Notes in Computer Science; Slowinski, R. Springer: Berlin/Heidelberg, Germany, 2008; Volume 5252, pp. 285-327.
2. Deb, K. Evolutionary multi-objective optimization: Past, present and future. Proceedings of the GECCO ’20: Proceedings of the 22th annual Conference on Genetic and Evolutionary Computation; Cancún, Mexico, 8–12 July 2020; pp. 343-372.
3. Aguilera-Rueda, V.J.; Cruz-Ramírez, N.; Mezura-Montes, E. Data-Driven Bayesian Network Learning: A Bi-Objective Approach to Address the Bias-Variance Decomposition. Math. Comput. Appl.; 2020; 25, 37. [DOI: https://dx.doi.org/10.3390/mca25020037]
4. Frausto-Solis, J.; Hernández-Ramírez, L.; Castilla-Valdez, G.; González-Barbosa, J.J.; Sánchez-Hernández, J.P. Chaotic Multi-Objective Simulated Annealing and Threshold Accepting for Job Shop Scheduling Problem. Math. Comput. Appl.; 2021; 26, 8. [DOI: https://dx.doi.org/10.3390/mca26010008]
5. Utz, S.; Wimmer, M.; Hirschberger, M.; Steuer, R.E. Tri-criterion inverse portfolio optimization with application to socially responsible mutual funds. Eur. J. Oper. Res.; 2014; 234, pp. 491-498. [DOI: https://dx.doi.org/10.1016/j.ejor.2013.07.024]
6. Estrada-Padilla, A.; Lopez-Garcia, D.; Gómez-Santillán, C.; Fraire-Huacuja, H.J.; Cruz-Reyes, L.; Rangel-Valdez, N.; Morales-Rodríguez, M.L. Modeling and Optimizing the Multi-Objective Portfolio Optimization Problem with Trapezoidal Fuzzy Parameters. Math. Comput. Appl.; 2021; 26, 36. [DOI: https://dx.doi.org/10.3390/mca26020036]
7. Van Veldhuizen, D.A. Multiobjective Evolutionary Algorithms: Classifications, Analyses, and New Innovations; Technical report Air Force Institute of Technology: Kaduna, Nigeria, 1999.
8. Coello, C.A.C.; Cortés, N.C. Solving Multiobjective Optimization Problems Using an Artificial Immune System. Genet. Program. Evolvable Mach.; 2005; 6, pp. 163-190. [DOI: https://dx.doi.org/10.1007/s10710-005-6164-x]
9. Schütze, O.; Esquivel, X.; Lara, A.; Coello, C.A.C. Using the averaged Hausdorff distance as a performance measure in evolutionary multiobjective optimization. IEEE Trans. Evol. Comput.; 2012; 16, pp. 504-522. [DOI: https://dx.doi.org/10.1109/TEVC.2011.2161872]
10. Zitzler, E.; Thiele, L.; Laumanns, M.; Fonseca, C.M.; da Fonseca, V.G. Performance assessment of multiobjective optimizers: An analysis and review. IEEE Trans. Evol. Comput.; 2003; 7, pp. 117-132. [DOI: https://dx.doi.org/10.1109/TEVC.2003.810758]
11. Auger, A.; Bader, J.; Brockhoff, D.; Zitzler, E. Theory of the hypervolume indicator: Optimal μ-distributions and the choice of the reference point. Proceedings of the Foundations of Genetic Algorithms, 10th ACM SIGEVO International Workshop, FOGA 2009; Orlando, FL, USA, 9–11 January 2009; Garibay, I.I.; Jansen, T.; Wiegand, R.P.; Wu, A.S. ACM: New York, NY, USA, 2009; pp. 87-102. [DOI: https://dx.doi.org/10.1145/1527125.1527138]
12. Beume, N.; Naujoks, B.; Emmerich, M.T.M. SMS-EMOA: Multiobjective selection based on dominated hypervolume. Eur. J. Oper. Res.; 2007; 181, pp. 1653-1669. [DOI: https://dx.doi.org/10.1016/j.ejor.2006.08.008]
13. Bader, J.; Zitzler, E. HypE: An Algorithm for Fast Hypervolume-Based Many-Objective Optimization. Evol. Comput.; 2011; 19, pp. 45-76. [DOI: https://dx.doi.org/10.1162/EVCO_a_00009]
14. Schütze, O.; Domínguez-Medina, C.; Cruz-Cortés, N.; de la Fraga, L.G.; Sun, J.Q.; Toscano, G.; Landa, R. A scalar optimization approach for averaged Hausdorff approximations of the Pareto front. Eng. Optim.; 2016; 48, pp. 1593-1617. [DOI: https://dx.doi.org/10.1080/0305215X.2015.1124872]
15. Emmerich, M.; Deutz, A.; Beume, N. Gradient-Based/Evolutionary Relay Hybrid for Computing Pareto Front Approximations Maximizing the S-Metric. Proceedings of the Hybrid Metaheuristics; Bartz-Beielstein, T.; Blesa Aguilera, M.J.; Blum, C.; Naujoks, B.; Roli, A.; Rudolph, G.; Sampels, M. Springer: Berlin/Heidelberg, Germany, 2007; pp. 140-156.
16. Emmerich, M.; Deutz, A.H. Time Complexity and Zeros of the Hypervolume Indicator Gradient Field. Proceedings of the EVOLVE—A Bridge between Probability, Set Oriented Numerics, and Evolutionary Computation III EVOLVE 2012; Mexico City, Mexico, 7–9 August 2012; Studies in Computational Intelligence Schuetze, O.; Coello, C.A.C.; Tantar, A.; Tantar, E.; Bouvry, P.; Moral, P.D.; Legrand, P. Springer: Berlin/Heidelberg, Germany, 2012; Volume 500, pp. 169-193. [DOI: https://dx.doi.org/10.1007/978-3-319-01460-9_8]
17. Wang, H.; Ren, Y.; Deutz, A.; Emmerich, M. On steering dominated points in hypervolume indicator gradient ascent for bi-objective optimization. NEO 2015; Springer: Berlin/Heidelberg, Germany, 2017; pp. 175-203.
18. Deist, T.M.; Maree, S.C.; Alderliesten, T.; Bosman, P.A. Multi-objective optimization by uncrowded hypervolume gradient ascent. Proceedings of the International Conference on Parallel Problem Solving from Nature; Springer: Berlin/Heidelberg, Germany, 2020; pp. 186-200.
19. Wang, H.; Deutz, A.H.; Bäck, T.; Emmerich, M. Hypervolume Indicator Gradient Ascent Multi-objective Optimization. Proceedings of the Evolutionary Multi-Criterion Optimization—9th International Conference, EMO 2017; Münster, Germany, 19–22 March 2017; Trautmann, H.; Rudolph, G.; Klamroth, K.; Schütze, O.; Wiecek, M.M.; Jin, Y.; Grimme, C. Springer: Berlin/Heidelberg, Germany, 2017; Volume 10173, pp. 654-669. [DOI: https://dx.doi.org/10.1007/978-3-319-54157-0_44]
20. Sosa Hernández, V.A.; Schütze, O.; Emmerich, M. Hypervolume maximization via set based Newton’s method. EVOLVE-A Bridge between Probability, Set Oriented Numerics, and Evolutionary Computation V; Springer: Berlin/Heidelberg, Germany, 2014; pp. 15-28.
21. Sosa-Hernández, V.A.; Schütze, O.; Wang, H.; Deutz, A.H.; Emmerich, M. The Set-Based Hypervolume Newton Method for Bi-Objective Optimization. IEEE Trans. Cybern.; 2020; 50, pp. 2186-2196. [DOI: https://dx.doi.org/10.1109/TCYB.2018.2885974] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30596593]
22. Petersen, K.B.; Pedersen, M.S. The matrix cookbook. Tech. Univ. Den.; 2008; 7, 510.
23. Hillermeier, C. Nonlinear Multiobjective Optimization: A Generalized Homotopy Approach; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2001; Volume 135.
24. Miettinen, K. Nonlinear Multiobjective Optimization; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 12.
25. Klamroth, K.; Tind, J.; Wiecek, M. Unbiased Approximation in Multicriteria Optimization. Math. Methods Oper. Res.; 2002; 56, pp. 413-437. [DOI: https://dx.doi.org/10.1007/s001860200217]
26. Fliege, J. Gap-free computation of Pareto-points by quadratic scalarizations. Math. Methods Oper. Res.; 2004; 59, pp. 69-89. [DOI: https://dx.doi.org/10.1007/s001860300316]
27. Eichfelder, G. Adaptive Scalarization Methods in Multiobjective Optimization; Springer: Berlin/Heidelberg, Germany, 2008.
28. Das, I.; Dennis, J.E. Normal-Boundary Intersection: A New Method for Generating the Pareto Surface in Nonlinear Multicriteria Optimization Problems. SIAM J. Optim.; 1998; 8, pp. 631-657. [DOI: https://dx.doi.org/10.1137/S1052623496307510]
29. Fliege, J.; Drummond, L.G.; Svaiter, B.F. Newton’s method for multiobjective optimization. SIAM J. Optim.; 2009; 20, pp. 602-626. [DOI: https://dx.doi.org/10.1137/08071692X]
30. Dellnitz, M.; Schütze, O.; Hestermeyer, T. Covering Pareto Sets by Multilevel Subdivision Techniques. J. Optim. Theory Appl.; 2005; 124, pp. 113-155. [DOI: https://dx.doi.org/10.1007/s10957-004-6468-7]
31. Hernández, C.; Naranjani, Y.; Sardahi, Y.; Liang, W.; Schütze, O.; Sun, J.Q. Simple Cell Mapping Method for Multi-objective Optimal Feedback Control Design. Int. J. Dyn. Control.; 2013; 1, pp. 231-238. [DOI: https://dx.doi.org/10.1007/s40435-013-0021-1]
32. Sun, J.Q.; Xiong, F.R.; Schütze, O.; Hernández, C. Cell Mapping Methods—Algorithmic Approaches and Applications; Springer: Berlin/Heidelberg, Germany, 2019.
33. Zhang, Q.; Li, H. MOEA/D: A Multi-objective Evolutionary Algorithm Based on Decomposition. IEEE Trans. Evol. Comput.; 2007; 11, pp. 712-731. [DOI: https://dx.doi.org/10.1109/TEVC.2007.892759]
34. Deb, K.; Jain, H. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints. IEEE Trans. Evol. Comput.; 2014; 18, pp. 577-601. [DOI: https://dx.doi.org/10.1109/TEVC.2013.2281535]
35. Jain, H.; Deb, K. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point Based Nondominated Sorting Approach, Part II: Handling Constraints and Extending to an Adaptive Approach. IEEE Trans. Evol. Comput.; 2014; 18, pp. 602-622. [DOI: https://dx.doi.org/10.1109/TEVC.2013.2281534]
36. Deb, K.; Agrawal, S.; Pratap, A.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput.; 2002; 6, pp. 182-197. [DOI: https://dx.doi.org/10.1109/4235.996017]
37. Ishibuchi, H.; Masuda, H.; Nojima, Y. A Study on Performance Evaluation Ability of a Modified Inverted Generational Distance Indicator; Association for Computing Machinery: New York, NY, USA, 2015; [DOI: https://dx.doi.org/10.1145/2739480.2754792]
38. Dilettoso, E.; Rizzo, S.A.; Salerno, N. A Weakly Pareto Compliant Quality Indicator. Math. Comput. Appl.; 2017; 22, 25. [DOI: https://dx.doi.org/10.3390/mca22010025]
39. Rudolph, G.; Schütze, O.; Grimme, C.; Domínguez-Medina, C.; Trautmann, H. Optimal averaged Hausdorff archives for bi-objective problems: Theoretical and numerical results. Comput. Optim. Appl.; 2016; 64, pp. 589-618. [DOI: https://dx.doi.org/10.1007/s10589-015-9815-8]
40. Bogoya, J.M.; Vargas, A.; Schütze, O. The Averaged Hausdorff Distances in Multi-Objective Optimization: A Review. Mathematics; 2019; 7, 894. [DOI: https://dx.doi.org/10.3390/math7100894]
41. Schütze, O.; Dell’Aere, A.; Dellnitz, M. On Continuation Methods for the Numerical Treatment of Multi-Objective Optimization Problems. Proceedings of the Practical Approaches to Multi-Objective Optimization; Number 04461 in Dagstuhl Seminar Proceedings Branke, J.; Deb, K.; Miettinen, K.; Steuer, R.E. Internationales Begegnungs- und Forschungszentrum (IBFI): Schloss Dagstuhl, Germany, 2005; Available online: http://drops.dagstuhl.de/opus/volltexte/2005/349 (accessed on 1 November 2022).
42. Martin, B.; Goldsztejn, A.; Granvilliers, L.; Jermann, C. On continuation methods for non-linear bi-objective optimization: Towards a certified interval-based approach. J. Glob. Optim.; 2014; 64, pp. 3-16. [DOI: https://dx.doi.org/10.1007/s10898-014-0201-3]
43. Martín, A.; Schütze, O. Pareto Tracer: A predictor-corrector method for multi-objective optimization problems. Eng. Optim.; 2018; 50, pp. 516-536. [DOI: https://dx.doi.org/10.1080/0305215X.2017.1327579]
44. Schütze, O.; Cuate, O.; Martín, A.; Peitz, S.; Dellnitz, M. Pareto Explorer: A global/local exploration tool for many-objective optimization problems. Eng. Optim.; 2020; 52, pp. 832-855. [DOI: https://dx.doi.org/10.1080/0305215X.2019.1617286]
45. Beltrán, F.; Cuate, O.; Schütze, O. The Pareto Tracer for General Inequality Constrained Multi-Objective Optimization Problems. Math. Comput. Appl.; 2020; 25, 80. [DOI: https://dx.doi.org/10.3390/mca25040080]
46. Bolten, M.; Doganay, O.T.; Gottschalk, H.; Klamroth, K. Tracing Locally Pareto-Optimal Points by Numerical Integration. SIAM J. Control. Optim.; 2021; 59, pp. 3302-3328. [DOI: https://dx.doi.org/10.1137/20M1341106]
47. Zitzler, E.; Thiele, L. Multiobjective Optimization Using Evolutionary Algorithms—A Comparative Case Study. Proceedings of the Parallel Problem Solving from Nature—PPSN V, 5th International Conference; Amsterdam, The Netherlands, 27–30 September 1998; Lecture Notes in Computer Science Eiben, A.E.; Bäck, T.; Schoenauer, M.; Schwefel, H. Springer: Berlin/Heidelberg, Germany, 1998; Volume 1498, pp. 292-304. [DOI: https://dx.doi.org/10.1007/BFb0056872]
48. Emmerich, M.; Yang, K.; Deutz, A.H.; Wang, H.; Fonseca, C.M. A Multicriteria Generalization of Bayesian Global Optimization. Advances in Stochastic and Deterministic Global Optimization; Pardalos, P.M.; Zhigljavsky, A.; Zilinskas, J. Springer: Berlin/Heidelberg, Germany, 2016; Volume 107, pp. 229-242. [DOI: https://dx.doi.org/10.1007/978-3-319-29975-4_12]
49. DiBenedetto, E.; Debenedetto, E. Real Analysis; Springer: Berlin/Heidelberg, Germany, 2002.
50. Paquete, L.; Schulze, B.; Stiglmayr, M.; Lourenço, A.C. Computing representations using hypervolume scalarizations. Comput. Oper. Res.; 2022; 137, 105349. [DOI: https://dx.doi.org/10.1016/j.cor.2021.105349]
51. Margossian, C.C. A Review of Automatic Differentiation and its Efficient Implementation. WIREs Data Mining Knowl. Discov.; 2019; 9, e1305. [DOI: https://dx.doi.org/10.1002/widm.1305]
52. Nocedal, J.; Wright, S.J. Numerical Optimization; Springer: Berlin/Heidelberg, Germany, 1999; [DOI: https://dx.doi.org/10.1007/b98874]
53. Baydin, A.G.; Pearlmutter, B.A.; Radul, A.A.; Siskind, J.M. Automatic Differentiation in Machine Learning: A Survey. J. Mach. Learn. Res.; 2017; 18, pp. 153:1-153:43.
54. Griewank, A.; Walther, A. Evaluating Derivatives—Principles and Techniques of Algorithmic Differentiation; 2nd ed. SIAM: Philadelphia, PA, USA, 2008; [DOI: https://dx.doi.org/10.1137/1.9780898717761]
55. Cuate, O.; Uribe, L.; Lara, A.; Schütze, O. A benchmark for equality constrained multi-objective optimization. Swarm Evol. Comput.; 2020; 52, 100619. [DOI: https://dx.doi.org/10.1016/j.swevo.2019.100619]
56. Cuate, O.; Uribe, L.; Lara, A.; Schütze, O. Dataset on a Benchmark for Equality Constrained Multi-objective Optimization. Data Brief; 2020; 29, 105130. [DOI: https://dx.doi.org/10.1016/j.dib.2020.105130]
57. Fu, C.; Wang, P.; Zhao, L.; Wang, X. A distance correlation-based Kriging modeling method for high-dimensional problems. Knowl. Based Syst.; 2020; 206, 106356. [DOI: https://dx.doi.org/10.1016/j.knosys.2020.106356]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Recently, the Hypervolume Newton Method (HVN) has been proposed as a fast and precise indicator-based method for solving unconstrained bi-objective optimization problems with objective functions. The HVN is defined on the space of (vectorized) fixed cardinality sets of decision space vectors for a given multi-objective optimization problem (MOP) and seeks to maximize the hypervolume indicator adopting the Newton–Raphson method for deterministic numerical optimization. To extend its scope to non-convex optimization problems, the HVN method was hybridized with a multi-objective evolutionary algorithm (MOEA), which resulted in a competitive solver for continuous unconstrained bi-objective optimization problems. In this paper, we extend the HVN to constrained MOPs with in principle any number of objectives. Similar to the original variant, the first- and second-order derivatives of the involved functions have to be given either analytically or numerically. We demonstrate the applicability of the extended HVN on a set of challenging benchmark problems and show that the new method can be readily applied to solve equality constraints with high precision and to some extent also inequalities. We finally use HVN as a local search engine within an MOEA and show the benefit of this hybrid method on several benchmark problems.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 Leiden Institute of Advanced Computer Science, Leiden University, 2333 CA Leiden, The Netherlands
2 School of Engineering and Sciences, Tecnológico de Monterrey, Av. Lago de Guadalupe Km 3.5, Atizapán de Zaragoza, Mexico City 52926, Mexico
3 Computer Science Department, Cinvestav-IPN, Mexico City 07360, Mexico