This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
The main goal herein is to derive an efficient optimization method for minimization of an objective function
In one of the first algorithms for solving unconstrained optimization problems, denoted as the steepest descent gradient method which is exposed by Cauchy, the iteration is defined as
Furthermore, in the general Newton method
In the quasi-Newton methods the Hessian of the goal function or its inverse is approximated by the adequately defined matrix. Using this type of methods we generally reduce the time of computations since we avoid the complicated calculations in deriving the Hessian of the objective function. Nevertheless, the methods of quasi Newton type preserve good properties of the Newton method. For these reasons, in this paper, we propose the method of quasi Newton type where the value of the iterative step size parameter
In the second section we give an overview of some accelerated gradient methods and hybrid iterations. We elaborate the deriving of the hybrid accelerated double direction method and restate the algorithm in the third section of this paper. In the fourth section we give a convergence analysis regarding the proposed iteration. Numerical experiments and comparison are presented in the last section of this paper.
2. Preliminaries: Accelerated Gradient Methods and Hybrid Iterations
The authors in [1] rightfully detected a class of accelerated gradient descent methods, defined by the general iterative scheme
An interesting concept of merging iterations through the hybrid expression was suggested in some research articles (see [6–8]). Some of representations are given by the next set of iterations:
In [9] it was proved that the hybrid method
3. HADD Algorithm
We are motivated by the confirmed advantages which were approved in [3] when the scheme (15) was applied on the SM method. As a result, the hybrid SM model (called HSM) was defined and tested in [3]. Herein, we apply the same hybridization strategy to the accelerated double direction method (ADD method, shortly), introduced in [4]. Derived scheme will be based on the hybrid scheme (15) and with that it keeps the accelerated features of the ADD iterations.
In order to complete the presentation, we start from the ADD iteration:
Algorithm 1: Procedure Second direction
(calculation of the second direction vector
Require: Objective function
1: Compute
where
2: Return
Remark 1.
For further investigation within this topic, the second direction
Applying the hybrid scheme (15) on the iterative rule (17), we get the hybrid iterative scheme
To simplify further calculation, we will use a constant value for the parameter
Yet, we need to determine the iterative value of the accelerated parameter
With the aim of preserving the Second-Order Necessary Condition and Second-Order Sufficient Condition, we assume positivity of the acceleration parameter:
In order to present the main HADD algorithm, we need two additional auxiliary procedures. The first one is previously displayed Algorithm 1, by which we calculate the second vector direction,
Algorithm 3 describes the main algorithm, termed the HADD algorithm.
Algorithm 2: The Backtracking line search procedure.
Require: Objective function
1:
2: While
3: Return
Algorithm 3: HADD method.
Require:
1: Set
2: If
3: Compute the iterative step length,
4: Compute second vector direction,
5: Compute
6: Calculate the approximation parameter
7: If
8: Set
9: Return
4. Convergence of the HADD Method
The convergence properties of the established HADD iterative method are considered on the set of uniformly convex and strictly convex quadratic functions. In the case of uniformly convex functions the statements are the same as exposed in [1, 4]. For that reason, we just restate the following lemma, in which decreasing of the objective function in two successive points is estimated with respect to the HADD scheme. Thereupon, the upcoming theorem confirms linear convergence of our hybrid accelerated model.
Lemma 2.
Suppose the function
Theorem 3.
For the twice continuously differentiable and uniformly convex function
We show now that the iteration (20) is convergent regarding the set of strictly convex quadratic functions
Lemma 4.
Let
Proof.
Let us calculate the difference in two successive iterative points of the goal function (29):
Including the iteration (20) we continue computations:
The right hand side of the previous expression can be further transformed as follows:
After some calculations, we obtain
Theorem 5.
Let the iterations (19) be applied on strictly convex quadratic function
Proof.
Let us consider the orthonormal system of eigenvectors
Applying (20), further we conclude that
(1)
This case implies the next set of inequalities:
As a consequence, we can conclude that
(2)
In this case, one can verify the following estimations:
Remark 6.
The assumption (42) used in the previous theorem is required in order to prove that the HADD process is convergent for the strictly convex quadratics. Therewith, knowing that the hybrid parameter
5. Computational Tests and Comparisons
The performance of the C
We compare the hybrid accelerated HADD method with its forerunner ADD scheme, as well as with the hybrid accelerated HSM method. The number of function evaluations is the performance profile measured in all tests. The dominance of the ADD method regarding the number of iterations among the other comparative models was confirmed in [4]. However, from that research we do not have any information about the behavior of the ADD method when the number of function evaluations is involved. With respect to this parameter, the HSM scheme upgrades the accelerated SM method as well as Nesterov’s line search algorithm; see [3]. For these reasons, our experimental goal is to numerically prove better performance feature of the HADD method, considering the number of function evaluations, when compared with the ADD and the HSM method.
In Table 1, we display the number of problems, out of 630, for which an algorithm achieved the minimum number of function evaluations. In the same table, we also display the total number of problems for which all three algorithms achieved an equal number of function evaluations. Based on the results displayed in this table, it is obvious that the HADD scheme convincingly outperforms the other two comparative models.
Table 1
Comparison between the HADD, ADD, and HSM methods regarding the minimal number of function evaluations.
Comparative methods | HADD | ADD | HSM | = |
---|---|---|---|---|
Number of function evaluations | 142 | 0 | 18 | 50 |
For more clear visualization of the performance of the HADD algorithm versus the ADD and the HSM algorithms, we display in Figure 1 the Dolan-Moré’s performance profile subject to the number of function evaluations metric. As we can see, the HADD scheme is more robust and therewith more efficient than the other two methods.
[figure omitted; refer to PDF]Obtained numerical results confirm that applied hybridization process is a good way to improve some important characteristics of chosen accelerated methods. Preferable outcomes of the HADD scheme, regarding analyzed characteristic, come from the properly chosen hybrid value
6. Conclusion
We present a hybrid accelerated double direction gradient method for solving unconstrained optimization problems. The HADD method is derived by applying good properties of the hybrid representation introduced in [3] in conjunction with the form of double direction optimization model with accelerated parameter presented in [4]. The convergence of defined optimization model is provided on the set of uniformly convex and strictly convex quadratic functions.
The HADD scheme reserves preferable features of both forerunner methods. Therewith, according to conducted numerical experiments, it outperforms the ADD and HSM methods regarding the requested number of function evaluations. We evaluated the Dolan-Moré performance profiles of comparative methods and showed that the HADD iteration is the most efficient compared to the other two algorithms.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Acknowledgments
The first and the third authors acknowledge support from the internal research project IS01-17 supported by the Faculty of Sciences and Mathematics, University of Priština, Serbia. The second author gratefully acknowledges support from the Project supported by Ministry of Education and Science of Republic of Serbia, Grant No. 174013.
[1] P. S. Stanimirović, M. B. Miladinović, "Accelerated gradient descent methods with line search," Numerical Algorithms, vol. 54 no. 4, pp. 503-520, DOI: 10.1007/s11075-009-9350-8, 2010.
[2] M. J. Petrović, "An accelerated double step size model in unconstrained optimization," Applied Mathematics and Computation, vol. 250, pp. 309-319, DOI: 10.1016/j.amc.2014.10.104, 2015.
[3] M. J. Petrovic', V. Rakocevic', N. Kontrec, S. Panic', D. Ilic', "Hybridization of accelerated gradient descent method," Numer. Algor, 2018.
[4] M. J. Petrović, P. S. Stanimirović, "Accelerated Double Direction Method for Solving Unconstrained Optimization Problems," Mathematical Problems in Engineering, vol. 2014,DOI: 10.1155/2014/965104, 2014.
[5] P. S. Stanimirović, G. V. Milovanović, M. J. Petrović, "A Transformation of Accelerated Double Step Size Method for Unconstrained Optimization," Mathematical Problems in Engineering, vol. 2015,DOI: 10.1155/2015/283679, 2015.
[6] S. Ishikawa, "Fixed points by a new iteration method," Proceedings of the American Mathematical Society, vol. 44, pp. 147-150, DOI: 10.1090/S0002-9939-1974-0336469-5, 1974.
[7] W. R. Mann, "Mean value methods in iteration," Proceedings of the American Mathematical Society, vol. 4, pp. 506-510, DOI: 10.1090/S0002-9939-1953-0054846-3, 1953.
[8] E. Picard, "Memoire sur la theorie des equations aux derivees partielles et la methode des approximations successives," Journal de Mathématiques Pures et Appliquées, vol. 6, pp. 145-210, 1890.
[9] S. H. Khan, "A Picard-Mann hybrid iterative process," Fixed Point Theory and Applications, vol. 2013, article 69,DOI: 10.1186/1687-1812-2013-69, 2013.
[10] N. I. Djuranović-Miličić, M. Gardaševic-Filipović, "A multi-step curve search algorithm in nonlinear optimization: nondifferentiable convex case," Facta Universitatis, Series: Mathematics and Informatics, vol. 25, pp. 11-24, 2010.
[11] A. Kumar, D. K. Gupta, E. Martinez, S. Singh, "Directional k -step Newton methods in n variables and its semilocal convergence analysis," Mediterranean Journal of Mathematics, vol. 15 no. 2,DOI: 10.1007/s00009-018-1077-0, 2018.
[12] B. Molina, M. Raydan, "Preconditioned Barzilai-BORwein method for the numerical solution of partial differential equations," Numerical Algorithms, vol. 13 no. 1-2, pp. 45-60, DOI: 10.1007/bf02143126, 1996.
[13] N. Andrei, "An unconstrained optimization test functions collection," Advanced Modeling and Optimization, vol. 10 no. 1, pp. 147-161, 2008.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2018 Milena J. Petrović et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
We present a hybridization of the accelerated gradient method with two vector directions. This hybridization is based on the usage of a chosen three-term hybrid model. Derived hybrid accelerated double direction model keeps preferable properties of both included methods. Convergence analysis demonstrates at least linear convergence of the proposed iterative scheme on the set of uniformly convex and strictly convex quadratic functions. The results of numerical experiments confirm better performance profile in favor of derived hybrid accelerated double direction model when compared to its forerunners.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 University of Priština, Faculty of Science, Lole Ribara 29, 28000 Kosovska Mitrovica, Serbia
2 University of Niš, Faculty of Science and Mathematics, Višegradska 33, 18000 Niš, Serbia
3 Politehnika, School for New Technologies, Autoput 18, 11000 Belgrade, Serbia