1. Background
Our main focus concerns a special form of accelerated gradient methods that contain two vector directions. Aiming to develop as efficient of an optimization method as possible, we explored an approach that combines several different search vectors. In a calculative sense, iterations with two directions have become the preferred choice above others. In [1], the authors suggested the double-direction iterative method for solving non-differentiable problems
(1)
In (1), is the iterative step length, while and are two differently defined vector directions. These three parameters are described by the following algorithms listed below. These procedures are denoted as the curve search algorithm, the algorithm for deriving vector direction , and the algorithm for deriving vector direction , respectively.Curve search algorithm
(2)
where is the smallest integer from such that(3)
In (3), is estimated as , stands for the second-order Dini upper-directional derivative at in direction d, and the function F presents the Moreau–Yosida regularization of the objective function f associated to the metric M, defined as follows:Algorithm for deriving vector direction
, , is an index set at k-th iteration. and , is a vector that satisfies .Algorithm for deriving vector direction
is the solution to the problem(4)
The results presented in [1] motivated the authors in [2] to define an accelerated gradient version of the iterative rule (1). In ref. [2], the accelerated double-direction method, denoted as the ADD, is presented as follows:
(5)
In (5), is the current iterative point, is the iterative step size, is the gradient of the objective function, is the accelerated parameter of the ADD method, and is the second vector direction.The step size parameter of the iteration (5) is calculated in the following way:
is the smallest integer from such that where is a real number such that .An alternative method for deriving the iterative step size is using the backtracking line search procedure, originally presented in [3]. In this procedure, we find the iterative step length , generally by starting with initial value . Applying the Backtracking parameters and , this algorithm checks the following condition:
(6)
while updating the reduced value of the iterative step size as next . The optimal initial step length, , is obtained after fulfilling the exit condition of the backtracking algorithm.There are two main approaches for calculating the step length parameter in an optimization method:
1. exact line search;
2. inexact line search.
Using procedure 1, in each iteration, the step size value is derived as the solution to the following problem:
(7)
Clearly, the exact line search requests additional CPU time. Accordingly, the required number of iterations and the number of the function evaluations are certainly increasing. In relation to that, in many contemporary optimization schemes, the step length parameter is derived using the second approach, i.e., through inexact line search procedures. We list some of the commonly applied inexact line search techniques as follows:-
Weak Wolfe’s line search algorithm is given by the following relations [4]:
-
Strong Wolfe’s line search rule is expressed as follows:
-
Armijo’s Backtracking procedure is introduced in [3] and defined by Relation (6).
-
Goldstein’s rule [5] can be stated as a generalization of the Armijo’s with more strict conditions:
where ρ is the parameter, bounded as in Armijo’s procedure, .
The authors in [2] modified the algorithms for deriving vector directions of the iteration (5) in gradient terms. These modifications are illustrated in Algorithms 1 and 2.
Algorithm 1 Calculation of the vector :
Algorithm 2 Calculation of the vector :
where is the solution of the transformed minimization problem (4)
The vector of the search direction is one of the important elements of each gradient minimization scheme for solving unconstrained optimization problems. In solving minimization tasks, it is assumed that the iterative search direction satisfies the following inequality:
(8)
where is the gradient of the objective function at point . Relation 8 is known as the descending condition. We list only some of the approaches for generating search directions that fulfill Condition (8).One of the common methods for deriving a vector direction that fulfills the descending condition is by expressing it as a multiplication of a negative gradient and a positive accelerated parameter. The acceleration factor can be determined using the features of the first or second-order Taylor expansion of the relevant iteration.
An unpublished idea for accelerated parameter construction relates to the properties of the logarithmic and local double-logarithmic reconstructions described in [6,7,8].
A similar approach concerns third-order accurate non-polynomial reconstructions and hyperbolic reconstructions of the objective function as a basis for developing an efficient search direction [9,10].
The vector direction can be generated as a linear combination of the negative gradient vector and the vector defined in Algorithm 2, as proposed in [11].
In [12], the direction vector is presented as multiplication of the gradient and the accelerated parameter , where ,
The search direction of the conjugate gradient algorithm can be formulated as a linear combination of the gradient and the difference between two successive iterative points, as presented in [13].
The last suggestion from the previous Remark 3 induces an idea that can be used as a basis for further studies. This research can relate to comparisons of a chosen conjugate gradient method and the hybrid accelerated method proposed in this paper.
The general form of the conjugate gradient method is given as follows:
(9)
where the iterative step length variable is calculated via the exact line search or via one of the listed inexact line searches given in Remark 1. The main specific of a conjugate gradient scheme comes from the method of generating the vector direction , which is defined as follows: i.e.,(10)
In (10), symbolize the scalar product of the gradient vectors.For suggested comparative studies, in relation to the research presented in this paper, it would be valuable to pay special attention on the set of quadratic functions. A quadratic function is defined by the following expression:
(11)
where A is a symmetric, positive definite matrix, , and . Starting with the initial condition , after some calculations, an update for the vector is obtained as follows:(12)
where(13)
The conjugate gradient method (9) with a vector direction defined by (12), where is calculated using Relation (13), is known as the Fletcher–Reeves formulation of the conjugate gradient method [14].We list several significant variants of the conjugate gradient method that differ with respect to the expressions that define quotients [15,16,17]:
For example, as a comparative minimization model, the conjugate method proposed in [13] can be taken. Providing the suggested comparative analysis would certainly contribute to the optimization community in general.One of the crucial variables of the ADD scheme (5) is the acceleration factor, calculated using the second-order Taylor expansion of the objective function:
(14)
The most significant contribution achieved in ref. [2] likely concerns the importance of the accelerated parameter. To substantiate this fact, the authors constructed and tested the non-accelerated version of the ADD model, called NADD method. In these studies, the incomparable effectiveness of the ADD method was confirmed.Recently, in [18], the authors introduced a new hybrid approach for generating accelerated gradient optimization methods. This calculative technique is denoted as s-hybridization. In developing this new approach of constructing an efficient minimization scheme, the authors were guided by research regarding the nearly contraction mappings and nearly asymptotically nonexpansive mappings and the existence of fixed points of these classes of mappings [19,20]. The main idea regarding the s- schemes arrives from the study presented in [20], where the following three-termed s-iterative rule is presented:
(15)
In (15), and are sequences of real numbers satisfying the following conditions:(16)
The authors in [18] simplified S-Iteration (15) by applying Condition (17)(17)
which transforms limits (16)–(18)(18)
Therewith, the s-iteration with one corrective parameter is expressed as follows:(19)
Guided by Iteration (19), in connection with the SM method from [21], the authors in [18] proposed the SHSM optimization method (20):(20)
The authors proved that this model is well defined and established comprehensive convergence analysis.This paper is organized as follows: in Section 2, we develop the s-hybrid double-direction minimization method based on the obtained results from the relevant studies described in the first section. The convergence analysis is presented in Section 3. Numerical investigations are illustrated in Section 4.
2. S-Hybridization of the Accelerated Double-Direction Method
In this section, we generate the s-hybrid model using the ADD iterative rule,
as a guiding operator in the three-term process (19). Applying the previously stated facts regarding the s-hybridization technique and the ADD method, we develop the shADD process trough the following three-term relations:(21)
Before we state and prove that the three-termed process (21), rewritten in a merged form, presents an accelerated gradient descent method, we induce the following two important terms.
(Second-order necessary conditions—unconstrained case [22]). Let be an interior point of the set Ω, and suppose that is a relative minimum point over Ω of the function . Then,
,
for all .
(Second-ordered sufficient conditions—unconstrained case [22]). Let be a function defined in a region in which the point is an interior point. Suppose in addition that
,
Hessian of , is positive definite.
The accelerated gradient iterative form of the shADD process (21) is given by the following relation:
(22)
The merged iterative rule of the process (21) can be derived by substituting the expression of from (21) into the previous relation of the same three-term method, i.e., the one that defines :
which proves (22).Now, we show that Method (22) fulfills the gradient descent property. For this purpose, let us rewrite relation (22) as follows:
where(23)
Knowing that implies . Further, can be considered as a linear combination of the gradient vector, since the vector direction is derived by Algorithm 2. With that, the parameter , as an acceleration parameter, is a positive constant. Therefore, direction is the gradient descent vector.Now, we derive the iterative value of the acceleration parameter for Method (22). To achieve this goal, we use the second-order Taylor series of the objective function f:
where satisfies the following: Instead of the function’s Hessian , we use the diagonal scalar matrix approximation in the previous Taylor expression, i.e., acceleration matrix : This gives us the expression of the acceleration parameter of the process:(24)
We assume the positiveness of the derived acceleration parameter . This fact confirms that the second-order necessary and sufficient conditions have been fulfilled. In the case of , we assign and derive the next iterative point as Knowing that induces , together with the fact that confirms that the previous scheme is the gradient descent method. □We end this section by exposing the algorithm of the shADD method, derived on the basis of the previously provided analysis.
Taking the initial values , , , , the algorithm of the shADD method is given by the following steps:
Set , compute , , and take ;
If , then go to Step 9; else, continue to Step 3;
Apply the backtracking algorithm to calculate the iterative step length ;
Compute the first vector direction using Algorithm 1;
Compute the second vector direction using Algorithm 2;
Compute using the iterative rule (22);
Determine the acceleration parameter using (24);
If , then take ;
Set , go to Step 2;
Return and .
3. Convergence Features of the shADD Method
We start this section with some relevant known statements that can be found in [23,24].
If the function is twice continuously differentiable and uniformly convex on , then:
-
the function f has a lower bound on , where is available;
-
the gradient g is Lipschitz continuous in an open convex set B that contains , i.e., there exists such that
(25)
Under the assumptions of Lemma 3, there exist real numbers m, M satisfying the following:
(26)
such that has an unique minimizer and(27)
(28)
(29)
Depending on the degree of complexity that a particular non-linear problem may have, the examination of its convergence resorts to establishing convergence on specific sets. Therefore, we expose in this section the convergence analysis of the derived process on the set of strictly convex quadratic functions. The general expression of the strictly convex quadratics is given by (30).
(30)
In (30), A is the real positive definite symmetric matrix, and . Further on, we use the following notations regarding the relevant eigenvalues of matrix A:Previous research showed that for the strictly convex quadratic an adequate relation, usually, a connection between the smallest and the largest eigenvalues must be fulfilled in order to establish the convergence of the objective optimization method [2,21,25,26]. In the next lemma, we define that connection by applying the method.
The relation between the smallest and largest eigenvalues of symmetric positive definite matrix that defines the strictly convex quadratic function (30) to which the method (22) is applied is given as follows:
(31)
where β is the parameter defined in the backtracking procedure.To prove (31), we start with the estimation of the difference of the function (30) values in two successive points:
The expression above, which describes the difference between function’s values for two successive points, is determined based on the following facts:Matrix A is symmetric, so
The gradient of Function (30) is
(32)
(33)
According to findings revealed in [21], the value of the iterative step length of the accelerated gradient method derived via the backtracking inexact algorithm satisfies the following:(34)
where L is the Lipschitz constant that figures in Proposition (3), so the following is valid:(35)
Considering relation (32), which defines the gradient of the strictly convex function, we have the following: The previous relation confirms that the largest eigenvalue of the symmetric matrix A fulfills the property of the Lipschitz constant L in (34). Additionally, according to the limitations of the backtracking parameters and , we derive the following estimations: which confirms the right side of Estimation (31). Based on (33) and the fact that the iterative step size is less than 1, the first inequality of (31) arises. □On the basis of proven estimations (31) relating to the acceleration parameter, backtracking parameter, and the lowest and the largest eigenvalues of the symmetric, positive definite matrix A that figures in the expression (30), using the following theorem, we establish the convergence of the method on the set of strictly convex quadratics.
For the strictly convex quadratic function (30), the process (22) is linearly convergent when . More precisely, the following relations are valid:
-
for some real constants and , such that
(36)
When the vectors represent the orthonormal set of eigenvectors of matrix A, the following inequations are fulfilled:
(37)
including
(38)
-
For the gradient (32) of the function (30), the following is valid:
(39)
Taking the expression of Gradient (32) of Function (30) at the ()-th iteration, we obtain the following:
Applying the orthonormal representations (36) in the previous equation leads us to the following:(40)
Knowing that we will prove (38) by showing that For this purpose, let us first assume that , i.e., Applying (31), we obtain the following: We rewrite the previous inequalities as follows:(41)
We assume now the opposite case: The last inequality gives which directly implies(42)
Estimations (41) and (42) prove (38).Finally, in order to prove (39), we use the gradient representation from (36),
which results in the following conclusion:(43)
Applying the fact that on inequalities (37) directly proves (39). □Non-Convex Case Overview
In the previous section, we proved that the shADD method linearly converges to a set of strictly convex quadratics. Although it is not the main subject of this research, in this subsection, we analyze a possible application of the presented scheme when the objective function is non-convex. The importance of providing this introductory discussion on this topic simply arises from the endless array of contemporary non-convex problems such as matrix completion, low-rank models, tensor decomposition, and deep neural networks.
Neural networks, considered as universal function approximators, contain significant symmetric properties. These features define them as non-convex structures. Some of the known techniques for solving machine learning problems and other non-convex problems are as follows:
Stochastic gradient descent methods,
Mini batch approach,
Stochastic variance reduced gradient (SVRG) method,
Alternating minimization methods,
Branch and bound methods.
Confirming convergence properties in non-convex optimization is quite difficult. In a theoretical sense, there exist no regular approaches to achieving this goal, as is the case of convex problems. Additionally, there are potentially many local minimums and the existence of saddle points and flat regions when the objective function is non-convex.
Generally, when solving non-convex optimization tasks, theoretical guarantees are very weak, and there is no tried and tested way for ending this process successfully.
Principal component analysis (PCA) is a technique for linear dimensionality reduction, which is useful in proving global convergence of minimization methods when applied on non-convex functions. We propose connecting this approach with the shADD method in further studies. The PCA process can be characterized through the following steps:
Standardizing the range of continuous initial variables;
Computing the covariance matrix to identify correlations;
Computing the eigenvectors and eigenvalues of the covariance matrix to identify the principal components;
Creating a feature vector to decide which principal components to keep;
Recasting the data along the principal components axes.
We set the goal problem as follows: determine the dominant eigenvector and eigenvalue of a positive symmetric semidefinite matrix A. We can write this problem as follows:
(44)
The equivalent of problem (44) is (45):(45)
where is the Frobenius norm, i.e., . Taking the objective function, defined in terms of the Frobenius norm,(46)
we see that the gradient of this function is given by the following expression:(47)
The classical gradient descent update step for Function (46) can be written as follows:(48)
where the adaptive step size parameter fulfills the following relation:(49)
Applying (49) in (48) leads to the following:(50)
After inductively taking the previous relation, we conclude that(51)
The previous relation (51) confirms that the gradient descent iteration (48) converges linearly, since
(52)
A similar analysis can be applied to the shADD iteration (22) for Function (46). According to the construction of the vector (Algorithm 2) in (22), we can modify this vector direction as a linear combination of the gradient vector, as explained in the proof of Lemma 1. Considering this fact allows us to rewrite iteration (22) in a simpler form for which Property (52) can be easily proved.
4. Numerical Test Results
In this section, we analyze the numerical performance of the shADD method depending on the choice of parameter (18), which is aptly named the corrective parameter. For the selected values of this parameter, we track standard numerical metrics, including the number of iterations performed, CPU time, and the number of function evaluations.
As proposed in ref. [27], which presents an extensive comparative analysis of several Khan-hybrid models, for a range of values (18), we take a specific numerical value for all . This further reinforces the corrective expression of the shADD iteration (22):
We have observed that for specific values of parameter , we have the following values of the expression
(53)
All of this motivated us to test the shADD method for these five specified values of . For this purpose, we selected five test functions from [28]. We tested these functions for the five given values of the corrective parameter and for ten different values of the number of variables . For each test function, we summarized the obtained outcomes for all selected numbers of variables. The results obtained from the measured performance metrics (number of iterations, CPU time, and number of evaluations) are presented in Table 1, Table 2, and Table 3, respectively.
We can observe that for the first test function, the Extended Penalty function, the values of the output results regarding the number of iterations and the number of evaluations do not depend on changes in the corrective parameter . For the same function, the changes in CPU time for different values of the corrective parameter are also minimal. However, for the other test functions, there are differences in the final values of all measured metrics depending on the choice of the corrective parameter value. In terms of metrics, regarding the number of iterations, the best results are achieved for and . In the case of measuring CPU time, tests for and take the least time. In terms of the number of evaluations, the smallest values are achieved for and . A general conclusion we can draw, based on a total of 250 tests, is that it is advisable to use a corrective parameter value of or .
The tests were conducted using a standard termination criterion:
The code was written in the C++ programming language.
5. Conclusions
In this paper, we propose a new s-hybrid variant of the accelerated double-direction minimization model. The explored convergence properties of the developed iterative rule on the set of strictly convex quadratic functions confirm that the method has linear consistency.
Numerical tests were performed for different values of the corrective parameter. This study leaves an open space for further investigations of the numerical performance of the defined process and comparative analysis regarding similar models.
Conceptualization, V.R. and M.J.P.; methodology, V.R.; software, M.J.P.; validation, V.R. and M.J.P.; formal analysis, V.R. and M.J.P.; investigation, M.J.P.; resources, V.R.; data curation, V.R. and M.J.P.; writing—original draft preparation, V.R. and M.J.P.; writing—review and editing, V.R.; visualization, M.J.P.; supervision, V.R and M.J.P.; project administration, M.J.P.; funding acquisition, V.R. All authors have read and agreed to the published version of the manuscript.
The data supporting this study are available from the authors upon request.
The author Milena J. Petrović gratefully acknowledges support from the Project supported by Ministry of Science, Technological Development and Innovation of the Republic of Serbia project no. 451-03-65/2024-03/200123.
The authors declare no conflicts of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Number of iterations with respect to the corrective parameter value.
Test Functions | | | | | |
---|---|---|---|---|---|
Extended Penalty function | 60 | 60 | 60 | 60 | 60 |
Diagonal 1 function | 54 | 70 | 93 | 81 | 87 |
Diagonal 3 function | 190 | 155 | 166 | 177 | 223 |
Generalized Tridiagonal 1 function | 128 | 120 | 120 | 120 | 120 |
Quadratic Diagonal Perturbed function | 2588 | 2407 | 1455 | 1618 | 3779 |
CPU with respect to the corrective parameter value.
Test Functions | | | | | |
---|---|---|---|---|---|
Extended Penalty function | 5 | 3 | 5 | 5 | 4 |
Diagonal 1 function | 3 | 5 | 6 | 6 | 6 |
Diagonal 3 function | 10 | 12 | 14 | 19 | 17 |
Generalized Tridiagonal 1 function | 10 | 9 | 10 | 10 | 10 |
Quadratic Diagonal Perturbed function | 321 | 200 | 249 | 175 | 294 |
Number of evaluations with respect to the corrective parameter value.
Test Functions | | | | | |
---|---|---|---|---|---|
Extended Penalty function | 3729 | 3729 | 3729 | 3729 | 3729 |
Diagonal 1 function | 13,979 | 13,785 | 19,634 | 14,440 | 14,870 |
Diagonal 3 function | 7140 | 4289 | 3741 | 9486 | 7767 |
Generalized Tridiagonal 1 function | 1053 | 1006 | 1013 | 1019 | 1027 |
Quadratic Diagonal Perturbed function | 8962 | 8419 | 5563 | 6052 | 12,535 |
References
1. Djuranović-Miličić, N.I.; Gardašević-Filipović, M. A multi-step curve search algorithm in nonlinear optimization: Nondifferentiable case. Facta Univ. Ser. Inform.; 2010; 25, pp. 11-24.
2. Petrović, M.J.; Stanimirovic, P.S. Accelerated Double Direction Method For Solving Unconstrained Optimization Problems. Math. Probl. Eng.; 2014; 2014, 965104. [DOI: https://dx.doi.org/10.1155/2014/965104]
3. Armijo, L. Minimization of functions having Lipschitz first partial derivatives. Pac. J. Math.; 1966; 6, pp. 1-3. [DOI: https://dx.doi.org/10.2140/pjm.1966.16.1]
4. Wolfe, P. Convergence conditions for ascent methods. SIAM Rev.; 1968; 11, pp. 226-235. [DOI: https://dx.doi.org/10.1137/1011036]
5. Goldstein, A.A. On steepest descent. SIAM J. Control; 1965; 3, pp. 147-151.
6. Petrović, M. A Truly Third Order Finite Volume Scheme on Quadrilateral Mesh. Master’s Thesis; Lund University: Lund, Sweden, 2006.
7. Artebrant, R.; Schroll, H.J. Conservative Logarithmic Reconstruction and Finite Volume Methods. SIAM J. Sci. Comput.; 2005; 27, pp. 294-314. [DOI: https://dx.doi.org/10.1137/03060240X]
8. Artebrant, R.; Schroll, H.J. Limiter-free Third Order Logarithmic Reconstruction. SIAM J. Sci. Comput.; 2006; 28, pp. 359-381. [DOI: https://dx.doi.org/10.1137/040620187]
9. Artebrant, R. Third order acurate non-polynomial reconstruction on ractangular and triangular meshes. J. Sci. Comput.; 2007; 30, pp. 193-221. [DOI: https://dx.doi.org/10.1007/s10915-005-9063-7]
10. Schroll, H.J.; Svensson, F. A Bi-Hyperbolic Finite Volume Method on Quadrilateral Meshes. J. Sci. Comput.; 2006; 26, pp. 237-260. [DOI: https://dx.doi.org/10.1007/s10915-004-4927-9]
11. Potra, F.A.; Shi, Y. Efficient line search algorithm for unconstrained optimization. J. Optim. Theory Appl.; 1995; 85, pp. 677-704. [DOI: https://dx.doi.org/10.1007/BF02193062]
12. Andrei, N. An acceleration of gradient descent algoritham with backtracing for unconstrained optimization. Numer. Algorithms; 2006; 42, pp. 63-173. [DOI: https://dx.doi.org/10.1007/s11075-006-9023-9]
13. Andrei, N. An accelerated conjugate gradient algorithm with guaranteed descent and conjugacy conditions for unconstrained optimization. Optim. Methods Softw.; 2012; 27, pp. 583-604. [DOI: https://dx.doi.org/10.1080/10556788.2010.501379]
14. Fletcher, R.; Reeves, C.M. Function minimization by conjugate gradients. Comput. J.; 1964; 7, pp. 149-154. [DOI: https://dx.doi.org/10.1093/comjnl/7.2.149]
15. Dai, Y.H.; Yuan, J.Y.; Yuan, Y. Modified two-point step-size gradient methods for unconstrained optimization. Comput. Optim. Appl.; 2002; 22, pp. 103-109. [DOI: https://dx.doi.org/10.1023/A:1014838419611]
16. Dai, Y.H.; Yuan, Y. Alternate minimization gradient method. IMA J. Numer. Anal.; 2003; 23, pp. 377-393. [DOI: https://dx.doi.org/10.1093/imanum/23.3.377]
17. Polak, E.; Ribiére, G. Note sur la convergence de méthodes de directions conjuguées. Rev. Fr. d’Inform. Rech. Opér. Sér. Rouge; 1969; 16, pp. 35-43. [DOI: https://dx.doi.org/10.1051/m2an/196903R100351]
18. Petrović, M.J.; Rakočević, V.; Ilić, D. Hybrid optimization models based on S-iteration process. Filomat; 2024; 38.
19. Sahu, D.R. Fixed points of demicontinuous nearly Lipschitzian mappings in Banach spaces. Comment. Math. Univ. Carol.; 2005; 46, pp. 653-666.
20. Agarwal, R.P.; O’Regan, D.; Sahu, D.R. Iterative construction of fixed points of nearly asymptotically nonexpansive mappings. J. Nonlinear Convex Anal.; 2007; 8, 61.
21. Stanimirović, P.S.; Miladinović, M.B. Accelerated gradient descent methods with line search. Numer. Algorithms; 2010; 54, pp. 503-520. [DOI: https://dx.doi.org/10.1007/s11075-009-9350-8]
22. Luenberg, D.G.; Ye, Y. Linear and Nonlinear Programming; Springer Science+Business Media, LLC: New York, NY, USA, 2008.
23. Ortega, J.M.; Reinboldt, W.C. Iterative Solution of Nonlinear Equation in Several Variables; Academic: London, UK, 1970.
24. Rockafellar, R.T. Convex Analysis; Princeton University Press: Princeton, NJ, USA, 1970.
25. Long, J.; Hu, X.; Zhang, L. Improved Newton’s method with exact line searches to solve quadratic matrix equation. J. Comput. Appl. Math.; 2008; 222, pp. 645-654. [DOI: https://dx.doi.org/10.1016/j.cam.2007.12.018]
26. Zhou, Q. An adaptive nonmomotonic trust region method with curlinear searches. J. Comput. Math.; 2006; 24, pp. 761-770.
27. Rakočević, V.; Petrović, M.J. Comparative analysis over accelerated models for solving unconstrained optimization problems with application of Khan’s hybrid rule. Mathemathics; 2022; 10, 4411. [DOI: https://dx.doi.org/10.3390/math10234411]
28. Andrei, N. An Unconstrained Optimization Test Functions Collection. Adv. Model. Optim.; 2008; 10, pp. 1-15. Available online: https://camo.ici.ro/journal/vol10/v10a10.pdf (accessed on 7 February 2025).
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In this paper, we study a recently established s-hybrid approach for generating gradient descent methods for solving optimization tasks. We present an s-hybrid variant of the accelerated double-direction method. The results obtained based on convergence analysis confirm the first-order consistency of the newly defined method on a set of strictly convex quadratic functions.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Faculty of Sciences and Mathematics, University of Niš, Višegradska 33, 18108 Niš, Serbia;
2 Faculty of Sciences and Mathematics, University of Priština in Kosovska Mitrovica, Lole Ribara 29, 38220 Kosovska Mitrovica, Serbia