ABSTRACT
Fractional calculus is the branch of mathematics that studies the several possibilities of generalizing the derivative and integral of a function to noninteger order. Recent studies found in literature have confirmed the importance of fractional calculus for minimization problems. However, the study of fractional calculus in interior point methods for solving optimization problems is still new. In this study, inspired in applications of fractional calculus in many fields, was developed the so-called fractional order log barrier interior point algorithm by replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions of Karush-Kuhn-Tucker to solve polynomial regression models in the ℓ p −norm for 1 < p < 2. Finally, numerical experiments are performed to illustrate the proposed algorithm.
Keywords: nonlinear programming; polynomial regression; ℓp−norm; interior point method; fractional derivative
1 INTRODUCTION
It is fundamentally important to make predictions based upon scientific data. The problem of fitting curves to data points has many practical applications Bard (1974); Sevaux & Mineur (2007); Chatterjee & Hadi (2012).
Given a set of m data points in , where a i is an argument value and b i a corresponding dependent value, with for all . The curve fitting procedure tries to build a linear or nonlinear function y = f (x), defined for all possible choices of x, that approximately fits the data set. The fitted curves to the data by f are most often chosen to be polynomials Süli & Mayers (2003).
Let y = f (x) be a polynomial function of degree n− 1 of the form , the candidate function to fit data. The procedure to fit the polynomial function to data, in the ℓ p −norm, determines the vector , where the superscript T represents transpose, that minimizes the ℓ p −norm of the residual error as follows
where denotes ℓ p −norm, and is a Vandermonde matrix, with rank (A) = m, which can be written as
This is a nonlinear (or linear) regression model if n > 2 (or n = 2). If m ≤ n, there is an (n − 1)th degree polynomial satisfying . If m > n, the problem cannot be exactly solved and one needs to find the vector x in (1).
In many industrial applications, missing data or anomalous values can arise as errors of measurement instruments during the data generation. In this cases, polynomial regression models in the ℓ p −norm, with , are more robust Forsythe (1972). It is important to choose an appropriate value for p and several criteria for the choice of p have been studied Rice & White (1964).
The calculus of integrals and derivatives of arbitrary order, named as fractional calculus, was conceptualized in connection with the infinitesimal calculus since 1695 Oldham & Spanier (1974). Some of the application areas include: viscoelasticity, signal processing, probability, statistics, electrochemistry, diffusion in porous media, fluid flow, backpropagation training of neural networks, fuzzy control method, and so on Kilbas et al. (2006); Dalir & Bashour (2010); Mohammadzadeh & Kayacan (2020); Wang et al. (2017a); Chen et al. (2017a); Grigoletto & Oliveira (2020). Also, its importance in minimization problems has been confirmed by recent works found in literature. For example, Chen et al. Chen et al. (2017b) presented the fractional order gradient methods (FOGMs) by writing the Riemann-Liouville and Caputo fractional derivatives as Taylor series, and Wang et al. Wang et al. (2017b) proposed the fractional gradient descent method employing the Caputo derivative for the backpropagation training of neural networks. However, the study of fractional calculus in the field of interior point methods for solving optimization problems is still new.
The goal of this study is to investigate the fractional order log barrier interior point algorithm, involving the Caputo fractional derivative. It is based on replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions for solving polynomial regression models in the ℓ p −norm for 1 < p < 2. Functions of the form , with κ > 0 and 1 < p < 2, arise when solving problem 1), according to the approach which will be discussed throughout this study. The Caputo fractional derivative, in addition to generalize the integer order derivative, is a useful tool to obtain the derivative for functions as g(x), as p is on the interval 1 < p < 2, then it will be a non integer number.
This paper is organized as follows. Preliminary concepts of fractional calculus are presented in Section 2. In Section 3, the fractional order log barrier interior point algorithm for solving polynomial regression models in the ℓ p −norm is discussed. Numerical experiments are performed to illustrate the proposed algorithm in Section 4, and the Section 5 contains the conclusions.
2 PRELIMINARY CONCEPTS OF FRACTIONAL CALCULUS
Some basic concepts and definitions involving special functions and the Caputo fractional derivatives will be presented in this section.
Definition 1. The gamma function Γ(z) is originally defined by Erdélyi et al. (1981); Andrews et al. (1999):
For gamma function, the reduction formula
holds. In particular, when , then .
Definition 2. The beta function B(z, w) is defined by Erdélyi et al. (1981); Andrews et al. (1999):
It is connected with the gamma function by the following relation
The fractional calculus basically defines several integral and derivative operators of arbitrary order. The Caputo fractional derivative is one of the fractional derivative operators Kilbas et al. (2006).
Definition 3. The Caputo left-sided fractional derivative with respect to x, of order 0 < α < 1, of functions f on a subset Ω = [a, b] of real axis , where , denoted by , is given by
Note that the Caputo fractional derivative is a nonlocal fractional derivative, because it depends on the choice of the order α, function f (x), x, and also relies on the total effects on the interval [a, x]. Usually, this is called memory effect.
In particular, .
Property 1. Let be defined for , with κ ≥ 0, and 1 < p < 2, and let 0 < α < 1. Then
Proof. Applying the Caputo left-sided fractional derivative (7) to the function g(x) with a = −κ,
Using the change of variables , then:
The integral (9) under this change of variables and by means of (4)-(6) takes the form
□
3 FRACTIONAL ORDER LOG BARRIER INTERIOR POINT ALGORITHM
Raising to the power p, the problem (1) can be rewritten in the alternative form as follows
where x ∈ ℝn . The problem (10) is similar to the following nonlinear optimization problem
where is the residual vector of regression and ℓ p −norm is defined of the form and .
The parameter values p most commonly used to the problem (11) are p = 1, p = 2 and p → ∞. Linear programming procedures Charnes et al. (1955); Oliveira et al. (2000); Oliveira & Lyra (2004) can be used for p = 1 and p → ∞ cases. For p = 2, the direct solution is given by x = (A ⊺ A)−1 A ⊺ b. For other values of p, unconstrained minimization procedures can be used Dennis Jr. & Schnabel (1983); Li (1993); Cantante et al. (2012).
Considering the unrestricted residual term r i as the difference between two nonnegative variables u i and v i: r i = u i −v i , with u i and v i being defined by
for all i = 1, 2, . . . , m, then |r i | = u i + v i and the problem (11) can be converted into a convex programming problem Charnes et al. (1955); Cantante et al. (2012):
where denote the set of m-dimensional nonnegative vectors.
Interior point methods are widely used to solve convex optimization problems because of their good performance in practice Biegler (2010); Gondzio (2012); Lilian et al. (2016). By adding a logarithmic barrier function to the objective function in (13), the barrier problem is given by
where µ > 0 is the barrier parameter. An optimal solution of (13) can be found by solving a series of barrier problems of the form (14) while µ is decreasing and going to zero. The Lagrangian function associated with the problem (14) is
where is a Lagrange multiplier vector.
Let ϕ and φ be the functions given by
and
then the Lagrangian function can be written as
The fractional order log barrier interior point algorithm, involving the Caputo fractional derivative, is based on replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions (𝛻L = 𝛻ϕ + 𝛻φ = 0), of the following form
where 𝛻x , 𝛻λ , 𝛻u and 𝛻v represent the gradient with respect to x, λ , u and v, respectively, and represent the fractional order gradient with respect to u and v, respectively, given by Caputo fractional derivative (7).
Furthermore, , where diag(w) denote the diagonal matrix from a vector w, and . The ith component of the vector g α is . By Property 1, for i = 1, 2, . . . , m, (g α )i is given by
The nonlinear system of equations (19) can be rewritten in the alternative form as follows
where .
For given µ > 0, if α = 1 in the system (21), then it recovers the first order optimality conditions, since , and has a unique solution . When the classical gradient is replaced by the fractional one (0 < α < 1), then if the system (21) has a solution , it can be called fractional solution and .
For given µ > 0 and 0 < α ≤ 1, suppose that the system (21) has a solution. Given an initial point (x 0, λ 0, u 0, v 0) such that , where denote the set of m-dimensional positive vectors, and Ax 0 + u 0 −v 0 = b, by applying Newton method to the system (21), then the search direction (Δx, Δλ, Δu, Δv) can be found by solving the following Newton system
where
Im is the identity matrix of order m, D u = G α + λI m +UH α , D v = G α −λI m +VH α , where G α = diag(g α ), H α = diag(h α ), where h α ∈ ℝm and for i = 1, 2, . . . , m, the ith component of the vector h α is given by
The (h α )i component is obtained by evaluating the derivative
and taking (4) into account, Γ(p + 1 −α) can be rewritten in the form (p−α)Γ(p−α). So, (h α )i in equation (24) can be obtained.
For given µ > 0, the new iterate barrier parameter is updated in the following form Wright (1996):
where β > 1 is added to control your decay and to improve the convergence process.
To keep off the next û and in the border region is required to shorten the step size α uv by introducing a parameter σ, with 0 < σ < 1, which is often a value close to 1, as follows Biegler (2010); Vanderbei (2020):
Given a point (x, λ, u, v), the new iterate is given by
3.1 Search direction
The search direction (∆x, ∆λ, ∆u, ∆v) is unique and can be obtained by solving the Newton system (22):
Isolating ∆u in (30),
Substituting (32) into (31) and isolating ∆v,
where
Substituting (32) into (29),
Now, substituting (33) into (35),
where
Then, ∆x can be obtained by replacing ∆λ given by (36) into (28). In this case, ∆x can be written as follows
where
3.2 Initialization
The initial point (x 0, λ 0, u 0, v 0) is chosen so that and , and is given by Coleman & Li (1992); Oliveira & Cantante (2004):
where ε 1 is a positive value close to zero.
The x 0 choice is the direct solution to the problem (11) for p = 2, and has proved to be a good choice in numerical experiments performed for polynomial regression models in the ℓ p −norm for values of p reasonably closer to 2 Cantante et al. (2012); Oliveira & Cantante (2004); Cantane (2004); Grigoletto (2011).
3.3 Termination criteria
The fractional order log barrier interior point algorithm terminates if at least one of the two following convergence criteria is satisfied:
where m is the number of rows of the matrix A, and
where ε 2 and ε 3 are positive values close to zero, and is the new iterate, given by
where is obtained by considering the new iterates and in equation (19).
3.4 Algorithm
The fractional order log barrier interior point algorithm works as follows. For given p, α, µ, β, σ, ε 1, ε 2, ε 3, δ, and initial point (x, λ, u, v) from (40), the Newton system (22) is solved, and its solution (Δx, Δλ, Δu, Δv) is obtained. The step size α uv is determined by (27) and the new iterate is obtained. The new barrier parameter is updated by (26). This procedure repeat until the termination criterion is satisfied.
In particular, when α = 1, the fractional order log barrier interior point algorithm recover the classical log barrier interior point algorithm.
Fractional order log barrier interior point algorithm for polynomial regression models in the ℓ p −norm.
4 NUMERICAL EXPERIMENTS
In order to illustrate the proposed fractional order log barrier interior point algorithm to solve polynomial regression models in the ℓ p −norm, numerical experiments were performed to compare it with the classical log barrier interior point algorithm. The implementation of the fractional order log barrier interior point algorithm was performed under Windows 10 and Matlab (R2016a) running on a desktop with 2.20 GHz Intel Core i5-5200 central processing unit (CPU) and 4G random-access memory (RAM).
A data set containing daily interest rates observed over 40 years was used for the analysis of polynomial regressions. The data set contains 10958 observations: , where a i represents the day of the week for a given date and b i the interest rate (in percentage) for the specific day a i . The ai values were normalized to the [0, 1] interval of the real line ℝ to avoid numerical stability problems of the algorithm Oliveira & Cantante (2004); Cantane (2004). Figure 1 shows the data set.
For the numerical results below, a comparison of the fractional order log barrier interior point algorithm for different values of the order α considers: “It.”, the number of iterations; “Res. Err.”, the residual error, given by ∥r∥p = ∥b−Ax∥p ; and “Time (s)”, the CPU time in seconds.
The number of iterations and the value for the residual error for different values of α and p, obtained from the fractional order log barrier interior point algorithm for linear regression (n = 2) in the ℓ p −norm, are shown in Tables 1-3. If the algorithm does not converge (divergence of Newton’s method for solving the nonlinear system of equations (21): , or if the algorithm fails (ill conditioned matrix), then the results will be shown in Tables 1-3 as “-” or “*”, respectively. When α = 1, the results of the classical log barrier interior point algorithm are recovered.
Numerical results for linear regression in the ℓ p −norm (p = 1.1, 1.2, 1.3) for different values of α.
Numerical results for linear regression in the ℓ p −norm (p = 1.4, 1.5, 1.6) for different values of α.
Numerical results for linear regression in the ℓ p −norm (p = 1.7, 1.8, 1.9) for different values of α.
The algorithm failed for α = 0.5 and p = 1.1, as can be seen in Table 1, but for example, for α = 0.55 and p = 1.1, the algorithm converges.
According to the Tables 1-3, the fractional order log barrier interior point algorithm for linear regression models in the ℓ p −norm (p = 1.1, 1.2, . . . , 1.7) yield smaller residual error when α = p − 0.8. It is a relationship between p and α, which can also be written as p−α = 0.8. In this cases, the algorithm takes more iterations until convergence.
For linear regression in the ℓ 1.2 −norm, a smaller residual error, given by 5242, was obtained with 16 iterations when α = 0.39 (p−α = 0.81). For α = 0.36, the residual error is 5256 and the number of iterations is 22. For α = 0.37, the residual error is 5252 and the number of iterations is 28. For α = 0.38, the residual error is 5247 and the number of iterations is 14.
For linear regression in the ℓ 1.8 −norm, a smaller residual error, given by 509.967, was obtained with 26 iterations when α = 0.96 (p − α = 0.84). For α = 0.94, the residual error is 510.011 and the number of iterations is 28. For α = 0.95, the residual error is 509.987 and the number of iterations is 16. For α = 0.97, the residual error is 510.015 and the number of iterations is 17.
For linear regression in the ℓ 1.9 −norm, the algorithm does not converge for α ≤ 0.92. For α = 0.93, the residual error is 403.48 and the number of iterations is 4. For α = 0.95, the residual error is 403.44 and the number of iterations is 5. For α = 0.97, the residual error is 403.42 and the number of iterations is 6. For α = 0.99, the residual error is 403.41 and the number of iterations is 6. When α = 0.99, the algorithm takes fewer iterations than when α = 1 (See Table 3), and the residual error is very close to the residual error obtained when α = 1.
The number of iterations, the residual error and the CPU time for (n − 1)th degree polynomial regressions (n = 2, 5, 10, 15), in the ℓ 1.3 −norm, for different values of α, obtained from the fractional order log barrier interior point algorithm, are given in the following tables.
The smallest residual errors for n = 2 and n = 5 were obtained with a greater number of iterations when α = 0.5 (see Table 4), while for n = 10 and n = 15 (see Table 5), the residual errors were smallest for α = 0.4.
The performance of the fractional order log barrier interior point algorithm appears to be computationally consistent. For example, the values of the residual error at each iteration k for linear regression in the ℓ 1.3 −norm are shown in Figure 2.
In the Figure 2, one can observe that the use of any one of the fractional order derivatives (α = 0, 4, 0.5, . . . , 0.9) produces smaller residual errors than with the use of the integer order derivative (α = 1). The use of the fractional derivatives does not interfere in the iterations running times, that compare similarly.
The approximate solution for (n − 1)th degree polynomial regression in the ℓ p −norm, obtained at the kth iteration of the fractional order log barrier interior point algorithm and given by , provides the coefficients of the (n − 1)th degree polynomial that approximately fits the data set, given by . In particular, the approximate solutions x k , obtained at the kth iteration of the fractional order log barrier interior point algorithm, for (n − 1)th degree polynomial regressions (n = 2, 5, 10, 15) in the ℓ 1.3 −norm, with smaller residual errors are:
i.e., at the 20th iteration of the algorithm with α = 0.5, the linear regression (n = 2) provides the polynomial coefficients from the above vector x 20. So, the linear regression is given by f (x) = 4.5 + 5.9 x.
The polynomials from the solutions described are shown in Figure 3.
5 CONCLUSIONS
The fractional order log barrier interior point algorithm for polynomial regression models in the ℓ p −norm (1 < p < 2), obtained by replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions, was investigated in this paper. This algorithm appears to be computationally consistent but depending on the fractional order α, the algorithm does not converge or fails. The numerical experiments showed that the use of the fractional derivatives can be beneficial for solving optimization problems. In future, further studies on this subject could be undertaken.
Acknowledgments
The author thanks the reviewers for their numerous helpful suggestions.
References
- ANDREWS GE, ASKEY R & ROY R. 1999. Special Functions. Cambridge: Cambridge University Press.
- BARD Y. 1974. Nonlinear Parameter Estimation. New York: Academic Press.
- BIEGLER LT. 2010. Nonlinear Programming. Philadelphia: SIAM.
- CANTANE DR. 2004. Métodos de Pontos Interiores Aplicados ao Problema de Regressão pela Norma Lp. São Carlos: Dissertação de Mestrado, Instituto de Ciências Matemáticas e de Computação, Universidade de São Paulo.
- CANTANTE DR, GRIGOLETTO EC & OLIVEIRA ARL. 2012. Método de pontos interiores barreira logarítmica preditor-corretor especializado para o problema de regressão pela norma Lp. Tendências em Matemática Aplicada e Computacional - TEMA, 13: 219-231.
- CHARNES A, COOPER WW & FERGUSON R. 1955. Optimal estimation of executive compensation by linear programming. Management Sci., 1: 138-151.
- CHATTERJEE S & HADI AS. 2012. Regression Analysis by Example. Hoboken: Wiley.
- CHEN Y, GAO Q, WEI Y & WANG Y. 2017a. Study on fractional order gradient methods. Appl. Math. Comput., 314: 310-321.
- CHEN Y, GAO Q, WEI Y & WANG Y. 2017b. Study on fractional order gradient methods. Appl. Math. Comput. , 314: 310-321.
- COLEMAN TF & LI Y. 1992. A globally and quadratically convergent affine scaling method for linear ℓ1 problems. Math. Programming, 56: 189-222.
- DALIR M & BASHOUR M. 2010. Applications of fractional calculus. Appl. Math. Sci., 4: 1021-1032.
- DENNIS JR JE & SCHNABEL RB. 1983. Numerical Methods for Unconstrained Optimization. NJ: Prentice-Hall.
- ERDÉLYI A, MAGNUS W, OBERHETTINGER F & TRICOMI F. 1981. Higher Transcendental Functions. vol. I-III. Melbourne, Florida: Krieger Pub.
- FORSYTHE AB. 1972. Robust estimation of straight line regression coefficients by minimizing pth power deviations. Technometrics, 14: 159-166.
- GONDZIO J. 2012. Interior point methods 25 years later. Eur. J. Oper. Res., 218: 587-601.
- GRIGOLETTO EC. 2011. Implementação Eficiente dos Métodos de Pontos Interiores Especializados para o Problema de Regressão pela Norma Lp. Campinas: Dissertação de Mestrado, Instituto de Matemática, Estatística e Computação Científica, Universidade Estadual de Campinas.
- GRIGOLETTO EC & OLIVEIRA ARL. 2020. Fractional order gradient descent algorithm. In: Proceeding Series of the Brazilian Society of Computational and Applied Mathematics - XXXIX CNMAC. p. 010387. São Carlos: SBMAC.
- KILBAS AA, SRIVASTAVA HM & TRUJILLO JJ. 2006. Theory and Applications of Fractional Differential Equations. Amsterdam: Elsevier.
- LI Y. 1993. A globally convergent method for ℓp problems. SIAM J. Optimiz., 3: 609-629.
- LILIAN FB, OLIVEIRA ARL & GHIDINI CTLS. 2016. Use of continued iteration on the reduction of iterations of the interior point method. Pesqui. Oper., 36: 487-501.
- MOHAMMADZADEH A & KAYACAN E. 2020. A novel fractional-order type-2 fuzzy control method for online frequency regulation in ac microgrid. Eng. Appl. Artif. Intell., 90: 103483.
- OLDHAM KB & SPANIER J. 1974. The Fractional Calculus. New York: Academic Press .
- OLIVEIRA ARL & CANTANTE DR. 2004. Método de pontos interiores aplicados ao problema de regressão pela norma Lp. Tendências em Matemática Aplicada e Computacional - TEMA , 5: 281-291.
- OLIVEIRA ARL & LYRA C. 2004. Interior point methods for the polynomial ℓ∞ fitting problems. Int. Trans. Oper. Res., 11: 309322.
- OLIVEIRA ARL, NASCIMENTO MA & LYRA C. 2000. Efficient implementation and benchmark of interior point methods for the polynomial ℓ1 fitting problem. Comput. Statist. Data Anal., 35: 119-135.
- RICE JR & WHITE JS. 1964. Norms for Smoothing and Estimation. SIAM Review, 6: 243-256.
- SEVAUX M & MINEUR Y. 2007. A curve-fitting genetic algorithm for a styling application. Eur. J. Oper. Res. , 179: 895-905.
- SÜLI E & MAYERS DF. 2003. An Introduction to Numerical Analysis. New York: Cambridge University Press.
- VANDERBEI RJ. 2020. Linear Programming: Foundations and Extensions. Springer Nature.
- WANG J, WEN Y, GOU Y, YE Z & CHEN H. 2017a. Fractional-order gradient descent learning of BP neural networks with Caputo derivative. Neural Netw., 89: 19-30.
- WANG J, WEN Y, GOU Y, YE Z & CHEN H. 2017b. Fractional-order gradient descent learning of BP neural networks with Caputo derivative. Neural Netw. , 89: 19-30.
- WRIGHT SJ. 1996. Primal-Dual Interior-Point Methods. Philadelphia, USA: SIAM.
Publication Dates
-
Publication in this collection
25 Nov 2022 -
Date of issue
2022
History
-
Received
18 June 2021 -
Accepted
23 June 2022