Open-access FRACTIONAL ORDER LOG BARRIER INTERIOR POINT ALGORITHM FOR POLYNOMIAL REGRESSION IN THE p -NORM

ABSTRACT

Fractional calculus is the branch of mathematics that studies the several possibilities of generalizing the derivative and integral of a function to noninteger order. Recent studies found in literature have confirmed the importance of fractional calculus for minimization problems. However, the study of fractional calculus in interior point methods for solving optimization problems is still new. In this study, inspired in applications of fractional calculus in many fields, was developed the so-called fractional order log barrier interior point algorithm by replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions of Karush-Kuhn-Tucker to solve polynomial regression models in the p norm for 1 < p < 2. Finally, numerical experiments are performed to illustrate the proposed algorithm.

Keywords: nonlinear programming; polynomial regression; ℓp−norm; interior point method; fractional derivative

1 INTRODUCTION

It is fundamentally important to make predictions based upon scientific data. The problem of fitting curves to data points has many practical applications Bard (1974); Sevaux & Mineur (2007); Chatterjee & Hadi (2012).

Given a set of m data points in 2:ai,bii=1m, where a i is an argument value and b i a corresponding dependent value, with aiaj for all ij. The curve fitting procedure tries to build a linear or nonlinear function y = f (x), defined for all possible choices of x, that approximately fits the data set. The fitted curves to the data by f are most often chosen to be polynomials Süli & Mayers (2003).

Let y = f (x) be a polynomial function of degree n− 1 of the form f(x)=x0+x1x++xn-1xn-1, the candidate function to fit data. The procedure to fit the polynomial function to data, in the p norm, determines the vector x=x0,x1,x2,,xn-1n, where the superscript T represents transpose, that minimizes the p norm of the residual error as follows

min Φ x = b - A x p , (1)

where ·p denotes p norm, b=b1,b2,,bmm and Am×n is a Vandermonde matrix, with rank (A) = m, which can be written as

A = 1 a 1 a 1 2 a 1 n - 1 1 a 2 a 2 2 a 2 n - 1 1 a m a m 2 a m n - 1 . (2)

This is a nonlinear (or linear) regression model if n > 2 (or n = 2). If m ≤ n, there is an (n − 1)th degree polynomial satisfying fai=bii=1,2,,m. If m > n, the problem cannot be exactly solved and one needs to find the vector x in (1).

In many industrial applications, missing data or anomalous values can arise as errors of measurement instruments during the data generation. In this cases, polynomial regression models in the p norm, with p2, are more robust Forsythe (1972). It is important to choose an appropriate value for p and several criteria for the choice of p have been studied Rice & White (1964).

The calculus of integrals and derivatives of arbitrary order, named as fractional calculus, was conceptualized in connection with the infinitesimal calculus since 1695 Oldham & Spanier (1974). Some of the application areas include: viscoelasticity, signal processing, probability, statistics, electrochemistry, diffusion in porous media, fluid flow, backpropagation training of neural networks, fuzzy control method, and so on Kilbas et al. (2006); Dalir & Bashour (2010); Mohammadzadeh & Kayacan (2020); Wang et al. (2017a); Chen et al. (2017a); Grigoletto & Oliveira (2020). Also, its importance in minimization problems has been confirmed by recent works found in literature. For example, Chen et al. Chen et al. (2017b) presented the fractional order gradient methods (FOGMs) by writing the Riemann-Liouville and Caputo fractional derivatives as Taylor series, and Wang et al. Wang et al. (2017b) proposed the fractional gradient descent method employing the Caputo derivative for the backpropagation training of neural networks. However, the study of fractional calculus in the field of interior point methods for solving optimization problems is still new.

The goal of this study is to investigate the fractional order log barrier interior point algorithm, involving the Caputo fractional derivative. It is based on replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions for solving polynomial regression models in the p norm for 1 < p < 2. Functions of the form g(x)=x+κp, with κ > 0 and 1 < p < 2, arise when solving problem 1), according to the approach which will be discussed throughout this study. The Caputo fractional derivative, in addition to generalize the integer order derivative, is a useful tool to obtain the derivative for functions as g(x), as p is on the interval 1 < p < 2, then it will be a non integer number.

This paper is organized as follows. Preliminary concepts of fractional calculus are presented in Section 2. In Section 3, the fractional order log barrier interior point algorithm for solving polynomial regression models in the p norm is discussed. Numerical experiments are performed to illustrate the proposed algorithm in Section 4, and the Section 5 contains the conclusions.

2 PRELIMINARY CONCEPTS OF FRACTIONAL CALCULUS

Some basic concepts and definitions involving special functions and the Caputo fractional derivatives will be presented in this section.

Definition 1. The gamma function Γ(z) is originally defined by Erdélyi et al. (1981); Andrews et al. (1999):

Γ z = 0 t z - 1 e - t d t z > 0 . (3)

For gamma function, the reduction formula

Γ z + 1 = z Γ z , (4)

holds. In particular, when z=n, then Γn+1=n!.

Definition 2. The beta function B(z, w) is defined by Erdélyi et al. (1981); Andrews et al. (1999):

B z , w = 0 ξ z - 1 1 + ξ - z - w d ξ z > 0 ; w > 0 . (5)

It is connected with the gamma function by the following relation

B z , w = Γ z Γ w Γ z + w z > 0 ; w > 0 . (6)

The fractional calculus basically defines several integral and derivative operators of arbitrary order. The Caputo fractional derivative is one of the fractional derivative operators Kilbas et al. (2006).

Definition 3. The Caputo left-sided fractional derivative with respect to x, of order 0 < α < 1, of functions f on a subset Ω = [a, b] of real axis =-,, where fC1a,b, denoted by Da+αxCfx, is given by

D a + α x C f x : = 1 Γ ( 1 - α ) a x x - τ - α f ' τ d τ x Ω . (7)

Note that the Caputo fractional derivative is a nonlocal fractional derivative, because it depends on the choice of the order α, function f (x), x, and also relies on the total effects on the interval [a, x]. Usually, this is called memory effect.

In particular, Da+1xCfx:=f 'x.

Property 1. Let g(x)=x+κp be defined for -κx<, with κ ≥ 0, and 1 < p < 2, and let 0 < α < 1. Then

D - κ + α x C x + κ p x = Γ p + 1 Γ p - α + 1 x + κ p - α . (8)

Proof. Applying the Caputo left-sided fractional derivative (7) to the function g(x) with a = −κ,

D - κ + α x C x + κ p x 1 Γ ( 1 - α ) - κ x x - τ - α p τ + κ p - 1 d τ . (9)

Using the change of variables ξ=τ+κx-τ, then:

τ = ξ x - κ 1 + ξ , d τ = x + κ 1 + ξ 2 d ξ , τ - κ ξ 0 and τ x ξ .

The integral (9) under this change of variables and by means of (4)-(6) takes the form

D - κ + α x C x + κ p x = p Γ ( 1 - α ) - κ x x - τ - α τ + κ p - 1 d τ = p Γ ( 1 - α ) 0 x + κ 1 + ξ - α ξ ( x + κ ) 1 + ξ p - 1 x + κ 1 + ξ 2 d ξ = p x + κ p - α Γ ( 1 - α ) 0 ξ p - 1 1 + ξ α - p - 1 d ξ = p x + κ p - α Γ ( 1 - α ) B p , 1 - α = p x + κ p - α Γ ( 1 - α ) Γ p Γ 1 - α Γ p - α + 1 = Γ p + 1 Γ p - α + 1 x + κ p - α .

3 FRACTIONAL ORDER LOG BARRIER INTERIOR POINT ALGORITHM

Raising Φx=b-Axp to the power p, the problem (1) can be rewritten in the alternative form as follows

min b - A x p p , (10)

where x ∈ ℝn . The problem (10) is similar to the following nonlinear optimization problem

min r p p s.t. A x + r = b , x n , (11)

where r=r1,r2,,rmm is the residual vector of regression and p norm is defined of the form rp=i=1mrip1p 1p< and r=maxri.

The parameter values p most commonly used to the problem (11) are p = 1, p = 2 and p → ∞. Linear programming procedures Charnes et al. (1955); Oliveira et al. (2000); Oliveira & Lyra (2004) can be used for p = 1 and p → ∞ cases. For p = 2, the direct solution is given by x = (A A)−1 A b. For other values of p, unconstrained minimization procedures can be used Dennis Jr. & Schnabel (1983); Li (1993); Cantante et al. (2012).

Considering the unrestricted residual term r i as the difference between two nonnegative variables u i and v i: r i = u i −v i , with u i and v i being defined by

u i = r i , if r i 0 , 0 , if r i < 0 , v i = 0 , if r i 0 , - r i , if r i < 0 , (12)

for all i = 1, 2, . . . , m, then |r i | = u i + v i and the problem (11) can be converted into a convex programming problem Charnes et al. (1955); Cantante et al. (2012):

min i = 1 m u i + v i p s.t. A x + u - v = b , x n , u + m , v + m , (13)

where +m denote the set of m-dimensional nonnegative vectors.

Interior point methods are widely used to solve convex optimization problems because of their good performance in practice Biegler (2010); Gondzio (2012); Lilian et al. (2016). By adding a logarithmic barrier function to the objective function in (13), the barrier problem is given by

min i = 1 m u i + v i p - μ ln u i - μ ln v i s.t. A x + u - v = b , x n , (14)

where µ > 0 is the barrier parameter. An optimal solution of (13) can be found by solving a series of barrier problems of the form (14) while µ is decreasing and going to zero. The Lagrangian function associated with the problem (14) is

L x , λ , u , v = i = 1 m u i + v i p - μ ln u i - μ ln v i + λ A x + u - v - b , (15)

where λ=λ1,λ2,,λmm is a Lagrange multiplier vector.

Let ϕ and φ be the functions given by

ϕ x , λ , u , v = i = 1 m u i + v i p , (16)

and

φ x , λ , u , v = - μ i = 1 m ln u i + ln v i + λ A x + u - v - b , (17)

then the Lagrangian function can be written as

L x , λ , u , v = ϕ x , λ , u , v + φ x , λ , u , v . (18)

The fractional order log barrier interior point algorithm, involving the Caputo fractional derivative, is based on replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions (𝛻L = 𝛻ϕ + 𝛻φ = 0), of the following form

α ϕ + φ = x ϕ + x φ λ ϕ + λ φ u α ϕ + u φ v α ϕ + v φ = A λ A x + u - v - b g α - μ U - 1 1 m + λ g α - μ V - 1 1 m - λ = 0 0 0 0 , (19)

where 𝛻x , 𝛻λ , 𝛻u and 𝛻v represent the gradient with respect to x, λ , u and v, respectively, uα and vα represent the fractional order gradient with respect to u and v, respectively, given by Caputo fractional derivative (7).

Furthermore, 1m=1,1,,1m, U=diagu, V=diagv, where diag(w) denote the diagonal matrix from a vector w, and gαm. The ith component of the vector g α is gαi=D-vi+αuiCui+vipui=D-ui+αviCui+vipvi. By Property 1, for i = 1, 2, . . . , m, (g α )i is given by

g α i = Γ p + 1 Γ p + 1 - α u i + v i p - α 1 < p < 2 ; 0 < α < 1 . (20)

The nonlinear system of equations (19) can be rewritten in the alternative form as follows

A λ A x + u - v - b U g α + λ V g α - λ = 0 0 μ m μ m , (21)

where μm=μ,μ,,μm.

For given µ > 0, if α = 1 in the system (21), then it recovers the first order optimality conditions, since limα1αϕ+φ=ϕ+φ=L, and has a unique solution xμ*,λμ*,uμ*,vμ*. When the classical gradient is replaced by the fractional one (0 < α < 1), then if the system (21) has a solution xμα*,λμα*,uμα*,vμα*, it can be called fractional solution and limα1xμα*,λμα*,uμα*,vμα*=xμ*,λμ*,uμ*,vμ*.

For given µ > 0 and 0 < α ≤ 1, suppose that the system (21) has a solution. Given an initial point (x 0, λ 0, u 0, v 0) such that u0++m, v0++m, where ++m denote the set of m-dimensional positive vectors, and Ax 0 + u 0v 0 = b, by applying Newton method to the system (21), then the search direction (Δx, Δλ, Δu, Δv) can be found by solving the following Newton system

0 A 0 0 A 0 m I m - I m 0 U D u U H α 0 - V V H α D v Δ x Δ λ Δ u Δ v = r 1 r 2 r 3 r 4 , (22)

where

r 1 = - A λ , r 2 = - A x - u + v + b , r 3 = - U g α + λ + μ m , r 4 = - V g α - λ + μ m , (23)

Im is the identity matrix of order m, D u = G α + λI m +UH α , D v = G α −λI m +VH α , where G α = diag(g α ), H α = diag(h α ), where h α ∈ ℝm and for i = 1, 2, . . . , m, the ith component of the vector h α is given by

h α i = Γ p + 1 Γ p - α u i + v i p - α - 1 . (24)

The (h α )i component is obtained by evaluating the derivative

h α i = u i g α i = p - α Γ p + 1 Γ p + 1 - α u i + v i p - α - 1 = v i g α i , (25)

and taking (4) into account, Γ(p + 1 −α) can be rewritten in the form (p−α)Γ(p−α). So, (h α )i in equation (24) can be obtained.

For given µ > 0, the new iterate barrier parameter μ^ is updated in the following form Wright (1996):

μ ^ = μ β , (26)

where β > 1 is added to control your decay and to improve the convergence process.

To keep off the next û and v^ in the border region is required to shorten the step size α uv by introducing a parameter σ, with 0 < σ < 1, which is often a value close to 1, as follows Biegler (2010); Vanderbei (2020):

α u v = min σ min Δ u i < 0 u i Δ u i , σ min Δ v i < 0 v i Δ v i , σ . (27)

Given a point (x, λ, u, v), the new iterate x^, λ^, u^, v^ is given by

x ^ = x + α u v Δ x , λ ^ = λ + α u v Δ λ , u ^ = u + α u v Δ u , v ^ = v + α u v Δ v .

3.1 Search direction

The search direction (∆x,λ,u,v) is unique and can be obtained by solving the Newton system (22):

A Δ λ = r 1 , (28)

A Δ x + Δ u - Δ v = r 2 , (29)

U Δ λ + D u Δ u + U H α Δ v = r 3 , (30)

- V Δ λ + V H α Δ u + D v Δ v = r 4 . (31)

Isolating ∆u in (30),

Δ u = D u - 1 r 3 - U Δ λ - U H α Δ v . (32)

Substituting (32) into (31) and isolating ∆v,

Δ v = D 1 - 1 r 4 + V Δ λ + V H α D u - 1 U Δ λ - r 3 , (33)

where

D 1 = D v - U V H α 2 D u - 1 . (34)

Substituting (32) into (29),

A Δ x + D u - 1 r 3 - U Δ λ - U H α D u - 1 + I m Δ v = r 2 . (35)

Now, substituting (33) into (35),

Δ λ = D 2 - 1 r 2 - A Δ x - D u - 1 r 3 + D 1 - 1 U H α D u - 1 + I m r 4 - V H α D u - 1 r 3 , (36)

where

D 2 = - U D u - 1 - V D 1 - 1 U H α D u - 1 + I m 2 . (37)

Then, ∆x can be obtained by replacing ∆λ given by (36) into (28). In this case, ∆x can be written as follows

Δ x = A D 2 - 1 A - 1 r ¯ , (38)

where

r ¯ = - r 1 + A D 2 - 1 r 2 - D u - 1 r 3 + D 1 - 1 U H α D u - 1 + I m r 4 - V H α D u - 1 r 3 . (39)

3.2 Initialization

The initial point (x 0, λ 0, u 0, v 0) is chosen so that u0++m, v0++m and Ax0+u0-v0=b, and is given by Coleman & Li (1992); Oliveira & Cantante (2004):

x 0 = A A - 1 A b , r 0 = b - A x 0 , λ 0 = δ r 0 r 0 ,with δ = 0 . 975 , u i 0 = r i 0 + ϵ 1 , if r i 0 0 , ϵ 1 , if r i 0 < 0 , for i = 1 , 2 , , m , v i 0 = ϵ 1 , if r i 0 0 , ϵ 1 - r i 0 , if r i 0 < 0 , for i = 1 , 2 , , m , (40)

where ε 1 is a positive value close to zero.

The x 0 choice is the direct solution to the problem (11) for p = 2, and has proved to be a good choice in numerical experiments performed for polynomial regression models in the p norm for values of p reasonably closer to 2 Cantante et al. (2012); Oliveira & Cantante (2004); Cantane (2004); Grigoletto (2011).

3.3 Termination criteria

The fractional order log barrier interior point algorithm terminates if at least one of the two following convergence criteria is satisfied:

N = α ϕ + φ 1 + x + u + v + λ ( 2 m ) ϵ 2 , (41)

where m is the number of rows of the matrix A, and

N ^ - N ϵ 3 , (42)

where ε 2 and ε 3 are positive values close to zero, and N^ is the new iterate, given by

N ^ = α ϕ ^ + φ ^ 1 + x ^ + u ^ + v ^ + λ ^ ( 2 m ) ,

where αϕ^+φ^ is obtained by considering the new iterates x:=x^, λ:=λ^, u:=u^, v:=v^ and μ:=μ^ in equation (19).

3.4 Algorithm

The fractional order log barrier interior point algorithm works as follows. For given p, α, µ, β, σ, ε 1, ε 2, ε 3, δ, and initial point (x, λ, u, v) from (40), the Newton system (22) is solved, and its solution (Δx, Δλ, Δu, Δv) is obtained. The step size α uv is determined by (27) and the new iterate x^, λ^, u^, v^ is obtained. The new barrier parameter μ^ is updated by (26). This procedure repeat until the termination criterion is satisfied.

In particular, when α = 1, the fractional order log barrier interior point algorithm recover the classical log barrier interior point algorithm.

Algorithm 1
Fractional order log barrier interior point algorithm for polynomial regression models in the p norm.

4 NUMERICAL EXPERIMENTS

In order to illustrate the proposed fractional order log barrier interior point algorithm to solve polynomial regression models in the p norm, numerical experiments were performed to compare it with the classical log barrier interior point algorithm. The implementation of the fractional order log barrier interior point algorithm was performed under Windows 10 and Matlab (R2016a) running on a desktop with 2.20 GHz Intel Core i5-5200 central processing unit (CPU) and 4G random-access memory (RAM).

A data set containing daily interest rates observed over 40 years was used for the analysis of polynomial regressions. The data set contains 10958 observations: ai,bii=110958, where a i represents the day of the week for a given date and b i the interest rate (in percentage) for the specific day a i . The ai values were normalized to the [0, 1] interval of the real line ℝ to avoid numerical stability problems of the algorithm Oliveira & Cantante (2004); Cantane (2004). Figure 1 shows the data set.

Figure 1
Daily changes in the interest rate.

For the numerical results below, a comparison of the fractional order log barrier interior point algorithm for different values of the order α considers: “It.”, the number of iterations; “Res. Err.”, the residual error, given by ∥rp = ∥b−Axp ; and “Time (s)”, the CPU time in seconds.

The number of iterations and the value for the residual error for different values of α and p, obtained from the fractional order log barrier interior point algorithm for linear regression (n = 2) in the p norm, are shown in Tables 1-3. If the algorithm does not converge (divergence of Newton’s method for solving the nonlinear system of equations (21): b-Axk+1pb-Axkp), or if the algorithm fails (ill conditioned matrix), then the results will be shown in Tables 1-3 as “-” or “*”, respectively. When α = 1, the results of the classical log barrier interior point algorithm are recovered.

Table 1
Numerical results for linear regression in the p norm (p = 1.1, 1.2, 1.3) for different values of α.

Table 2
Numerical results for linear regression in the p norm (p = 1.4, 1.5, 1.6) for different values of α.

Table 3
Numerical results for linear regression in the p norm (p = 1.7, 1.8, 1.9) for different values of α.

The algorithm failed for α = 0.5 and p = 1.1, as can be seen in Table 1, but for example, for α = 0.55 and p = 1.1, the algorithm converges.

According to the Tables 1-3, the fractional order log barrier interior point algorithm for linear regression models in the p norm (p = 1.1, 1.2, . . . , 1.7) yield smaller residual error when α = p − 0.8. It is a relationship between p and α, which can also be written as p−α = 0.8. In this cases, the algorithm takes more iterations until convergence.

For linear regression in the 1.2 norm, a smaller residual error, given by 5242, was obtained with 16 iterations when α = 0.39 (p−α = 0.81). For α = 0.36, the residual error is 5256 and the number of iterations is 22. For α = 0.37, the residual error is 5252 and the number of iterations is 28. For α = 0.38, the residual error is 5247 and the number of iterations is 14.

For linear regression in the 1.8 norm, a smaller residual error, given by 509.967, was obtained with 26 iterations when α = 0.96 (p − α = 0.84). For α = 0.94, the residual error is 510.011 and the number of iterations is 28. For α = 0.95, the residual error is 509.987 and the number of iterations is 16. For α = 0.97, the residual error is 510.015 and the number of iterations is 17.

For linear regression in the 1.9 norm, the algorithm does not converge for α ≤ 0.92. For α = 0.93, the residual error is 403.48 and the number of iterations is 4. For α = 0.95, the residual error is 403.44 and the number of iterations is 5. For α = 0.97, the residual error is 403.42 and the number of iterations is 6. For α = 0.99, the residual error is 403.41 and the number of iterations is 6. When α = 0.99, the algorithm takes fewer iterations than when α = 1 (See Table 3), and the residual error is very close to the residual error obtained when α = 1.

The number of iterations, the residual error and the CPU time for (n − 1)th degree polynomial regressions (n = 2, 5, 10, 15), in the 1.3 norm, for different values of α, obtained from the fractional order log barrier interior point algorithm, are given in the following tables.

The smallest residual errors for n = 2 and n = 5 were obtained with a greater number of iterations when α = 0.5 (see Table 4), while for n = 10 and n = 15 (see Table 5), the residual errors were smallest for α = 0.4.

Table 4
Numerical results for polynomial regressions (n = 2, 5) in the 1.3 norm.

Table 5
Numerical results for polynomial regressions (n = 10, 15) in the 1.3 norm.

The performance of the fractional order log barrier interior point algorithm appears to be computationally consistent. For example, the values of the residual error at each iteration k for linear regression in the 1.3 norm are shown in Figure 2.

Figure 2
Linear regression in the 1.3norm.

In the Figure 2, one can observe that the use of any one of the fractional order derivatives (α = 0, 4, 0.5, . . . , 0.9) produces smaller residual errors than with the use of the integer order derivative (α = 1). The use of the fractional derivatives does not interfere in the iterations running times, that compare similarly.

The approximate solution for (n − 1)th degree polynomial regression in the p norm, obtained at the kth iteration of the fractional order log barrier interior point algorithm and given by xk=x0k,x1k,x2k,,xn-1k, provides the coefficients of the (n − 1)th degree polynomial that approximately fits the data set, given by f(x)=x0+x1x++xn-1xn-1. In particular, the approximate solutions x k , obtained at the kth iteration of the fractional order log barrier interior point algorithm, for (n − 1)th degree polynomial regressions (n = 2, 5, 10, 15) in the 1.3 norm, with smaller residual errors are:

x 20 = x 0 20 x 1 20 = 4 . 5 E + 00 5 . 9 E + 00 n = 2 ; α = 0 . 5 ,

i.e., at the 20th iteration of the algorithm with α = 0.5, the linear regression (n = 2) provides the polynomial coefficients from the above vector x 20. So, the linear regression is given by f (x) = 4.5 + 5.9 x.

x 20 = x 0 20 x 1 20 x 2 20 x 3 20 x 4 20 = 5 . 3 E + 00 - 3 . 4 E + 01 1 . 9 E + 02 - 2 . 8 E + 02 1 . 2 E + 02 n = 5 ; α = 0 . 5 , x 2 = x 0 2 x 1 2 x 2 2 x 3 2 x 4 2 x 5 2 x 6 2 x 7 2 x 8 2 x 9 2 = 1 . 8 E + 00 1 . 7 E + 02 - 3 . 7 E + 03 3 . 5 E + 04 - 1 . 6 E + 05 4 . 6 E + 05 - 7 . 6 E + 05 7 . 2 E + 00 - 3 . 7 E + 05 7 . 9 E + 04 ( n = 10 ; α = 0 . 4 ) , x 1 = x 0 1 x 1 1 x 2 1 x 3 1 x 4 1 x 5 1 x 6 1 x 7 1 x 8 1 x 9 1 x 10 1 x 11 1 x 12 1 x 13 1 x 14 1 = 3 . 6 E + 00 2 . 6 E + 00 - 1 . 9 E + 02 1 . 3 E + 04 - 2 . 1 E + 05 1 . 4 E + 06 - 5 . 6 E + 06 1 . 1 E + 07 - 1 . 3 E + 07 5 . 9 E + 06 - 2 . 6 E + 06 1 . 2 E + 07 - 1 . 9 E + 07 1 . 2 E + 07 - 2 . 8 E + 06 n = 15 ; α = 0 . 4 .

The polynomials from the solutions described are shown in Figure 3.

Figure 3
Polynomial regressions in the 1.3 norm.

5 CONCLUSIONS

The fractional order log barrier interior point algorithm for polynomial regression models in the p norm (1 < p < 2), obtained by replacing some integer derivatives for the corresponding fractional ones on the first order optimality conditions, was investigated in this paper. This algorithm appears to be computationally consistent but depending on the fractional order α, the algorithm does not converge or fails. The numerical experiments showed that the use of the fractional derivatives can be beneficial for solving optimization problems. In future, further studies on this subject could be undertaken.

Acknowledgments

The author thanks the reviewers for their numerous helpful suggestions.

References

  • ANDREWS GE, ASKEY R & ROY R. 1999. Special Functions. Cambridge: Cambridge University Press.
  • BARD Y. 1974. Nonlinear Parameter Estimation. New York: Academic Press.
  • BIEGLER LT. 2010. Nonlinear Programming. Philadelphia: SIAM.
  • CANTANE DR. 2004. Métodos de Pontos Interiores Aplicados ao Problema de Regressão pela Norma Lp. São Carlos: Dissertação de Mestrado, Instituto de Ciências Matemáticas e de Computação, Universidade de São Paulo.
  • CANTANTE DR, GRIGOLETTO EC & OLIVEIRA ARL. 2012. Método de pontos interiores barreira logarítmica preditor-corretor especializado para o problema de regressão pela norma Lp. Tendências em Matemática Aplicada e Computacional - TEMA, 13: 219-231.
  • CHARNES A, COOPER WW & FERGUSON R. 1955. Optimal estimation of executive compensation by linear programming. Management Sci., 1: 138-151.
  • CHATTERJEE S & HADI AS. 2012. Regression Analysis by Example. Hoboken: Wiley.
  • CHEN Y, GAO Q, WEI Y & WANG Y. 2017a. Study on fractional order gradient methods. Appl. Math. Comput., 314: 310-321.
  • CHEN Y, GAO Q, WEI Y & WANG Y. 2017b. Study on fractional order gradient methods. Appl. Math. Comput. , 314: 310-321.
  • COLEMAN TF & LI Y. 1992. A globally and quadratically convergent affine scaling method for linear ℓ1 problems. Math. Programming, 56: 189-222.
  • DALIR M & BASHOUR M. 2010. Applications of fractional calculus. Appl. Math. Sci., 4: 1021-1032.
  • DENNIS JR JE & SCHNABEL RB. 1983. Numerical Methods for Unconstrained Optimization. NJ: Prentice-Hall.
  • ERDÉLYI A, MAGNUS W, OBERHETTINGER F & TRICOMI F. 1981. Higher Transcendental Functions. vol. I-III. Melbourne, Florida: Krieger Pub.
  • FORSYTHE AB. 1972. Robust estimation of straight line regression coefficients by minimizing pth power deviations. Technometrics, 14: 159-166.
  • GONDZIO J. 2012. Interior point methods 25 years later. Eur. J. Oper. Res., 218: 587-601.
  • GRIGOLETTO EC. 2011. Implementação Eficiente dos Métodos de Pontos Interiores Especializados para o Problema de Regressão pela Norma Lp. Campinas: Dissertação de Mestrado, Instituto de Matemática, Estatística e Computação Científica, Universidade Estadual de Campinas.
  • GRIGOLETTO EC & OLIVEIRA ARL. 2020. Fractional order gradient descent algorithm. In: Proceeding Series of the Brazilian Society of Computational and Applied Mathematics - XXXIX CNMAC. p. 010387. São Carlos: SBMAC.
  • KILBAS AA, SRIVASTAVA HM & TRUJILLO JJ. 2006. Theory and Applications of Fractional Differential Equations. Amsterdam: Elsevier.
  • LI Y. 1993. A globally convergent method for ℓp problems. SIAM J. Optimiz., 3: 609-629.
  • LILIAN FB, OLIVEIRA ARL & GHIDINI CTLS. 2016. Use of continued iteration on the reduction of iterations of the interior point method. Pesqui. Oper., 36: 487-501.
  • MOHAMMADZADEH A & KAYACAN E. 2020. A novel fractional-order type-2 fuzzy control method for online frequency regulation in ac microgrid. Eng. Appl. Artif. Intell., 90: 103483.
  • OLDHAM KB & SPANIER J. 1974. The Fractional Calculus. New York: Academic Press .
  • OLIVEIRA ARL & CANTANTE DR. 2004. Método de pontos interiores aplicados ao problema de regressão pela norma Lp. Tendências em Matemática Aplicada e Computacional - TEMA , 5: 281-291.
  • OLIVEIRA ARL & LYRA C. 2004. Interior point methods for the polynomial ℓ fitting problems. Int. Trans. Oper. Res., 11: 309322.
  • OLIVEIRA ARL, NASCIMENTO MA & LYRA C. 2000. Efficient implementation and benchmark of interior point methods for the polynomial ℓ1 fitting problem. Comput. Statist. Data Anal., 35: 119-135.
  • RICE JR & WHITE JS. 1964. Norms for Smoothing and Estimation. SIAM Review, 6: 243-256.
  • SEVAUX M & MINEUR Y. 2007. A curve-fitting genetic algorithm for a styling application. Eur. J. Oper. Res. , 179: 895-905.
  • SÜLI E & MAYERS DF. 2003. An Introduction to Numerical Analysis. New York: Cambridge University Press.
  • VANDERBEI RJ. 2020. Linear Programming: Foundations and Extensions. Springer Nature.
  • WANG J, WEN Y, GOU Y, YE Z & CHEN H. 2017a. Fractional-order gradient descent learning of BP neural networks with Caputo derivative. Neural Netw., 89: 19-30.
  • WANG J, WEN Y, GOU Y, YE Z & CHEN H. 2017b. Fractional-order gradient descent learning of BP neural networks with Caputo derivative. Neural Netw. , 89: 19-30.
  • WRIGHT SJ. 1996. Primal-Dual Interior-Point Methods. Philadelphia, USA: SIAM.

Publication Dates

  • Publication in this collection
    25 Nov 2022
  • Date of issue
    2022

History

  • Received
    18 June 2021
  • Accepted
    23 June 2022
location_on
Sociedade Brasileira de Pesquisa Operacional Rua Mayrink Veiga, 32 - sala 601 - Centro, 20090-050 , Tel.: +55 21 2263-0499 - Rio de Janeiro - RJ - Brazil
E-mail: sobrapo@sobrapo.org.br
rss_feed Acompanhe os números deste periódico no seu leitor de RSS
Acessibilidade / Reportar erro