Fractional Chebyshev functional link neural network‐optimization method for solving delay fractional optimal control problems with Atangana‐Baleanu derivative

Summary In this article, we propose a higher order neural network, namely the functional link neural network (FLNN), for the model of linear and nonlinear delay fractional optimal control problems (DFOCPs) with mixed control‐state constraints. We consider DFOCPs using a new fractional derivative wit...

Full description

Saved in:
Bibliographic Details
Published inOptimal control applications & methods Vol. 41; no. 3; pp. 808 - 832
Main Authors Kheyrinataj, Farzaneh, Nazemi, Alireza
Format Journal Article
LanguageEnglish
Published Glasgow Wiley Subscription Services, Inc 01.05.2020
Subjects
Online AccessGet full text
ISSN0143-2087
1099-1514
DOI10.1002/oca.2572

Cover

More Information
Summary:Summary In this article, we propose a higher order neural network, namely the functional link neural network (FLNN), for the model of linear and nonlinear delay fractional optimal control problems (DFOCPs) with mixed control‐state constraints. We consider DFOCPs using a new fractional derivative with nonlocal and nonsingular kernel that was recently proposed by Atangana and Baleanu. The derivative possesses more important characteristics that are very useful in modelling. In the proposed method, a fractional Chebyshev FLNN is developed. At the first step, the delay problem is transformed to a nondelay problem, using a Padé approximation. The necessary optimality condition is stated in a form of fractional two‐point boundary value problem. By applying the fractional integration by parts and by constructing an error function, we then define an unconstrained minimization problem. In the optimization problem, trial solutions for state, co‐state and control functions are utilized where these trial solutions are constructed by using single‐layer fractional Chebyshev neural network model. We then minimize the error function using an unconstrained optimization scheme based on the gradient descent algorithm for updating the network parameters (weights and bias) associated with all neurons. To show the effectiveness of the proposed neural network, some numerical results are provided.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0143-2087
1099-1514
DOI:10.1002/oca.2572