Open access peer-reviewed chapter

Conjugate Gradient Approach for Discrete Time Optimal Control Problems with Model-Reality Differences

Written By

Sie Long Kek, Sy Yi Sim, Wah June Leong and Kok Lay Teo

Submitted: 21 August 2019 Reviewed: 13 September 2019 Published: 22 November 2019

DOI: 10.5772/intechopen.89711

From the Edited Volume

Control Theory in Engineering

Edited by Constantin Volosencu, Ali Saghafinia, Xian Du and Sohom Chakrabarty

Chapter metrics overview

761 Chapter Downloads

View Full Metrics

Abstract

In this chapter, an efficient computation approach is proposed for solving a general class of discrete-time optimal control problems. In our approach, a simplified optimal control model, which is adding the adjusted parameters into the model used, is solved iteratively. In this way, the differences between the real plant and the model used are calculated, in turn, to update the optimal solution of the model used. During the computation procedure, the equivalent optimization problem is formulated, where the conjugate gradient algorithm is applied in solving the optimization problem. On this basis, the optimal solution of the modified model-based optimal control problem is obtained repeatedly. Once the convergence is achieved, the iterative solution approximates to the correct optimal solution of the original optimal control problem, in spite of model-reality differences. For illustration, both linear and nonlinear examples are demonstrated to show the performance of the approach proposed. In conclusion, the efficiency of the approach proposed is highly presented.

Keywords

  • optimal control
  • conjugate gradient
  • adjusted parameters
  • iterative solution
  • model-reality differences

1. Introduction

Optimal control problems are existing in engineering and natural sciences for so long, and the applications of the optimal control have been well defined in the literature [1, 2, 3, 4]. With the rapid evolution of computer technology, the development of the optimal control techniques is reached a mature level, from classical control to modern control, from proportional-integral-derivative (PID) control to feedback control, and from adaptive control to intelligent control [5, 6, 7, 8]. The studies in the optimal control field are still progressing, and attract the interest of, not only engineers and applied mathematicians but also biologists and financialists, to investigate and contribute to the optimal control theory.

In particular, the optimal control algorithm, which integrates system optimization and parameter estimation, gives a new insight into the control community. This algorithm is known as the integrated system optimization and parameter estimation (ISOPE), and its dynamic version is called the dynamic ISOPE (DISOPE). Both of these algorithms were introduced by Robert [9, 10, 11], and Robert and Becerra [12, 13, 14], respectively. The basic idea of DISOPE is applying the model-based optimal control, which has different structures and parameters compared to the original optimal control problem, to obtain the correct optimal solution of the original optimal control problem, in spite of model-reality differences. Recently, this algorithm has been extended to cover both deterministic and stochastic versions, and it is known as an integrated optimal control and parameter estimation (IOCPE) algorithm [15, 16]. On the other hand, the application of the optimization techniques, particularly, using the conjugate gradient method for solving the optimal control problem [17, 18, 19] has also been studied, where the open-loop control strategy is concerned [3, 8].

In this chapter, the conjugate gradient approach [17, 19] is employed to solve the linear model-based optimal control problem for obtaining the optimal solution of the original optimal control problem. In our approach, the simplified model, which is adding the adjusted parameters, is formulated initially. Then, an expanded optimal control problem, which combines the system dynamic and the cost function from the original optimal control problem into the simplified model, is introduced. By defining the Hamiltonian function and the augmented cost function, the corresponding necessary conditions for optimality are derived. Among these necessary conditions, a set of necessary conditions is for the modified model-based optimal control problem, a set of necessary conditions defines the parameter estimation problem, and a set of necessary conditions calculates the multipliers [15].

By virtue of the modified model-based optimal control problem, an equivalence optimization problem is defined, and the related gradient function is determined. With an initial control sequence, the initial gradient and the initial search direction are computed. Then, the control sequences are updated through the line search technique, where the gradient and search direction would satisfy the conjugacy condition [17, 18]. During the iteration, the state and the costate are updated by the control sequence obtained from the conjugate gradient approach. When the convergence is achieved within a tolerance given, the iterative solution approximates to the correct optimal solution of the original optimal control problem, in spite of model-reality differences. For illustration, examples of linear and nonlinear cases, which are damped harmonic oscillator [7] and continuous stirred-tank chemical tank [8], are studied.

The chapter is organized as follows. In Section 2, the problem statement is described in detail, where the original optimal control problem and the simplified model are discussed. In Section 3, the methodology used is further explained. The necessary conditions for optimality are derived, and the use of the conjugate gradient method is delivered in solving the equivalence optimization problem. In Section 4, examples of a damped harmonic oscillator and a continuous stirred-tank chemical reactor are studied. The results show the efficiency of the algorithm proposed. Finally, concluding remarks are made.

Advertisement

2. Problem statement

Consider a general class of the discrete-time nonlinear optimal control problem, given by

min u k J 0 u = φ x N N + k = 0 N 1 L x k u k k subject to x k + 1 = f x k u k k , x 0 = x 0 E1

where u k m , k = 0 , 1 , , N 1 , and x k n , k = 0 , 1 , , N , are the control sequences and the state sequences, respectively, while f : n × m × n represents the real plant, L : n × m × is the cost under summation, and φ : n × is the terminal cost. Here, J 0 is the scalar cost function, and x 0 is the known initial state vector. It is assumed that all functions in (1) are continuously differentiable with respect to their respective arguments.

This problem, which is referred to as Problem (P), is regarded as the real optimal control problem. Due on the complex and nonlinear structure, solving Problem (P) actually requires the efficient computation techniques. For this reason, the simplified model of Problem (P) is identified to be solved such that the true optimal solution of Problem (P) could be approximated. Hence, this simplified model-based optimal control problem is defined as follows:

min u k J 1 u = 1 2 x N T S N x N + γ N + k = 0 N 1 1 2 x k T Qx k + u k T Ru k + γ k subject to x k + 1 = Ax k + Bu k + α k , x 0 = x 0 E2

where γ k , k = 0 , 1 , , N , and α k , k = 0 , 1 , , N 1 , are introduced as the adjusted parameters, whereas A is an n × n transition matrix, and B is an n × m control coefficient matrix. Besides, S N and Q are n × n positive semi-definite matrices, and R is a m × m positive definite matrix. Here, J 1 is the scalar cost function.

Let this problem is referred to as Problem (M). It can be seen that, because of the different structures and parameters, only solving Problem (M) would not obtain the optimal solution of Problem (P) for not taking the adjusted parameters into account. Notice, adding the adjusted parameters into Problem (M) could let us calculate the differences between the real plant and the model used. On this basis, Problem (M) would be solved iteratively to give the correct optimal solution of Problem (P), in spite of model-reality differences.

Advertisement

3. System optimization with parameter estimation

Now, an expanded optimal control problem, which combines the real plant and the cost function in Problem (P) into Problem (M) and is referred to as Problem (E), is introduced by

min u k J 2 u = 1 2 x N T S N x N + γ N + k = 0 N 1 1 2 x k T Qx k + u k T Ru k + γ k + 1 2 r 1 u k v k 2 + 1 2 r 2 x k z k 2 subject to x k + 1 = Ax k + Bu k + α k , x 0 = x 0 1 2 z N T S N z N + γ N = φ z N N 1 2 z k T Qz k + v k T Rv k + γ k = L z k v k k Az k + Bv k + α k = f z k v k k v k = u k z k = x k E3

where v k m , k = 0 , 1 , , N 1 , and z k n , k = 0 , 1 , , N , are introduced to separate the sequences of control and state in the optimization problem from the respective signals in the parameter estimation problem, and denotes the usual Euclidean norm. The terms 1 2 r 1 u k v k 2 and 1 2 r 2 x k z k 2 with r 1 , r 2 are introduced to improve the convexity and to facilitate the convergence of the resulting iterative algorithm. Here, we clarify that the algorithm is designed such that the constraints v k = u k and z k = x k are satisfied upon termination of the iterations, assuming that convergence is achieved. Moreover, the state constraint z k and the control constraint v k are used for the computation of the parameter estimation and matching scheme, while the corresponding state constraint x k and control constraint u k are reserved for optimizing the model-based optimal control problem. Therefore, system optimization and parameter estimation are declared and mutually integrated.

3.1 Necessary conditions for optimality

Define the Hamiltonian function for Problem (E), given by:

H 2 (k)= 1 2 (x (k) T Qx(k)+u (k) T Ru(k))+γ(k) + 1 2 r 1 ||u(k)v(k)| | 2 + 1 2 r 2 ||x(k)z(k)| | 2 +p (k+1) T (Ax(k)+Bu(k)+α(k)) λ (k) T u(k)β (k) T x(k) E4

where λ k m , k = 0 , 1 , , N 1 , β k n , k = 0 , 1 , , N , and p k n , k = 0 , 1 , , N , are modifiers. Using this Hamiltonian function in (4), write the cost function in (3) to be the augmented cost function, that is,

J 2 (u)= 1 2 x (N) T S(N)x(N)+γ(N)+p (0) T x(0)p (N) T x(N) +ξ(N)(φ(z(N),N) 1 2 z (N) T S(N)z(N)γ(N)) + Γ T (x(N)z(N))+ k=0 N1 H 2 (k) p (k) T x(k)+λ (k) T v(k)+β (k) T z(k) +ξ(k)(L(z(k),v(k),k) 1 2 (z (k) T Qz(k)+v (k) T Rv(k))γ(k)) +μ (k) T (f(z(k),v(k),k)Az(k)Bv(k)α(k)) E5

where p k , ξ k , λ k , β k , μ k and Γ are the appropriate multipliers to be determined later.

Applying the calculus of variation [7, 9, 11, 13, 15] to the augmented cost function in (5), the following necessary conditions for optimality are obtained:

(a) Stationary condition:

Ru k + B T p k + 1 λ k + r 1 u k v k = 0 E6

(b) Co-state equation:

p k = Qx k + A T p k + 1 β k + r 2 x k z k E7

(c) State equation:

x k + 1 = Ax k + Bu k + α k E8

(d) Boundary conditions:

p N = S N x N + Γ and x 0 = x 0 E9

(e) Adjusted parameter equations:

φ z N N = 1 2 z N T S N z N + γ N E10
L z k v k k = 1 2 z k T Qz k + v k T Rv k + γ k E11
f z k v k k = Az k + Bv k + α k E12

(f) Modifier equations:

Γ = z N φ S N z N E13
λ k = v k L Rv k f v k B T p ̂ k + 1 E14
β k = z k L Qz k f z k A T p ̂ k + 1 E15

with ξ k = 1 and μ k = p ̂ k + 1 .

(g) Separable variables:

v k = u k , z k = x k , p ̂ k = p k . E16

Notice that for the optimality necessary conditions obtained above, they are divided into three sets of necessary conditions. The first set of necessary conditions in (6)(9) is the necessary conditions for the system optimization problem. The second set of necessary conditions in (10)(12) defines the parameter estimation problem. The third set of necessary conditions in (13)(15) provides the computation of multipliers. In fact, the necessary conditions, which are defined in (6)(9), are the optimality for the modified model-based optimal control problem, and the adjusted parameters, which are calculated from the necessary conditions in (10)(12), measure the differences between the real plant and the model used.

3.2 Modified model-based optimal control problem

As a consequence, the modified model-based optimal control problem, which is referred to as Problem (MM), is defined by

min u k J 3 u = 1 2 x N T S N x N + Γ T x N + γ N + k = 0 N 1 1 2 x k T Qx k + u k T Ru k + γ k + 1 2 r 1 u k v k 2 + 1 2 r 2 x k z k 2 λ k T u k β k T x k subject to x k + 1 = Ax k + Bu k + α k , x 0 = x 0 E17

with the specified α k , γ k , λ k , β k , Γ , v k and z k , where the boundary conditions are given by x 0 and p N with the specified multiplier Γ .

It is obvious that Problem (MM), which is derived from Problem (E), is a modification of optimal control problem and is also known as a modified linear quadratic regular problem. Importantly, the set of the necessary conditions in (6)(9) for Problem (E) is the necessary conditions that are satisfied by Problem (MM). In addition, due to the quadratic criterion feature of the objective function, the conjugate gradient method [17, 18], which is one of the numerical optimization techniques, could be applied to solve Problem (MM).

3.3 Conjugate gradient algorithm

For simplicity [19], establish Problem (MM) as a nonlinear optimization problem with the initial control given by u 0 = u k 0 as follows:

min u k J 3 u subject to u = u k m for k = 0 , 1 , , N 1 E18

Let this problem as Problem (Q). Moreover, the Hamiltonian function defined in (4) is taken into consideration as an equivalent objective function. Hence, this Hamiltonian function allows the evaluation of the gradient function, which is the stationary condition in (6), and by using the iterative solution u i = u k i to satisfy the state Eq. (8), which is solved forward in time, and the co-state Eq. (7), which is solved backward in time.

Define the gradient function g : m m as

g u i = u J 3 u i E19

which is represented by the stationary condition in (6). For arbitrary initial control u 0 m , the initial gradient and the initial search direction are calculated from

g 0 = g u 0 E20
d 0 = g 0 . E21

The following line search equation is applied to update the control sequence:

u i + 1 = u i + a i d i E22

where a i is the step size, and its value can be determined from

a i = arg min a 0 J 3 u i + a d i . E23

After that, the gradient and the search direction are updated by

g i + 1 = g u i + 1 E24
d i + 1 = g i + 1 + b i d i E25

with

b i = g i + 1 T g i + 1 g i T g i E26

for i = 0, 1, 2, … represents the iteration numbers.

From the discussion above, we present the result as a proposition given as follows:

Proposition 1. Consider Problem (Q). The control sequence u i , which is defined in (22) and is represented by

u i = u 0 T u 1 T u N 1 T ,

is generated through a set of the search direction vector d i whose components are linearly independent. Also, the direction d i is conjugacy.

The conjugate gradient algorithm is summarized below:

Conjugate gradient algorithm

Data: Choose the arbitrary initial control u 0 and the tolerance ε .

Step 0: Compute the initial gradient g 0 from (20) and the initial search direction d 0 from (21), respectively. Set i = 0 .

Step 1: Solve the state Eq. (8) forward in time from k = 0 to k = N with the initial condition (9) to obtain x k i , k = 0 , 1 , , N .

Step 2: Solve the costate Eq. (7) backward in time from k = N to k = 0 with the boundary condition (9), where p k i is the solution obtained.

Step 3: Calculate the value of the cost functional J 3 u i from (17).

Step 4: Solve (23) to obtain the step size a i .

Step 5: Calculate the control u i + 1 from (22).

Step 6: Evaluate the gradient g i + 1 and the search direction d i + 1 , respectively, from (24) and (25) with computing b i from (26). If the gradient g i + 1 = g i , within a given tolerance, stop, else set i = i + 1 , go to Step 1.

Remark 1:

  1. Step 0 is the preliminary step for setting the initial search direction based on the gradient direction in using the conjugate gradient algorithm.

  2. Steps 1, 2, and 3 are performed to solve the system optimization by using the corresponding control sequence u i .

  3. Steps 4, 5, and 6 are the computation steps in implementing the conjugate direction.

3.4 Iterative calculation procedure

Accordingly, Problem (Q) is solved by using the conjugate gradient algorithm. Indeed, the solution procedure for system optimization with parameter estimation could be described by joining the conjugate gradient algorithm with the parameters estimated. A summary of the calculation procedure including the principle of model-reality differences is listed as follows:

Iterative algorithm based on model-reality differences

Data: A , B , Q , R , S N , N , x 0 , r 1 , r 2 , k v , k z , k p , f , L . Note that A and B could be determined based on the linearization of f at x 0 or from the linear terms of f .

Step 0: Compute a nominal solution. Assume that α k = 0 , k = 0 , 1 , , N 1 , and r 1 = r 2 = 0 . Solve Problem (M) defined by (2) to obtain u k 0 , k = 0 , 1 , , N 1 , and x k 0 , p k 0 , k = 0 , 1 , , N . Then, with α k = 0 , k = 0 , 1 , , N 1 , and using r 1 , r 2 from the data. Set i = 0 , v k 0 = u k 0 , z k 0 = x k 0 and p ̂ k 0 = p k 0 .

Step 1: Compute the parameters γ k i , k = 0 , 1 , , N , and α k i , k = 0 , 1 , , N 1 , from (10)(12). This is called the parameter estimation step.

Step 2: Compute the modifiers Γ i , λ k i and β k i , k = 0 , 1 , , N 1 , from (13)(15). Notice that this step requires taking the derivatives of f and L with respect to v k i and z k i .

Step 3: With γ k i , α k i , Γ i , λ k i , β k i , v k i , and z k i , solve Problem (Q) using the conjugate gradient algorithm. This is called the system optimization step.

Step 4: Test the convergence and update the optimal solution of Problem (P). In order to provide a mechanism for regulating convergence, a simple relaxation method is employed:

v k i + 1 = v k i + k v u k i v k i E27
z k i + 1 = z k i + k z x k i z k i E28
p ̂ k i + 1 = p ̂ k i + k p p k i p ̂ k i E29

where k v , k z , k p 0 1 are scalar gains. If v k i + 1 = v k i , k = 0 , 1 , , N 1 , and z k i + 1 = z k i , k = 0 , 1 , , N , within a given tolerance, stop; else set i = i + 1 , and repeat the procedure starting from Step 1.

Remark 2:

  1. In Step 0, the nominal solution could be obtained by using the standard procedure of the linear quadratic regulator approach, where the feedback gain and the Riccati equation are calculated offline.

  2. In Step 3, applying the conjugate gradient algorithm to obtain the new control sequence will give a good effect if the conjugacy of the search direction is satisfied.

  3. In Step 4, the simple relaxation method in (27)(29) is used, so that the matching scheme for the parameters and the optimal solution can be established.

Advertisement

4. Illustrative examples

In this section, two examples are studied. The first example is for optimizing and controlling a damped harmonic oscillator [7], and the second example is related to optimal control of a continuous stirred-tank chemical reactor [8]. The mathematical models of these examples are discussed, and their optimal solution is obtained by using the algorithm discussed in Section 3. Here, the algorithm is implemented in the Octave 5.1.0 environment.

4.1 Example 1: a damped harmonic oscillator

Consider a damped harmonic oscillator [7] given by

x ̇ = 0 1 ω 2 2 δω x + 0 1 u E30

with the natural frequency ω = 0.8, the damping ratio δ = 0.1, and the initial state x 0 = 10 10 T . Define the state x = x 1 x 2 T , where x 1 is the displacement and x 2 is the velocity. For the purpose of controlling this oscillator, the following objective function

J 0 = 1 2 0.0 9.4 x 1 t 2 + x 2 t 2 + u t 2 dt E31

is minimized. This problem is a continuous-time linear optimal control problem, and the equivalence discrete time optimal control problem, which is regarded as Problem (P), is given by:

min u J u = k = 0 10 1 2 Δ t x 1 k 2 + x 2 k 2 + u k 2 subject to x k + 1 = 1.00 0.94 0.60 0.85 x k + 0.00 0.94 u k E32

with the initial state x 0 = 10 10 T . and the sampling time Δ t = 0.94 s is taken for the discretization transform.

Consider the model-based optimal control problem, which is regarded as Problem (M), given by:

min u J u = k = 0 10 1 2 x 1 k 2 + x 2 k 2 + u k 2 + γ k Δ t subject to x k + 1 = 1 0 0 1 x k + 1 0 u k + α k E33

with the initial state x 0 = 10 10 T , and the adjusted parameters γ k , k = 0 , 1 , , N , and α k , k = 0 , 1 , , N 1 , are supplied to the model used.

By using the algorithm proposed, the simulation result is shown in Table 1. Notice that the minimum cost for Problem (M) is 546.05 units without adding the adjusted parameters. Once the adjusted parameters are taken into consideration, the iterative solution approximates to the true optimal solution of the original optimal control problem, in spite of model-reality differences. It is highlighted that there is a 99% of the cost reduction to obtain the final cost of 128.50 units.

Number of iteration Initial cost Final cost Elapsed time (s)
20 17053.11 128.50 1.38021

Table 1.

Simulation result, Example 1.

Figures 1 and 2 show the trajectories of control and state, respectively. With this control effort, the state reaches at the steady state after 4 units of time, which presents the oscillator stopped from moving. Figure 3 shows the changes of the costate at the first 2 units of time. The optimal solution obtained is verified by satisfying the stationary condition as shown in Figure 4. Figures 5 and 6 show the adjusted parameters after the convergence is achieved, where the model-reality differences are measured during the iterative procedure.

Figure 1.

Final control u(k), Example 1.

Figure 2.

Final state x(k), Example 1.

Figure 3.

Final costate p(k), Example 1.

Figure 4.

Stationary Hu(k), Example 1.

Figure 5.

Adjusted parameter α(k), Example 1.

Figure 6.

Adjusted parameter γ(k), Example 1.

Therefore, this damped harmonic oscillator is controlled, and the cost function is minimized as desired.

4.2 Example 2: a continuous stirred-tank chemical reactor

Consider a continuous stirred-tank chemical reactor, which consists of two state equations [8]. The flow of a coolant through a coil inserted in the reactor is to control the first order, irreversible exothermic reaction taking place in the reactor. Assume that x 1 t is the deviation from the steady-state temperature, x 2 t is the deviation from the steady-state concentration, and u t is the normalized control variable that represents the effect of coolant flow on the chemical reaction. The corresponding state equations are given by

x ̇ 1 t = 2 x 1 t + 0.25 + x 2 t + 0.5 exp 25 x 1 t x 1 t + 2 x 1 t + 0.25 u t E34
x ̇ 2 t = 0.5 x 2 t x 2 t + 0.5 exp 25 x 1 t x 1 t + 2 E35

with the initial state x 0 = 0.05 0.00 T . The cost function to be minimized is given by

J 0 = 0.0 0.8 x 1 t 2 + x 2 t 2 + 0.1 u t 2 dt . E36

Here, the desired objective is to maintain the temperature and the concentration close to their respective steady-state values without expending large amounts of the control effort.

This problem is a continuous time nonlinear optimal control problem. For doing the discretization transform, the sampling time Δ t = 0.0057 s is used to formulate the equivalence discrete-time optimal control problem, which is referred to as Problem (P), given by:

min u J u = k = 0 80 1 2 2 x 1 k 2 + 2 x 2 k 2 + 0.2 u k 2 Δ t subject to x 1 k + 1 = x 1 k 2 x 1 k + 0.25 Δ t + x 2 k + 0.5 Δ t exp 25 x 1 k x 1 k + 2 x 1 k + 0.25 u k Δ t ) x 2 k + 1 = x 2 k + 0.5 x 2 k Δ t x 2 k + 0.5 Δ t exp 25 x 1 k x 1 k + 2 E37

with the initial state x 0 = 0.05 0.00 T .

By applying the algorithm proposed to obtain the optimal solution for Problem (P), the following model, which is referred to as Problem (M), is introduced,

min u J u = k = 0 80 1 2 2 x 1 k 2 + 2 x 2 k 2 + 0.2 u k 2 Δ t subject to x k + 1 = 1.048 0.010 0.062 0.984 x k + 0.002 0.000 u k + α k E38

with the initial state x 0 = 0.05 0.00 T , and the adjusted parameters γ k , k = 0 , 1 , , N , and α k , k = 0 , 1 , , N 1 , are added into the model.

Table 2 shows the simulation result obtained by using the algorithm proposed. It is mentioned that the minimum cost for the linear model-based optimal control problem is 5.9589 units. At the beginning of the iteration calculation procedure, the initial cost is 0.147463 unit, and a 90% of cost reduction is addressed to give the final cost of 0.014167 unit.

Number of iteration Initial cost Final cost Elapsed time (s)
9 0.147463 0.014167 4.60934

Table 2.

Simulation result, Example 2.

The trajectories of the final control and the final state are, respectively, shown in Figures 7 and 8. It is noted that the state reaches to the steady state after 40 units of time by associating the control effort taken. This situation indicates that the temperature and the concentration are maintained at their steady state. Thus, the desired objective is confirmed. Figure 9 shows the costate behavior, which is reduced gradually to zero at the terminal time, and Figure 10 shows the stationary condition, which examines the existing of the optimal solution. The adjusted parameters, which are shown in Figures 11 and 12, respectively, measure the differences between the model used and the real plant.

Figure 7.

Final control u(k), Example 2.

Figure 8.

Final state x(k), Example 2.

Figure 9.

Final costate p(k), Example 2.

Figure 10.

Stationary Hu(k), Example 2.

Figure 11.

Adjusted parameter α(k), Example 2.

Figure 12.

Adjusted parameter γ(k), Example 2.

Hence, the correct optimal solution of Problem (P) is approximated successfully by solving the model in Problem (M), and the efficiency of the algorithm proposed is demonstrated.

Advertisement

5. Concluding remarks

The approach, which integrates system optimization and parameter estimation, was discussed in this chapter. The use of the conjugate gradient method in solving the model-based optimal control problem has been examined, and the applicability of the conjugate gradient approach in associating the principle of model-reality differences was identified. Definitely, many computational approaches could be used to solve the model-based optimal control; however, the algorithm proposed in this chapter gives a tractable solution procedure for handling the optimal control problems with different structures and parameters, especially for obtaining the optimal solution for the nonlinear optimal control problem. In conclusion, the efficiency of the algorithm is highly recommended. In future research, it is strongly suggested to investigate the application of optimization techniques in stochastic optimization and control.

Advertisement

Acknowledgments

The authors would like to acknowledge the Universiti Tun Hussein Onn Malaysia (UTHM) and the Ministry of Higher Education (MOHE) for the financial support for this study under the research grant FRGS VOT. 1561.

Advertisement

Conflict of interest

The authors declare no conflict of interest.

References

  1. 1. Ahmed NU. Linear and Nonlinear Filtering for Scientists and Engineers. Singapore: World Scientific Publishers; 1999
  2. 2. Bagchi A. Optimal Control of Stochastic Systems. New York: Prentice-Hall; 1993
  3. 3. Bryson AE, Ho YC. Applied Optimal Control. New York: Hemisphere Publishing Company; 1975
  4. 4. Bryson AE. Applied Linear Optimal Control, Examples and Algorithms. UK: Cambridge University Press; 2002
  5. 5. Kalman RE. Contributions to the theory of optimal control. Boletín de la Sociedad Matemática Mexicana. 1960;5:102-119
  6. 6. Simon D. Optimal State Estimation: Kalman, H-Infinity and Nonlinear Approaches. Hoboken, New Jersey: John Wiley & Sons; 2006
  7. 7. Lewis FL, Vrabie V, Symos VL. Optimal Control. 3rd ed. New York: John Wiley & Sons, Inc.; 2012
  8. 8. Kirk DE. Optimal Control Theory: An Introduction. New York: Dover Publications; 2004
  9. 9. Roberts PD. An algorithm for steady-state system optimization and parameter estimation. International Journal of Systems Science. 1979;10:719-734
  10. 10. Roberts PD, Williams TWC. On an algorithm for combined system optimization and parameter estimation. Automatica. 1981;17:199-209
  11. 11. Roberts PD, Becerra VM. Optimal control of a class of discrete-continuous nonlinear systems decomposition and hierarchical structure. Automatica. 2001;37:1757-1769
  12. 12. Becerra VM. Development and applications of novel optimal control algorithms [Ph.D. thesis]. UK: City University; 1994
  13. 13. Becerra VM, Roberts PD. Dynamic integrated system optimization and parameter estimation for discrete time optimal control of nonlinear systems. International Journal of Control. 1996;63:257-281
  14. 14. Becerra VM, Roberts PD. Application of a novel optimal control algorithm to a benchmark fed-batch fermentation process. Transaction of Instrument Measurement Control. 1998;20:11-18
  15. 15. Kek SL, Teo KL, Mohd Ismail AA. An integrated optimal control algorithm for discrete-time nonlinear stochastic system. International Journal of Control. 2010;83:2536-2545
  16. 16. Kek SL, Teo KL, Mohd Ismail AA. Filtering solution of nonlinear stochastic optimal control problem in discrete-time with model-reality differences. Numerical Algebra, Control and Optimization. 2012;2:207-222
  17. 17. Chong EKP, Zak SH. An Introduction to Optimization. 4th ed. Hoboken, New Jersey: John Wiley & Sons, Inc.; 2013
  18. 18. Mostafa EME. A nonlinear conjugate gradient method for a special class of matrix optimization problems. Journal of Industrial and Management Optimization. 2014;10:883-903
  19. 19. Lasdon LS, Mitter SK. The conjugate gradient method for optimal control problems. IEEE Transactions on Automatic Control. 1967;12:132-138

Written By

Sie Long Kek, Sy Yi Sim, Wah June Leong and Kok Lay Teo

Submitted: 21 August 2019 Reviewed: 13 September 2019 Published: 22 November 2019