 Research
 Open Access
 Published:
Realtime obstacle avoidance of mobile robots using statedependent Riccati equation approach
EURASIP Journal on Image and Video Processing volume 2018, Article number: 79 (2018)
Abstract
In this paper, statedependent Riccati equation (SDRE) methodbased optimal control technique is applied to a robot. In recent years, issues associated with the robotics have become one of the developing fields of research. Accordingly, intelligent robots have been embraced greatly; however, control and navigation of those robots are not easy tasks as collision avoidance of stationary obstacles to doing a safe routing has to be taken care of. A moving robot in a certain time has to reach the specified goals. The robot in each time step needs to identify criteria such as velocity, safety, environment, and distance in respect to defined goals and then calculate the proper control strategy. Moreover, getting information associated with the environment to avoid obstacles, do the optimal routing, and identify the environment is necessary. The robot must intelligently perceive and act using adequate algorithms to manage required control and navigation issues. In this paper, smart navigation of a mobile robot in an environment with certain stationary obstacles (known to the robot) and optimal routing through Riccati equation depending on SDRE is considered. This approach enables the robot to do the optimal path planning in static environments. In the end, the answer SDRE controller with the answer linear quadratic controller will be compared. The results show that the proposed SDRE strategy leads to an efficient control law by which the robot avoids obstacles and moves from an arbitrary initial point × 0 to a target point. The robust performance of SDRE method for a robot to avoid obstacles and reach the target is demonstrated via simulations and experiments. Simulations are done using MATLAB software.
Introduction
A mobile intelligent robot is a useful tool which can lead to the target and at the same time avoid an obstacle when faced with it. Obstacle avoidance means that the robot avoids colliding with obstacles such as fixed objects or moving objects. So when a robot encounters with an obstacle, it must decide to avoid it and at the same time, consider the most efficient path to the target with a good decision; this decision is performed in this research using the statedependent Riccati equation (SDRE) control method. The optimal control of nonlinear systems cannot be done similar to the methods of linear systems. One of the significant viewpoints for optimal control of nonlinear systems is using the SDRE. This method in addition to stability creates a proper functioning and robust for a wide range of nonlinear systems. Nonlinear optimal controller SDRE is a developed linear optimal control LQR, in which equations Riccati is statedependent. Strategy SDRE has been introduced in the past decade. This strategy is a very effective algorithm for feedback nonlinear analysis that its states are nonlinear; however, a flexibility idea of through the weight matrix is dependent on the state. This involves factoring (ie, parameterization) of nonlinear dynamics to state vector and produces a value function for the matrix which is dependent on its state.
There are conventional methods of obstacle avoidance such as the path planning method [1], the navigation function method [2], and the optimal regulator [3]. Hence, an SDRE regulator [4, 5] is used for this paper. In recent years, many researchers have investigated the obstacle avoidance problem from different perspectives. SDRE technique developed as a design method, which provides a systematic and effective design of nonlinear controllers, filters, and observers. Because of its versatile features, SDRE is broadly used for different cases [6,7,8,9,10,11,12,13,14]. SDRE can also be used in the field of medicine, for example, in [15, 16] the presentation of the optimal chemotherapy protocol for cancer treatment, considering the metastasis, has used the full optimal SDRE feedback. In [17], the SDRE algorithm is studied on motion design of the cablesuspended robot with uncertainties and moving obstacles. A method for controlling the tracing of a robot has been developed by the SDRE. Vibration control of flexiblelink manipulator in [18] is used in SDRE controller and Kalman filtering. The problem of estimating flexural states during the use of the SDRE for flexible control is the focus of this paper. Wanga et al. [19] present a novel H_{2} − H_{∞}SDRE control approach with the purpose of providing a more effective control design framework for continuoustime nonlinear systems to achieve a mixed nonlinear quadratic regulator and H_{∞}control performance criteria.
It should be noted that the algorithms discussed for avoiding an obstacle are different from each other, and each algorithm has proposed a separate and new method for avoiding obstacles. So far, novel methods to avoid obstacles in addition to the mentioned methods are also presented [20,21,22,23].
This paper is organized as follows: a description of the SDRE formulation is given in Section 2. Then, motion equations of the robot and its constraints are presented in Section 3. In Section 4, SDRE formulation of the robot is derived. And simulation results have been reported in Section 5. Finally, in Section 6, some conclusions are drawn from this research.
Methods
Problem formulation
Consider a nonlinear system that is autonomous, fullstate observable and input affine that is presented as follows:
Where x ∈ R^{n} is the state vector, u ∈ R^{m} is the input vector, and t ∈ [0, ∞), with functions f : R^{n} → R^{n}and B : R^{n} → R^{n × m}, and B(x) ≠ 0 ∀ x. Without any loss of generality, the origin is assumed to be an equilibrium point, such that f(0) = 0. In this context, the minimization of the infinitetime performance criterion is as follows:
Consider the above equation which is nonquadratic in x but quadratic in u. The state and input weighting matrices are assumed statedependent such that Q : R^{n} → R^{n × n} and R : R^{n} → R^{n × m}. These design parameters satisfy Q(x) ≥ 0 and R(x) > 0 for all x [24].
SDRE nonlinear regulator problem
The SDRE methodology uses extended linearization as the key design concept in formulating the nonlinear optimal control problem. The linear control synthesis method for this case is the LQR synthesis method. SDRE feedback control is an “extended linearization control method” that provides a similar approach to the nonlinear regulation problem for the inputaffine system (1) with cost functional (2). The nonlinear system of (1) can be rewritten in the linear structure using SDC form:
SDC form is not unique and should be chosen in such a way that \( \left\{{Q}^{\raisebox{1ex}{$1$}\!\left/ \!\raisebox{1ex}{$2$}\right.}(x),A(x)\right\} \) and {A(x), B(x)} are pointwise observable and pointwise stabilizable, respectively.
Where W(x) is obtained from the solution of algebraic SDRE:
to get W(x) ≥ 0.
Where W(x) is the unique, symmetric, positivedefinite solution of the algebraic statedependent Riccati equation. As a result of this, the nonlinear control is obtained as:
The response obtained is locally asymptotically stable and is optimal [25].
SDRE nonlinear tracking problem
Consider an infinitetime nonlinear tracking problem. Minimize the following nonlinear cost function that is nonquadratic in x but is quadratic in u [26]:
in which
Q, R, x, and uare like the previous model and with the same conditions. By solving the Eq. (6), the input is obtained as the following equation:
The fourwheeled car
Here is considered a fourwheeled robot with frontwheel drive as Fig. 1. In this robot model, R_{tu} instantaneous turning radius, L length between the rear axle and the front of the robot, and (x, y) the robot’s position is measured at the center point of the rear wheels of the robot. The state vector consists of twoposition variables (x, y), one orientation variable (θ), the speed of the robot (V), and frontwheel drive angle (α). Only the front wheels swivel in two directions, and the rear wheels swivel without slipping. The main reason for choosing this system for this project is the nonholonomic boundary movement of this system [27].
Here, (V) and (θ) are considered as input vectors, but using these control inputs, this system is faced with solutions that are not smooth. It is very obvious that in practice mode, the speed of the vehicle’s robot and the wheel angle cannot change quickly and instantaneously. Therefore, to smooth the (V, θ) controls and make it possible to apply inputs to a real robot, a control level is added. In other words, (V, θ) is recorded as part of the state vector and (β, ψ) is added as a new control vector. These changes create smoother paths for (V, θ) that can be used in a real robot. Finally, with regard to all these points, the equation of motion for the robot kinematics derived as follows [27].
Vehicle’s constraints
The constraints of the state and control inputs are as follows:
So far, dynamic equations of the robot and corresponding constraints on states and control input have been presented. In the following section, the proper formulation for utilizing SDRE approach for robot motion control will be presented.
Implementation of SDRE controller for the robot
SDC parameterization
The first step in designing the SDRE controller is to formulate the SDC. The most important point in SDC parameterization is that factorization should provide appropriate control over the entire area. SDC presentation of the system is written as follows [28].
Obstacle avoidance
In this paper, the concept of the APF method and the navigation function [29] is used to avoid collisions with obstacles during the robot’s movement. The navigation function that is similar to the APF, with the help of the sensor, will detect obstacles and avoid them and converge the robot toward the target. This function has the local minimum at the destination and a local maximum at the obstacles.
In the modified model, OBS can be decomposed into obs_{a} and obs_{b}. Where obs_{a} is the component force under the direction along the line between the robot and the obstacle, and obs_{b} is the component force under the direction along the line between the robot and the target. m is a weighting factor that one can set the distance of the robot from an obstacle during movement. By setting mand z, the desired result is achieved which is obstacle avoidance and reach the target. The proposed navigation function added to the cost function is as follows:
The form of obstacles is a sphere with center (x_{ob}, y_{ob}) and radius R_{ob}, and it is supposed that they do not overlap each other. The purpose of this study is to design a control law by which the controlled object avoids obstacles and moves from an arbitrary initial point x_{0} to the destination point x_{d}. In the next step, this obstacle avoidance equation is added to the cost function. The purpose here is to minimize the performance of the criterion function that is considered as follows:
Then, using the factoring, the matrix Q(x) is obtained as follows:
E_{X} and E_{Y} are errors of position. Q(x) is not a single matrix and many matrices are obtained for it. By using this formulation, the robot can now start from an initial point and reach a destination while avoiding obstacles. The useful property of the obstacle avoidance term in the cost function led to the increase in the cost function while the robot is nearing obstacle and decreasing when leaving it. Figure 2shows mesh plot of obstacle avoidance term in which the function has the global minimum at the destination and a local maximum at the obstacle. By rewriting (15), the second obstacle in the motion planning algorithm can be constructed as follows.
Results and discussion
In this section, the effectiveness of the proposed method is verified by the simulation model of the robot. The general parameters include all states are as follows:
Figure 2 shows the mesh plot of the obstacle avoidance term in which the function has the global minimum at the destination and a local maximum at the obstacle. It shows that the obstacles like the summit and the target act like the cavity and absorbs the robot into the inside itself.
In Fig. 3, obstacles have been considered as a circle of radius 0.4 and show that with SDRE method, the robot avoids obstacles and reaches the target with minimal cost. When the robot is faced by an obstacle; Q parameter is designed so that increases the cost function and in other words, finds the highest possible value to avoid colliding with the obstacle. Here, to avoid collision with the obstacle and reach the desired target, the value of m and z in the navigation obsis respectively put 5 and 2.
Figure 4 shows two variables x and y in reaching the target point x_{d} = (1, 5), and these figures show that the robot stops as soon as reaching to the target. Figure 5 shows that other states also regulated to zero after 3.1 s. Figure 6 shows control inputs that are applied to reach the desired target, after reaching the target became zero. The purpose of this control method is to find the control inputs that are applied to the system with Eq. (11); while stabilizing the system and satisfying the constraints defined for it, the defined cost function (16) is minimized, and the system state variables converge to zero with the least control effort. The figures obtained from the simulation show that SDRE method has achieved the desired goals. By comparing this method with the modified APF method, it is observed that this method selects a more optimal route according to the definition of the cost function for avoiding the obstacle and reaching the goal. The other difference is that in SDRE method, three parameters are added: vehicle velocity (V), the angle of the front wheels (α), and length of the car between the front and rear axles (L), to design appropriate control function in order to avoidance obstacles and reach the desired target.
Without taking navigation function and supposing Q(x) = diag[100, 100, 1, 1, 1], Fig. 7 shows that the robot cannot reach the target with obstacle avoidance. Figure 8 of the LQR method in which a nonlinear system is linearized around the vector X_{v} = [1, 1, 1, 1, 1]^{T}and initial conditionsX_{p} = [−3,3,0.0001,0.0001,0.0001]^{T}. Figure 9 shows that although the control inputs became zero, as Fig. 8 depicts, the robot cannot avoid obstacles in reaching to the target. Like Fig. 8, in LQR method, with linearization of around none of the operating points and initial conditions, the robot cannot avoid obstacles in reaching to the target, so is concluded, by SDRE method where the matrix Adepends on its states, the answer is better.
In the second section, the effectiveness of the proposed method is verified by the simulation model of the robot. The general parameters include all states are as follows:
Figure 10 shows the mesh plot of the obstacle’s avoidance term in which the function has the global minimum at the destination and a local maximum at the obstacle. It shows that the obstacles like the summit and the target act like the cavity and absorbs the robot into the inside itself. Here, to avoid collision with the obstacle and reach the desired target, the value of m and z in the navigation obs is respectively put 0.5 and 2. In the face of obstacles more than an obstacle. The principles of work for other obstacles are like one obstacle, but in this case, a navigation function is put. For example, in Fig. 11, two obstacles have been considered as a circle of radius 0.25 and shows that with SDRE method; the robot avoids obstacles and reaches the target with minimal cost. Figure 12 shows the curve of two variable’s xand y in reaching the target point x_{d} = (2, 5.5), and these figures show that the robot stops as soon as reaching to the target. Figure 13 shows that other states also regulated to zero after 4.2 s. Figure 14 shows control inputs that are applied to reach the desired target, after reaching the target the control attempt becomes zero.
Without taking navigation function and supposing Q(x) = diag[100, 100, 1, 1, 1], Fig 15 shows that the robot cannot reach the target with obstacle avoidance. Figure 16 of the LQR method in which a nonlinear system is linearized around the vector X_{v} = [1, 1, 1, 1, 1]^{T}and initial conditions X_{p} = [−3,3,0.0001,0.0001,0.0001]^{T}. Figure 17 shows that although the control inputs became zero, as Fig. 16 depicts, the robot cannot avoid obstacles in reaching to the target. Like Fig. 16, in LQR method, with linearization of around none of the operating points and initial conditions, the robot cannot avoid obstacles in reaching to the target, so is concluded, by SDRE method where the matrix Adepends on its states, the answer is better.
Conclusions
This paper focuses on the SDRE nonlinear regulator for solving the nonlinear optimal control problems. The existence of solutions as well as optimality and stability properties associated with SDRE controllers are the main of this paper. The paper is organized as follows. In Section 2, the formulation of the nonlinear optimal control problem, the concept of extended linearization and the SDRE controller for nonlinear optimal regulation are presented, then the additional degrees of freedom provided by the nonuniqueness of the SDC parameterization is reviewed. The necessary and sufficient conditions on the existence of solutions to the nonlinear optimal control problem, in particular, by SDRE feedback control, are reviewed. A theoretical study on the stability and optimality properties of SDRE feedback controls is pursued. SDRE method is compared with LQR method. It is concluded for that SDRE is a semilinear method, give the best answer than linear methods for nonlinear systems, because the basis of this method is that takes nonlinearity of the system and creates the nonlinear system that its statedependent coefficient matrix structure has semilinear and minimum nonlinear performance index with semiquadratic structure. The result is that contrary to controllers like LQR that first, they linearize nonlinear controllers, then the laws are designed for stability that can cause to remove some of the important elements of nonlinear systems that have a key role in system stability. In the method of SDRE, the important elements are not removed from the nonlinear system and as a result, according to the results of the various articles, it is concluded that the system is robust against disturbances and uncertainties and achieves better performance than LQR. Our target in this research is finding an equation control using the method SDRE for routing a robot and avoid collisions with obstacles and path optimization for the robot to reach the target. In the end, the proper suggestion for each robot’s path planning is to create an optimal combination algorithm according to the specific structure of each robot so that each algorithm can be covered the constraints of the other algorithm.
Abbreviations
 LQR:

Linear quadratic regulator
 SDC:

Statedependent coefficient
 SDRE:

Statedependent Riccati equation
 R:

Weighting matrix inputs
 Q:

Weighting matrix of states
 W:

Matrix final cost
 J:

Cost function
 u:

Control signal
References
 1.
G. Han, C. Zhang, J. Jiang, X. Yang, M. Guizani, Mobile anchor nodes path planning algorithms usingnetworkdensitybased clustering in wireless sensor networks. J. Netw. Comput. Appl. 85, 64–75 (2017)
 2.
S. Hacohen, S. Shoval, N. Shvalbc, Applying probability navigation function in dynamic uncertain environments. Robot. Auton. Syst. 87, 237–246 (2017)
 3.
S. Herzog, F. Wörgötter, T. Kulvicius, Generation of movements with boundary conditions based on optimal control theory. Robot. Auton. Syst. 94, 1–11 (2017)
 4.
R. Iraji, S.M. Ghadami, Aids control using statedependent Riccati equation. Sci. Int. 27, 1183–1188 (2015)
 5.
Q. Jia, Y. Liu, G. Chen, H. Sun, Statedependent Riccati equation control for motion planning of space manipulator carrying a heavy payload. Open Mech. Eng. J. 9(1), 992–999 (2015)
 6.
K. Sinan, Tracking control of elastic joint parallel robots via statedependent Riccati equation. Turk. J. Elec. Eng. Comput. Sci. 23, 522–538 (2015)
 7.
F. Sağlam, Y.S. Ünlüsoy, State dependent Riccati equation control of an active hydropneumatic suspension system. J. Autom. Control Res. 1, 1–10 (2014)
 8.
A. Khamis, D.S. Naidu, Finitehorizon optimal nonlinear regulation and tracking using differential state dependent Riccati equation. Int. J. Control. 1–18 (2014) https://doi.org/10.1080/00207179.2014.977823
 9.
S.R. Nekoo, Nonlinear closed loop optimal control: a modified statedependent Riccati equation. ISA Trans. 52, 285–290 (2013)
 10.
M.H. Korayem, S.R. Nekoo, Statedependent differential Riccati equation to track control of timevarying systems with state and control nonlinearities. ISA Trans. 57, 117–135 (2015)
 11.
Y.L. Kuo, T.L. Wu, A suboptimal tracking control approach based on the statedependent Riccati equation method. J. Comput. Theor. Nanosci. 13, 1013–1021 (2016)
 12.
J. Slávka, S. Ján, in IEEE 11th International Symposium on Intelligent Systems and Informatics. Application of the statedependent Riccati equation method in nonlinear control design for inverted pendulum systems (2013), pp. 26–28
 13.
F. Liccardo, S. Strano, M. Terzo, in World Congress on Engineering. Optimal control using statedependent Riccati equation (SDRE) for a hydraulic actuator (2013), pp. 3–5
 14.
S.M.H. Rostami, M. Ghazaani, State dependent Riccati equation tracking control for a two link robot. J. Comput. Theor. Nanosci. 15, 1490–1494 (2018)
 15.
A. Hamdache, S. Saadi, A stochastic nominal control optimizing the adoptive immunotherapy for cancer using tumorinfiltrating lymphocytes. Int. J. Dynam. Control. 5, 783–798 (2017)
 16.
R. Padmanabhan, N. Meskin, W.M. Haddad, Reinforcement learningbased control of drug dosing for cancer chemotherapy treatment. Math. Biosci. 293, 11–20 (2017)
 17.
A. Behboodi, S.M. Salehi, SDRE controller for motion design of cablesuspended robot with uncertainties and moving obstacles. Mater. Sci. Eng. 248, 1–6 (2017)
 18.
M.P.P. Reddy, J. Jacob, Vibration control of flexible link manipulator using SDRE controller and Kalman filtering. Stud. Inform. Control 2, 143–150 (2017)
 19.
X. Wanga, E.E. Yazb, S.C. Schneiderb, Y.I. Yazc, H2−H∞ control of continuoustime nonlinear systems using the statedependent Riccati equation approach. Syst. Sci. Control Eng. 5(1), 224–231 (2017)
 20.
Y. Chen, H. Peng, J. Grizzle, Obstacle avoidance for lowspeed autonomous vehicles with barrier function. IEEE Trans. Control Syst. Technol. 99, 1–13 (2018)
 21.
M.G. Plessen, D. Bernardini, H. Esen, A. Bemporad, Spatialbased predictive control and geometric corridor planning for adaptive cruise control coupled with obstacle avoidance. IEEE Trans. Control Syst. Technol. PP(99), 1–13 (2018)
 22.
D. Han, H. Nie, J. Chen, M. Chen, Dynamic obstacle avoidance for manipulators using distance calculation and discrete detection. Robot. Comput. Integr. Manuf. 49, 98–104 (2018)
 23.
M. Mendesy, A.P. Coimbray, M.M. Crisostomoy, Assis  cicerone robot with visual obstacle avoidance using a stack of odometric data. IAENG Int. J. Comput. Sci. 45, 219–227 (2018)
 24.
P.D.H. Nguyen, C.T. Recchiuto, A. Sgorbissa, Realtime path generation and obstacle avoidance for multirotors: a novel approach. J. Intell. Robot. Syst. 89, 27–49 (2018)
 25.
Ç. Tayfun, in The International Federation of Automatic Control. Statedependent Riccati equation (SDRE) control: a survey (2008), pp. 6–11
 26.
P.K. Menon, T. Lam, L.S. Crawford, V.H.L. Cheng, in American Control Conference. Realtime computational methods for SDRE nonlinear control of missiles (2002), pp. 1–17
 27.
M.H. Hurni, P. Sekhavat, I.M. Ross, in AIAA Infotech Aerospace Conference AIAA. Issues on robot optimal motion planning and obstacle avoidance (2009), pp. 1–20
 28.
S.S. Mohammadi, H. Khaloozade, in 4th international conference on control, Instrumentation, and Automation. Optimal motion planning of unmanned ground vehicle using SDRE controller in the presence of obstacles (2016), pp. 167–171
 29.
S. Xie, P. Wu, Y. Peng, J. Luo, J. Gu, D. Qu, Q. Li, in Proceeding of the IEEEInternational Conference on Information and AutomationHailar. The obstacle avoidance planning of USV based on improved artificial potential field (2014), pp. 746–751
Acknowledgements
Authors of this would like to express the sincere thanks to the National Natural Science Foundation of China for their funding support to carry out this project.
Funding
This work is supported by the National Natural Science Foundation of China (61772454, 6171171570).
Availability of data and materials
The data is available on request with prior concern to the first author of this paper.
Author information
Affiliations
Contributions
All the authors conceived the idea, developed the method, and conducted the experiment. SMHR contributed to the formulation of methodology and experiments. AKS contributed in the data analysis and performance analysis. JW contributed to the overview of the proposed approach and decision analysis. HjK contributed to the algorithm design and data sources. All authors read and approved the final manuscript.
Corresponding author
Correspondence to Hyejin Kim.
Ethics declarations
Ethics approval and consent to participate
All the authors of this manuscript would like to declare that they mutually agreed no conflict of interest. All the results of this paper are with respect to the experiments on human subject.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Rostami, S.M.H., Sangaiah, A.K., Wang, J. et al. Realtime obstacle avoidance of mobile robots using statedependent Riccati equation approach. J Image Video Proc. 2018, 79 (2018) doi:10.1186/s1364001803191
Received
Accepted
Published
DOI
Keywords
 Optimal control
 Nonlinear control
 Obstacle avoidance
 SDRE