13
Research Article Three-Dimensional Path-Following Control of a Robotic Airship with Reinforcement Learning Chunyu Nie , 1 Zewei Zheng , 2,3 and Ming Zhu 1 1 School of Aeronautic Science and Engineering, Beihang University, Beijing 100191, China 2 School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China 3 The Seventh Research Division, Beihang University, Beijing 100191, China Correspondence should be addressed to Zewei Zheng; [email protected] Received 25 November 2018; Accepted 20 February 2019; Published 25 March 2019 Academic Editor: Zhiguang Song Copyright © 2019 Chunyu Nie et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. This paper proposed an adaptive three-dimensional (3D) path-following control design for a robotic airship based on reinforcement learning. The airship 3D path-following control is decomposed into the altitude control and the planar path-following control, and the Markov decision process (MDP) models of the control problems are established, in which the scale of the state space is reduced by parameter simplication and coordinate transformation. To ensure the control adaptability without dependence on an accurate airship dynamic model, a Q-Learning algorithm is directly adopted for learning the action policy of actuator commands, and the controller is trained online based on actual motion. A cerebellar model articulation controller (CMAC) neural network is employed for experience generalization to accelerate the training process. Simulation results demonstrate that the proposed controllers can achieve comparable performance to the well-tuned proportion integral dierential (PID) controllers and have a more intelligent decision-making ability. 1. Introduction The lift of an airship mainly comes from the buoyancy of the gas, and therefore, it does not have to do continuous move- ment to balance gravity, which makes it a promising platform for long-endurance and low-energy consumption. In the last two decades, airships such as heavy-duty airships and strato- spheric airships [1] have gained increasing interest given their potential in communication relay, space observation, and military reconnaissance. Control strategy has always been an important subject in airship research. After years of study, airship motion control has gained signicant achieve- ments with regard to hovering control, trajectory tracking, and path-following control [24]. The objective of the hover- ing control is to maintain the airship in a certain area against the wind, and the primary purpose of the trajectory tracking is to track a time-parameterized desired path, while the path-following control is mainly concerned with tracking a desired path without a specied temporal constraint. The control response of an airship is slow with a long time delay. The exibility and elasticity of the hull and the stochastic atmosphere disturbance make the control of the airship a problem with uncertainties. At present, common airship control approaches include PID control, backstep- ping, dynamic inversion, sliding mode control, and robust control [58]. The PID controller is widely used for its sim- plicity and eectiveness, but its parameters such as propor- tion, integral, and dierential should be tuned properly [9]. When the model of the system or environment changes, the performance of the PID controller may get worse before being tuned again. The modern model-based nonlinear control approaches can ensure the control robustness and global stability while being accurately controlled; however, these improvements incur the work of system modeling and parameter identication with ever increasing complexity [10, 11]. With the rapid development of information technol- ogy, machine learning algorithms have been successfully applied in various complex applications [12] and can make intelligent decisions similar to human beings. As an impor- tant unsupervised learning algorithm, reinforcement learn- ing is able to realize adaptive control and decision in an unknown environment through the mechanism of action Hindawi International Journal of Aerospace Engineering Volume 2019, Article ID 7854173, 12 pages https://doi.org/10.1155/2019/7854173

Three-Dimensional Path-Following Control of a Robotic

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Three-Dimensional Path-Following Control of a Robotic

Research ArticleThree-Dimensional Path-Following Control of a RoboticAirship with Reinforcement Learning

Chunyu Nie ,1 Zewei Zheng ,2,3 and Ming Zhu 1

1School of Aeronautic Science and Engineering, Beihang University, Beijing 100191, China2School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China3The Seventh Research Division, Beihang University, Beijing 100191, China

Correspondence should be addressed to Zewei Zheng; [email protected]

Received 25 November 2018; Accepted 20 February 2019; Published 25 March 2019

Academic Editor: Zhiguang Song

Copyright © 2019 Chunyu Nie et al. This is an open access article distributed under the Creative Commons Attribution License,which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

This paper proposed an adaptive three-dimensional (3D) path-following control design for a robotic airship based onreinforcement learning. The airship 3D path-following control is decomposed into the altitude control and the planarpath-following control, and the Markov decision process (MDP) models of the control problems are established, in which thescale of the state space is reduced by parameter simplification and coordinate transformation. To ensure the control adaptabilitywithout dependence on an accurate airship dynamic model, a Q-Learning algorithm is directly adopted for learning the actionpolicy of actuator commands, and the controller is trained online based on actual motion. A cerebellar model articulationcontroller (CMAC) neural network is employed for experience generalization to accelerate the training process. Simulationresults demonstrate that the proposed controllers can achieve comparable performance to the well-tuned proportion integraldifferential (PID) controllers and have a more intelligent decision-making ability.

1. Introduction

The lift of an airship mainly comes from the buoyancy of thegas, and therefore, it does not have to do continuous move-ment to balance gravity, which makes it a promising platformfor long-endurance and low-energy consumption. In the lasttwo decades, airships such as heavy-duty airships and strato-spheric airships [1] have gained increasing interest giventheir potential in communication relay, space observation,and military reconnaissance. Control strategy has alwaysbeen an important subject in airship research. After years ofstudy, airship motion control has gained significant achieve-ments with regard to hovering control, trajectory tracking,and path-following control [2–4]. The objective of the hover-ing control is to maintain the airship in a certain area againstthe wind, and the primary purpose of the trajectory trackingis to track a time-parameterized desired path, while thepath-following control is mainly concerned with tracking adesired path without a specified temporal constraint.

The control response of an airship is slow with a longtime delay. The flexibility and elasticity of the hull and the

stochastic atmosphere disturbance make the control of theairship a problem with uncertainties. At present, commonairship control approaches include PID control, backstep-ping, dynamic inversion, sliding mode control, and robustcontrol [5–8]. The PID controller is widely used for its sim-plicity and effectiveness, but its parameters such as propor-tion, integral, and differential should be tuned properly [9].When the model of the system or environment changes, theperformance of the PID controller may get worse beforebeing tuned again. The modern model-based nonlinearcontrol approaches can ensure the control robustness andglobal stability while being accurately controlled; however,these improvements incur the work of system modelingand parameter identification with ever increasing complexity[10, 11]. With the rapid development of information technol-ogy, machine learning algorithms have been successfullyapplied in various complex applications [12] and can makeintelligent decisions similar to human beings. As an impor-tant unsupervised learning algorithm, reinforcement learn-ing is able to realize adaptive control and decision in anunknown environment through the mechanism of “action

HindawiInternational Journal of Aerospace EngineeringVolume 2019, Article ID 7854173, 12 pageshttps://doi.org/10.1155/2019/7854173

Page 2: Three-Dimensional Path-Following Control of a Robotic

and reward” [13, 14], which does not need an accurate modelthat costs human effort.

Up to now, reinforcement learning has been extensivelystudied in robot control, traffic dispatch, communicationcontrol, and game decision-making [15–18]. In the fieldof aircraft navigation and control, reinforcement learningalso witnessed many successful applications. However, dueto the complexity of aircraft motion and the large scale ofthe state space, the reinforcement learning control can easilybe trapped in the “curse of dimensionality,” which confinesthe applications to outer-level decisions rather than inneractuator commands. In particular, Pearre and Brown [19]and Dunn et al. [20] took the flight path angle as the actionof the reinforcement learning algorithm and constructedthe reward function using energy consumption, time spent,and collision loss to flexibly plan and optimize the referenceflight path. To further optimize the flight path, Zhang et al.[21] proposed a “geometric reinforcement learning” algo-rithm to provide more available actions by transferringbetween nonadjacent states. Palunko et al. [22] and Faust[23] introduce a reinforcement learning algorithm to theadaptive acceleration control of a quadrotor with suspendedload. The primary purpose of the above studies is to generatethe outer loop guidance law; however, the inner loop control-lers still need to be designed. In [24], a reinforcement learn-ing approach for the airship model parameter identificationis suggested by Ko et al., which can reduce the control errorcaused by the model uncertainties. In a previous work in[25], Rottmann et al. designed an airship altitude control-ler based on reinforcement learning and completed theonline training process in a few minutes. However, the con-troller training of [24, 25] is performed in the restrictedlow-dimensional state space, and therefore, the approach isdifficult to be directly employed in the multidimensionalmotion control. Hwangbo et al. [26] proposed a neuralnetwork quadrotor control strategy based on a reinforce-ment learning algorithm, and the well-trained controllercan achieve autonomous flight control even if the quadrotoris thrown upside down into the air, which is superior tohuman operators. But in [26], the quadrotor’s model wasused for training the neural network controller off-line toavoid the “curse of dimensionality” problem, which broughtin the risk of control failure when the dynamic model is inac-curate or unavailable.

For the purpose of acquiring a feasible airship controlstrategy without dependence on an accurate dynamic model,this paper directly takes the inner actuator commands as theaction of the reinforcement learning algorithm and imple-ments the training by actual motion. The main contributionsof this paper are as follows:

(1) The MDP models of the airship 3D path-followingcontrol are established based on motion analysis,which provide new models and methods for airshipcontrol. Furthermore, in contrast to the previous rein-forcement learning airship control without depen-dence on an accurate dynamic model, for the firsttime, this study proposed an airship control strategythrough autonomous online training in 3D space

(2) To deal with the “curse of dimensionality” problemin the airship planar path-following control, a novelcoordinate frame is proposed to describe the relativestate between the airship and the target. This coor-dinate transformation form can reduce the scale ofthe state space and generalize the experience byrotation and makes it possible to learn the controlstrategy online

(3) In the proposed reinforcement learning airship con-trol strategy, a CMAC neural network is employedto generalize the experience in a local neighborhood,which effectively accelerates the training process

The rest of the paper is organized as follows. Section 2 pre-sents the problem formulation and establishes the MDPmodels of the airship control. Section 3 introduces the struc-ture, the principles, and the training process of the proposedairship reinforcement learning control strategy. Section 4 pre-sents the simulation results, and Section 5 concludes the paper.

2. Problem Formulation and the MDP Model ofthe Airship Control

2.1. Problem Formulation. After takeoff and ascent, the air-ship stays steadily near its design altitude. Thereafter, the alti-tude of the airship is mainly controlled by an elevator, andthe horizontal motion is mainly controlled by a propeller,rudder, and vector propeller. Compared with the fixed-wingaircraft, the airship has a smaller roll angle, and the rollingmotion has little effect on the horizontal motion; hence, thecoupling of the airship longitudinal motion and lateralmotion is weak [27]. To simplify the control problem, wedecompose the airship 3D path-following control into thealtitude control and the planar path-following control.

The autonomous training process of reinforcement learn-ing is similar to the animal learning process, in which an intu-itive objective will contribute to the acceleration of learning. Inthis paper, the airship completes the path-following missionby reaching the target points on the desired path one by one.Ogxgygzg represents the earth reference frame (ERF), wherexg, yg, zg denotes the position of the airship. Let h denote thealtitude of the airship which satisfies h = −zg; then, the posi-tion of the airship can be described as xg, yg, h .

In the airship altitude control, suppose the desired alti-tude is hat and the control objective is given by

h − hat < εh, 1

where εh is defined as the valid range of the desired altitude.In the airship planar path-following control, let xt, yt

denote the position of the target point, and the control objec-tive can be described as

xg − xt2 + yg − yt

2< εr , 2

where εr indicates the valid area of the target point.

2 International Journal of Aerospace Engineering

Page 3: Three-Dimensional Path-Following Control of a Robotic

Remark 1. During the training process, the desired altitudeand the target point of the two controllers can be set ran-domly and independently. In the 3D path-following control,we choose the planar path-following control as the main con-trol mission and the altitude control as the auxiliary. Toensure that the actual path is close to the desired 3D path,the horizontal position of the current target is taken as thetarget point in the planar path-following control and the alti-tude of the nearest point on the desired path to the airshipis taken as the desired altitude in the altitude control. Letx1, y1, h1 denote the current target, x0, y0, h0 denote thelast reached target, and xg, yg, h denote the position of theairship. Then, the target point xt, yt in the planarpath-following control can be set as x1, y1 and the desiredaltitude hat in the altitude control can be calculated as

hat =H1 ·Hg

H12 · h1 − h0 + h0, 3

where H1 and Hg are given by

H1 = x1 − x0, y1 − y0, h1 − h0 ,

Hg = xg − x0, yg − y0, h − h04

2.2. MDP Model of the Airship Control. MDP is the basis forthe implementation of reinforcement learning; thus, weestablish the MDP models of the airship altitude controland airship planar path-following control first. MDP can beexpressed by five elements S,A, r, P, J , where S indicatesthe state space constructed by parameters such as the posi-tion, speed, and attitude of the airship; A is the action setcomposed of available control actuator commands, r is thereward of the state and action, P is the state transition prob-ability, and J is the optimization objective function of thesequential decision. MDP satisfies the following property:

p st+1 = sj st = si, at = ak, st−1, at−1,⋯,s0, a0= p st+1 = sj st = si, at = ak = pij ak

 ∀si, sj ∈ S, ak ∈A,∀t ≥ 0,

5

where pij ak is defined as the transition probability fromstate si to state sj after taking the action ak, and t is the timeparameter. It can be seen from the property (5) that param-eters of the state space S are sufficient statistics of the airshipmotion. The primary work of airship control MDP model-ling is to construct the state space. By reasonably selectingparameters to construct state space, the property (5) can beapproximately satisfied.

2.2.1. State Space S of the Altitude Control. As shown inFigure 1, let Ogxgygzg denote the earth reference frame(ERF) and Obxbybzb denote the body reference frame (BRF)attached to the airship. The attitude of the airship isdescribed as the Euler angles ϕ, θ, ψ T in EBF. The speedand angular velocity of the airship in BRF are defined as

u, v,w T and p, q, r T , respectively. δR , δE, and FT representthe rudder deflection, the elevator deflection and the propel-ler thrust of the airship, respectively.

The airship resultant velocity is defined as V in ERF,which satisfies the following equation:

V = u2 + v2 +w2 6

According to the kinematics of the airship, the speed inERF xg, yg, zg

T can be calculated as

xg, yg, zgT=K u, v,w T 7

K is the rotation matrix from BRF to ERF, which isgiven by

K =

cθcψ sθcψsϕ − sψcϕ sθcψcϕ + sψsϕ

cθcψ sθsψsϕ + cψcϕ sθsψcϕ − cψsϕ

−sθ cθsϕ cθcϕ

, 8

where sx is sin x and cx is cos x.According to (7) and h = −zg, the altitude-changing rate h

can be calculated as

h = −zg = sin θ · u − cos θ sin ϕ · v − cos θ cos ϕ ·w 9

In the actual motion of the airship, the speeds v andw aremuch smaller than u, and due to the strong restoring torqueand large damping of the rolling motion, the roll angle ϕ issmall. Therefore, from (6) and (9), approximate equationscan be obtained as

V = u2 + v2 +w2 ≈ u,

h ≈ sin θ · u ≈ sin θ · V10

It is can be seen that the three parameters of the velocityV , the pitch angle θ, and the altitude-changing rate h havestrong correlation.

The airship motion has 12 state parameters of position,velocity, angle, and angular velocity in 3 coordinate axes. If

Rudder

Elevator

q

v r

p

u

xb

zg

yg

Og xg

FT

zb

yb

Ob

𝛿E

𝛿R

w

Figure 1: Coordinate frame of the airship.

3International Journal of Aerospace Engineering

Page 4: Three-Dimensional Path-Following Control of a Robotic

all the parameters are taken into account in the airship alti-tude control, the scale of the state space and the amount ofmemory to store value function estimates will be too largefor the controller to complete the training in a limited time.To avoid this “curse of dimensionality” problem, it is neces-sary to simplify the state parameters according to their rele-vance to motion control.

The parameters related to the airship altitude controlmainly include the altitude h, the altitude-changing rate h,the velocity V , and the pitch angle θ. From (10), h, V , andθ have strong correlation; thus, two of them can contain themain information. Since the speed parameters are relativelyeasy to measure, we construct the state space of the airshipaltitude control MDP model by h, h, V .

2.2.2. State Space S of the Planar Path-Following Control.According to the control objective (2), the state parametersrelated to the planar path-following control mainly includethe position of the airship (xg, yg), the position of the target(xt, yt), the velocity V , and the course angle ψa. We use theparameters xg, yg, xt, yt, ψa,V to describe the main infor-mation of the airship planar path-following control.

Remark 2. The yaw angle ψ of airship is also an importantparameter for the airship path-following control. However,the course angle ψa has strong correlation with the yawangle ψ due to the course stability of the airship. Underthe influence of the airship vertical tail, the course angleψa will soon follow the yaw angle ψ, and the differencebetween them is limited. Since the course angle ψa is directlyrelated to the planar motion, we choose ψa to describe thestate of the airship.

To avoid the “curse of dimensionality” problem duringthe training process in multidimensional space, the parame-ters xg, yg, xt, yt, ψa, V can be simplified and merged basedon airship motion analysis. A common method is to rebuildthe coordinate frame with a relative position. Denote targetposition xt, yt as the origin, and the relative position can

be calculated as xr = xg − xt and yr = yg − yt; then, the stateof the airship can be expressed as xr, yr, ψa, V . To furtheraccelerate the training process of the planar path-followingcontroller, the experience can be generalized by rotatingaround the origin. Inspired by this idea, we present a novelcoordinate frame lr, ψr, V to describe the relative statebetween the airship and target, as shown in Figure 2.

In Figure 2, lr represents the distance from the airship tothe target, ψt is the angle between the Ogxg axis and the linefrom the airship to the target, and ψr is the angle betweenthe line from the airship to the target and the direction ofV . According to the geometrical relation, the distance lr,the angle ψr, and the velocity V can describe the relative statebetween the airship and target, which are calculated by

lr = x2r + y2r = xg − xt2 + yg − yt

2,

ψr = ψa − ψd = ψa − tan−1yt − ygxt − xg

11

By means of coordinate transformation, the state spaceof the planar path-following control can be constructed bylr, ψr, V , and the dimensions of the planar path-followingcontrol state space is reduced from 6 to 3, which effectivelyavoids the “curse of dimensionality” problem in multidimen-sional online training.

2.2.3. A, r, P, J of the Airship Control. The action set A ofMDP can be flexibly adjusted according to the actual controlactuators. In this study, the action setA of the airship altitudecontrol is constructed by elevator deflection and A of the pla-nar path-following control is constructed by rudder deflec-tion. The reward function r consists of the positive rewardR+ wherein the airship reaches the valid target area and thenegative reward R− wherein the airship flies outside theborder. In this study, the controller is trained online andthe state transition probability P is determined by the actualresponse of the airship to the control output. The optimiza-tion objective function J is set as the total cumulative reward.Let π denote the action policy and J∗ denote the optimalcumulative reward, which is given by

J∗ =maxπ

J =maxπ

E 〠∞

t=0γtrt , 12

where rt is the reward at time t and γ ∈ 0, 1 is the rewarddiscount factor.

3. Path-Following Control Design

The reinforcement learning altitude controller and thereinforcement learning planar path-following controllerhave similar structures as shown in Figure 3. The mappingbetween the “state input” and the “control output” is createdautonomously through online training.

Og yg

xgTarget

(xt, yt)

(xg, yg)

lr

𝜓t𝜓r

𝜓a

V

Figure 2: Coordinate frame of the relative state between the airshipand the target.

4 International Journal of Aerospace Engineering

Page 5: Three-Dimensional Path-Following Control of a Robotic

As can be seen from Figure 3, the airship reinforcementlearning controller consists of three modules: the error cal-culation module, the CMAC neural network module, andthe action policy module. The functions of each moduleare as follows.

3.1. Error Calculation. The error of the value function estima-tion is calculated based on the reinforcement learning algo-rithm according to the actual state transition of the airship.This error signal will be used for updating the weights ofthe CMAC neural network.

3.2. CMAC Neural Network. The value function of eachaction is estimated by the CMAC neural network, which pro-vides the decision-making basis for the action policy module.

3.3. Action Policy. The main function of this module is tobalance between exploration and the greedy policy accordingto the value function estimations of actions and then outputthe corresponding control command.

Remark 3.Different from the supervised learning, the errorfor modifying the neural network in the reinforcementlearning controller is not from the sample’s label but iscalculated from the estimation of the value function andreward function. In fact, if there are enough training sampleswith labels provided by an expert controller, similar struc-tures can be used for imitating the control strategy of theexpert controller.

3.4. Q-Learning Algorithm. The reinforcement learning con-troller improves its action policy by interacting with theexternal environment. The controller selects the action basedon the estimation of the value function and updates the esti-mation to maximize the cumulative reward which is given bythe external environment with a random delay. Based on thisconcept, the Q-Learning algorithm estimates and stores theaction value function in a different state and uses a “trialand error” approach to update the estimation of the actionvalue function. Q-Learning is an appropriate reinforcementlearning algorithm for the online training of the airshipcontroller, and the properties of the airship also make ita suitable platform for the Q-Learning algorithm. Comparedwith an airplane and a helicopter, the attitude of the airship isrelatively stable, and the “trial and error” operations of theairship rarely cause a catastrophic consequence. Further-more, the airship is a platform with long endurance, whichcan provide enough time for the training and iterative

function estimation of the algorithm. In this paper, theQ-Learning algorithm is adopted for calculating the modify-ing error of the neural network.

Let Q s, a denote the value function estimation of theaction a in state s, which can be calculated as

Q s, a = E 〠∞

t=0γtrt s0 = s, a0 = a , 13

where rt is the reward at time t, γ is the reward discount fac-tor, s0 is the initial state, and a0 is the initial action.

According to the theory of operation research, Q s, asatisfies the following Bellman equation:

Q st , at =〠st+1

p st , at , st+1 · r st , at , st+1

+ γ 〠st+1,at+1

p st , at , st+1 ·Q st+1, at+1 ,14

where p st , at , st+1 denotes the transition probability fromstate st to state st+1 after taking the action at and r st , at ,st+1 indicates the reward of the action at and the transitionfrom st to st+1. Let π denote the action policy; then, the opti-mal action policy π∗ s and the optimal action estimationQπ∗

s, a corresponding to equation (14) can be describedas follows:

Qπ∗s, a =max

πQ s, a ,

π∗ s = arg maxπ

Qπ∗s, a

15

From (14) and (15), the one-step update formula ofQ s, a under the optimal action policy π∗ s can beobtained as

Q st , at =Q st , at + αQ r st , at , st+1

+ γ maxat+1

Q st+1, at+1 −Q st , at ,16

where αQ is the learning rate of the Q-Learning algorithm.

3.5. CMAC Neural Network. In actual control applications,the multidegrees of system freedom always result in the largescale of the state space. In order to accelerate the training

Target(altitude

orposition)

Stateinput CMAC

Error

Controloutput Airship

Actualstate

Rewardfunction

Actionpolicy 𝜋

Reinforcementlearning

controller

Figure 3: Structure of the airship reinforcement learning controller.

5International Journal of Aerospace Engineering

Page 6: Three-Dimensional Path-Following Control of a Robotic

process of the reinforcement learning controller, an effectiveapproach is to fit the value function based on the theory ofstochastic process and the neural network. Neural networkscan be divided into global approximation neural networksand local approximation neural networks. For global approx-imation neural networks like the Back Propagation (BP)neural network, each weight may affect the output and eachtraining sample will update all the weights. As a result, theconvergence speed is slow and can hardly meet the require-ment of online training. For local approximation neural net-works like the CMAC neural network, only a few weightsaffect the output and each training sample only updates thelocal weights, which is able to quickly approximate time-varying nonlinear functions.

During the training of the reinforcement learning con-troller, there are not enough training samples but onlylimited data from online motion; hence, a rapid convergencespeed is necessary. For an airship, the control strategies arealike in similar states, and the control strategies may becompletely different in very dissimilar states. The advantagesof the local approximation neural network suit these proper-ties of the airship. The CMAC neural network fits the non-linear function by table querying and updates the weightsin the local neighborhood, and it has less computation,which is convenient to be applied in the embedded systemof a flight control computer. For the above advantages, inthis study, the CMAC neural network is employed to gener-alize the training experience of the airship reinforcementlearning controller.

The structure of a typical CMAC neural network isshown in Figure 4, where S is the state space formed bym states, W is the weights stored in n memory addresses,and ynet is the output of the network. The actual state inputis mapped to state s in the state space S, and each state sicorresponds to an activation vector F si = Fi. The W andFi are given by

W = w1,w2,⋯,wj⋯,wnT ,

Fi = f i1, f i2,⋯,f ij⋯,f inT i = 1, 2⋯,m,

 j = 1, 2⋯,n,

17

where f ij = 1 or 0 represents the activation state of si to wj.For any two states, the closer the states are, the more

overlaps their activation vectors have. And if two statesare far away from each other, their activation vectors haveno overlap.

The output ynet si of the CMAC neural network can beobtained as

ynet si =WTF si =WTFi = 〠n

j=1wjf ij 18

We establish one CMAC neural network for each actionin the action set A, and the network weights of a ∈A can bedefined asWa. Letting Q s, a denote the value function esti-mation of the state-action pair s, a , we have

Q s, a =WTa F s 19

Then, the partial derivative of Q s, a to Wa is

∂Q s, a∂Wa

= F s 20

Suppose at as the action at time t, then in state st , thevalue function estimation error δt of Q st , at is given by

δt =Q st , at −Q st , at 21

The network weightsWatof action at can be updated by a

gradient algorithm

“at t+1 =Wat t+ αWδt

∂Q st , at∂Wat

, 22

where αW is the learning rate of the weights. By substituting(20) and (21) into (22), we have

Wat t+1 =Wat t+ αW Q st , at −Q st , at F st , 23

but Q st , at is unknown. Based on the concept (16)of the Q-Learning algorithm, Q st , at can be replaced byr st , at , st+1 + γ max

at+1Q st+1, at+1 which is calculated from

the reward and value function estimation at time t + 1.Taking into account the impact of the state visit frequency,the updated formula of Wat

with a fitness coefficient can bewritten as

Wat t+1 =Wat t− αWet st , at r st , at , st+1

+ γ maxat+1

Q st+1, at+1 −Q st , at F st

=Wat t− αWet st , at r st , at , st+1

+ γ maxat+1

WTat+1

F st+1 −WTatF st F st ,

24

InOut

ynet

S

1~m

1~n

W

Σ

Figure 4: Structure of the CMAC neural network.

6 International Journal of Aerospace Engineering

Page 7: Three-Dimensional Path-Following Control of a Robotic

where e s, a is the fitness coefficient with an initial value of 1,which reflects the state visit frequency:

et+1 s, a =γet s, a , s, a ≠ st , at ,

1, s, a = st , at25

3.6. Process of Training and Control. To avoid being trappedin a local optimum and ensure the convergence speed of thetraining, a proper action policy π is necessary for the rein-forcement learning controller. In this paper, a random actionpolicy based on the Boltzmann distribution is adopted forbalancing between the exploration and greedy policy. Let Adenote the available action set and p a s denote the proba-bility of selecting action a in state s; then, the action policyπ is given by

p at st =eQ st ,at /T temp

∑a∈AeQ st ,a /T temp

, 26

where T temp is the exploration coefficient which is similar tothe temperature coefficient of a simulated annealing algo-rithm. By progressively reducing T temp, this policy ensures a

high exploration probability in the initial states and graduallytrends to the greedy policy.

The training and control process of the proposed airshipreinforcement learning controller is described as shown inFigure 5. The flowchart before the dotted line is the onlinetraining process, and the rest is the actual control process.

As can be seen from Figure 5, during the training process,the airship starts one round of learning after another by ran-domly setting the position of the target. In each round oflearning, the controller selects the action based on policy πperiodically before reaching the target or boundary, and thefitness coefficient and the weights of the CMAC neural net-work are updated at the same time. When the algorithmand the neural network converge, the training process iscompleted and the controller can be applied in actual controlmissions. During the control process, the parameters of thealgorithm are fixed, and the action policy is turned into thegreedy policy. The controller maximizes the cumulativereward by selecting the action corresponding to the maxi-mum value function estimation.

4. Simulation

To test the performance of the proposed controllers, sim-ulations are implemented in MATLAB. We build themodel of a LS-S1200 robotic airship [28] based on theapproach in [9, 27]. The aerodynamic forces and torquecan be calculated as

FaX =Q∞ CX1 cos2α cos2β + CX2 sin 2α sinα

2,

FaY =Q∞ CY1 cosβ

2sin 2β + CY2 sin 2β

+ CY3 sin β sin β + CY4δR

,

FaZ =Q∞ CZ1 cosα

2sin 2α + CZ2 sin 2α

+ CZ3 sin α sin α + CZ4δE ,

La =Q∞ CL2 sin β sin β ,

Ma =Q∞ CM1 cosα

2sin 2α + CM2 sin 2α

+ CM3 sin α sin α + CM4δE

,

Na =Q∞ CN1 cosβ

2sin 2β + CN2 sin 2β

+ CN3 sin β sin β + CN4δR ,

27

where Q∞ = ρV2/2 is the dynamic pressure of the inflow,ρ is the atmospheric density, α is the attack angle, β isthe sideslip angle, FaX ~Na are the aerodynamic forcesand torque of the airship, and CX1 ~ CN4 are the hull aero-dynamic coefficients. The aerodynamic coefficients, the

Start training

Set a random target

N

Act based on policy 𝜋

Output and calculate 𝛿t

Update W, e(s,a)

Reach target

Finish training?

Start control

Input current target

Act based on greedy policy

Reach target?

Control output

N

Fail?

NY

Y

Y

N

Y

UpdateTtemp

TrainingControl

Figure 5: Training and control processes.

7International Journal of Aerospace Engineering

Page 8: Three-Dimensional Path-Following Control of a Robotic

structure parameters, and the reinforcement learning con-troller parameters used in the simulation are listed inTable 1, where ma is the mass of the airship, ∇ is the vol-ume of the airship, Ix ~ Iz are the moments of inertia, andk1 ~ k3 are the inertia factors.

During the training of the controller, the propeller thrustchanges randomly every 15 seconds to simulate the variationof the airship velocity. When the training is completed andthe control mission begins, the airship velocity is maintainedby a simple proportion integral (PI) propeller thrust control-ler, in which the cruising velocity is set to 10m/s.

For convenience in analysing the control performance,the well-tuned PID controller is chosen as a comparison. Inthe outer loop of the PID controller, the desired attitudeangle is calculated based on the line-of-sight (LOS) guidancelaw [29]. In the inner loop, we establish the dynamic equa-tions according to the airship’s structure and aerodynamicparameters and then linearize the equations with a small per-turbation method. After working out the transfer functionfrom the rudder deflection to the attitude angle, the gainparameters of the PID controller can be tuned well usingthe root locus tool RLTOOL in MATLAB.

4.1. Training of Reinforcement Learning Controllers. Thetraining of the proposed airship altitude control and planarpath-following control is independent and can be imple-mented simultaneously. The controller is trained online byrandomly setting the target near the current position of theairship. The setting range of the desired altitude is 10m,and the setting range of the target point is 100m. In thispaper, the weights of the CMAC neural network are savedevery 100 s for test and evaluation. Figure 6 shows the trendof the success rate and the average time spent on the trainingmissions during the training process.

In Figure 6(a), after training for about 1000 s, the missionsuccess rate of the airship altitude controller is stable at 100%;then, the altitude controller can be applied in actual controlmissions. And as shown in Figures 6(c) and 6(d), the training

of the airship planar path-following controller is completedin about 3 hours. In Figures 6(b) and 6(d), the average timespent on the training missions increased first, then decreasedgradually, and eventually tended to be stable. This processcorresponds to three stages of learning: at the initial stage,the controller could not control the airship and failed soon;then, the controller learned to lead the airship to reach thetarget in an indirect and suboptimal path; finally, the actionpolicy was continuously optimized and gradually stabilized.

4.2. Altitude Control. We use the trained reinforcementlearning controller and a well-tuned PID controller toimplement the airship altitude control mission. When thepropeller thrust FT remains constant and the airship keepsa straight flight (δR = 0), the simulation result is shown inFigure 7(a). When the propeller thrust FT and rudder deflec-tion δR change randomly every 15 seconds, the simulationresult is shown in Figure 7(b).

As shown in Figure 7(a), both the reinforcement learningcontroller and the PID controller perform well when the air-ship is in the steady flight state. The altitude control errors ofthe two controllers are within 5m, and the PID controllerperforms slightly better.

In Figure 7(b), it can be seen that when the airship is inthe maneuver flight state, the performance of the reinforce-ment learning controller is slightly affected, and its altitudecontrol error is still within 5m. But the PID controller is sig-nificantly affected and its maximum altitude control errorreaches about 10m, which is consistent with the actual exper-iment data of the airship [28]. The reinforcement learningcontroller performs better than the PID controller when theairship is in the maneuver flight state. This is because duringthe training process, the airship has been in a randommaneuver flight state, and the reinforcement learning con-troller gradually learned the action policy in that state byautonomous adjustment.

4.3. Planar Path-Following Control.A desired path consistingof continuous turns, a straight line, and arcs is used for test-ing the performance of the planar path-following controller.Figure 8 shows the planar path-following control simulationresults. Figure 9 shows the deviation distance of the airship tothe target points and the desired path.

In Figures 8 and 9, it can be seen that the planarpath-following control error of the reinforcement learningcontroller is close to that of the PID controller near the20th target point. At this point, the airship is in the steadyflight state on the desired straight path. During the rest ofthe flight, the airship is in the maneuver flight state, and thecontrol error of the reinforcement learning controller issmaller. This is because the action policy of the reinforce-ment learning controller is not determined by the currentdeviation but based on the total cumulative reward. Whenthe current deviation is small but a significant control erroris about to occur, the reinforcement learning controller canreact in time.

The deviation distance of the reinforcement learningcontroller and the PID controller is compared in Table 2.To complete the planar path-following control mission, the

Table 1: Parameters of the airship model and controller.

ma kg 100 CY2 -10.798 CN1 -36.364

∇ m3 79.0 CY3 -22.852 CN2 56.948

ρ kg/m3 1.225 CY4 2.336 CN3 42.655

Ix kg · m2 324 CZ1 -5.629 CN4 -12.324

Iy kg · m2 474 CZ2 -10.798 εh m 1.5

Iz kg · m2 202 CZ3 -21.622 εr m 15

k1 0.084 CZ4 -2.336 γ 0.9

k2 0.856 CL2 2.337 αW 0.003

k3 m2 5.543 CM1 36.364 T temp 40

CX1 -0.518 CM2 -56.948 R+ 10

CX2 -5.629 CM3 -42.655 R− -10

CY1 -5.629 CM4 -12.324

8 International Journal of Aerospace Engineering

Page 9: Three-Dimensional Path-Following Control of a Robotic

reinforcement learning controller takes 419.2 s and the PIDcontroller takes 443.7 s. By comparing the above data, it isclear that the reinforcement learning controller not only isslightly better in terms of control accuracy but also takes lesstime in the whole control mission.

To test the adaptability of the reinforcement learningcontroller, we can assume that asymmetric aerodynamicshape changes occur in the airship dynamic model. In thisexample, an additional yaw torque coefficient (CNadd = −4 9)is added to the dynamic equation, which will lead to theasymmetric changes in the airship’s turning radius andturning ability. As a result, the performance of the original

controllers may get worse, but the reinforcement learningcontroller can autonomously adjust its action policy by train-ing online. Figure 10 shows the simulation result of the con-trollers after the airship’s dynamic model changed.

As shown in Figure 10, the airship’s right turning abil-ity is weaker than the left due to the asymmetric aerody-namic shape. Controlled by the PID controller, the airshipkept circling around the 2nd target point but cannot getcloser. Through expanding the valid target area, the PIDcontroller can complete the control mission, but its controlerror becomes larger because of the relaxation of the con-trol requirements. After online training, the reinforcement

100

80

60

40

20

0

0 1500

Succ

ess r

ate (

%)

3000t (s)

4500 6000

(a) Success rate of missions in the altitude controller training

30

25

20

10

15

5

0

t miss

ion (

s)

0 1500 3000t (s)

4500 6000

(b) Average time spent of missions in the altitude controller training

100

80

60

40

20

0

Succ

ess r

ate (

%)

0 3000 6000t (s)

9000 12000

(c) Success rate of missions in the planar

path-following controller training

100

80

60

40

20

0

t miss

ion (

s)

0 3000 6000t (s)

9000 12000

(d) Average time spent of missions in the planar

path-following controller training

Figure 6: Trend of the success rate and the average time spent.

120

110

100

90

0 200

PIDReinforcement learningDesired altitude

400 600t (s)

h (m

)

800

(a) Altitude control with constant FT and δR

120

110

100

90

h (m

)

0 200

PIDReinforcement learningDesired altitude

400 600t (s)

800

(b) Altitude control with varied FT and δR

Figure 7: The airship altitude control.

9International Journal of Aerospace Engineering

Page 10: Three-Dimensional Path-Following Control of a Robotic

learning controller autonomously adapts to the modelchanges and is able to complete the control mission withoutexpanding the valid target area. Furthermore, it is remarkablethat near the 11th target point, the reinforcement learningcontroller shows an intelligent decision-making ability simi-lar to human operators. When the airship could not reach the11th target point directly because of the weak right turningability, the reinforcement learning controller implementedthe action policy of flying further at first and then turningleft, which successfully led the airship to reach the target.This is owing to the experience of reaching the target in an

indirect path that the reinforcement learning controllergained during the training process.

4.4. 3D Path-Following Control. The 3D path-following con-trol mission can be completed through the cooperation of theproposed controllers. In this paper, the target point of theplanar path-following controller and the desired altitude ofthe altitude controller are calculated according to the follow-ing method.

Planar path-following control: in the 3D path-followingcontrol, the planar path-following is chosen as the main con-trol mission; therefore, we project the 3D desired path ontothe horizontal plane and create a sequence of target points.

Altitude control: while the airship is flying to the targetpoint, the altitude controller works as an auxiliary andupdates its desired altitude periodically. As shown in Remark1, the altitude of the nearest point on the 3D desired path tothe airship is taken as the desired altitude.

Taking a spiral path as an example, the simulationresult of the proposed airship reinforcement learning con-troller is shown in Figure 11, where the horizontal distancebetween two adjacent target points is 100m and the verticaldistance is 10m. It can be seen from Figure 11 that the per-formance of the planar path-following controller in the 3Dpath-following control is similar to that in the independentplanar control, while the altitude of the airship slightly showsstep changes. This is because the desired altitude of the alti-tude controller is updated periodically. The overall simula-tion result shows that the reinforcement learning planar

800

600

400

200

0

2000 200

Target pointDesired path

PIDReinforcement learning

400 600xg (m)

yg (

m)

800 1000

20

1200

20

Figure 8: The airship planar path-following control.

20

15

10

5

0

Dist

ance

(m)

5

PIDReinforcement learning

10 15Target point number

4035302520

(a) Distance from the airship to the target points

40

30

20

10

0

Dist

ance

(m)

5000

PIDReinforcement learning

1000 1500Desired path length (m)

3500300025002000

(b) Distance from the airship to the desired path

Figure 9: The deviation distance in the planar path-followingcontrol.

Table 2: The maximum and the average deviation distance.

Deviation distance PID Reinforcement learning

Airship to target maximum (m) 17.6 13.2

Airship to target average (m) 10.1 4.2

Airship to path maximum (m) 40.9 29.2

Airship to path average (m) 19.9 4.8

800

600

400

200

0

200

yg (

m)

112

0 200

Target point

Desired path

PID with expandingthe valid target area

PIDReinforcement learning

400 600xg (m)

800 1000 1200

111122

Figure 10: The planar path-following control after the airship’sdynamic model changed.

10 International Journal of Aerospace Engineering

Page 11: Three-Dimensional Path-Following Control of a Robotic

path-following controller and altitude controller do notinterfere with each other and cooperate well.

5. Conclusions

In this paper, a reinforcement learning airship 3Dpath-following control is proposed, which is robust againstdynamic model changes. To avoid the “curse of dimensional-ity” problem, the MDP models of the airship altitude controland planar path-following control are established, and thescale of the state space is reduced by parameter simplificationand coordinate transformation. The airship reinforcementlearning controllers are designed based on the Q-Learningalgorithm and CMAC neural network, and the training pro-cess can converge in 3 hours. The reinforcement learningaltitude controller and planar path-following controller canachieve comparable performance to well-tuned PID control-lers, and the 3D path-following control mission can be com-pleted through cooperation of the proposed controllers.Compared with existing approaches, the main advantage ofthis control strategy is that when the airship model dynamicparameters are uncertain or wrong, the controller is able toadjust its action policy autonomously and make intelligentdecisions similar to human operators.

Data Availability

The data used to support the findings of this study areavailable from the corresponding author upon request.

Conflicts of Interest

The authors declare that there is no conflict of interestregarding the publication of this paper.

Acknowledgments

This work was supported by the National Natural ScienceFoundation of China (No. 61503010), the Aeronautical Sci-ence Foundation of China (No. 2016ZA51001), and the Fun-damental Research Funds for the Central Universities (No.YWF-19-BJ-J-118).

References

[1] L. Chen, D. P. Duan, and D. S. Sun, “Design of a multi-vectoredthrust aerostat with a reconfigurable control system,” AerospaceScience and Technology, vol. 53, pp. 95–102, 2016.

[2] Y. Yang, J. Wu, and W. Zheng, “Positioning control for anautonomous airship,” Journal of Aircraft, vol. 53, no. 6,pp. 1638–1646, 2016.

[3] Y. Yang, “A time-specified nonsingular terminal sliding modecontrol approach for trajectory tracking of robotic airships,”Nonlinear Dynamics, vol. 92, no. 3, pp. 1359–1367, 2018.

[4] Z. Zheng and L. Xie, “Finite-time path following control for astratospheric airship with input saturation and error con-straint,” International Journal of Control, pp. 1–26, 2017.

[5] Y. Yang and Y. Ye, “Backstepping sliding mode control foruncertain strict-feedback nonlinear systems using neural-network-based adaptive gain scheduling,” Journal of SystemsEngineering and Electronics, vol. 29, no. 3, pp. 140–146, 2018.

[6] Z. Zheng, L. Liu, and M. Zhu, “Integrated guidance and con-trol path following and dynamic control allocation for a strato-spheric airship with redundant control systems,” Proceedingsof the Institution of Mechanical Engineers, Part G: Journal ofAerospace Engineering, vol. 230, no. 10, pp. 1813–1826, 2016.

[7] Y. Yang and Y. Yan, “Attitude regulation for unmanned quad-rotors using adaptive fuzzy gain-scheduling sliding mode con-trol,” Aerospace Science and Technology, vol. 54, pp. 208–217,2016.

[8] S. Liu, Y. Sang, and H. Jin, “Robust model predictive controlfor stratospheric airships using LPV design,” Control Engineer-ing Practice, vol. 81, pp. 231–243, 2018.

[9] G. A. Khoury,Airship Technology, Cambridge university press,2012.

[10] T. Du, L. Guo, and J. Yang, “A fast initial alignment for SINSbased on disturbance observer and Kalman filter,” Transac-tions of the Institute of Measurement and Control, vol. 38,no. 10, pp. 1261–1269, 2016.

[11] T. Du and L. Guo, “Unbiased information filtering for systemswith missing measurement based on disturbance estimation,”Journal of the Franklin Institute, vol. 353, no. 4, pp. 936–954,2016.

[12] D. Silver, A. Huang, C. J. Maddison et al., “Mastering the gameof Go with deep neural networks and tree search,” Nature,vol. 529, no. 7587, pp. 484–489, 2016.

[13] R. S. Sutton and A. G. Barto, Reinforcement Learning: an Intro-duction, MIT press, 2018.

[14] J. Kober, J. A. Bagnell, and J. Peters, “Reinforcement learningin robotics: a survey,” The International Journal of RoboticsResearch, vol. 32, no. 11, pp. 1238–1274, 2013.

[15] D. L. Leottau, J. Ruiz-del-Solar, and R. Babuška, “Decentra-lized reinforcement learning of robot behaviors,” ArtificialIntelligence, vol. 256, pp. 130–159, 2018.

[16] E. Walraven, M. T. J. Spaan, and B. Bakker, “Traffic flow opti-mization: a reinforcement learning approach,” Engineering

Target point

250

200

150

100

50

h (m

)

0

xg (m) yg (m)

−600−500

−400−300

−200−100

0

Start

−300−200

−1000

100200

300

Desired pathActual path

Start

Figure 11: The airship 3D path-following control.

11International Journal of Aerospace Engineering

Page 12: Three-Dimensional Path-Following Control of a Robotic

Applications of Artificial Intelligence, vol. 52, pp. 203–212,2016.

[17] S. Kosunalp, Y. Chu, P. D. Mitchell, D. Grace, and T. Clarke,“Use of Q-learning approaches for practical medium accesscontrol in wireless sensor networks,” Engineering Applicationsof Artificial Intelligence, vol. 55, pp. 146–154, 2016.

[18] V. Mnih, K. Kavukcuoglu, D. Silver et al., “Human-level con-trol through deep reinforcement learning,” Nature, vol. 518,no. 7540, pp. 529–533, 2015.

[19] B. Pearre and T. X. Brown, “Model-free trajectory optimisationfor unmanned aircraft serving as data ferries for widespreadsensors,” Remote Sensing, vol. 4, no. 10, pp. 2971–3005, 2012.

[20] C. Duun, J. Valasek, and K. Kirkpatrik, “Unmanned air systemsearch and localization guidance using reinforcement learn-ing,” in Infotech@ Aerospace, pp. 1–8, Garden Grove, CA,USA, June 2012.

[21] B. Zhang, Z. Mao, W. Liu, and J. Liu, “Geometric reinforce-ment learning for path planning of UAVs,” Journal of Intelli-gent and Robotic Systems, vol. 77, no. 2, pp. 391–409, 2015.

[22] I. Palunko, A. Faust, P. J. Cruz, L. Tapia, and R. Feirro, “A rein-forcement learning approach towards autonomous suspendedload manipulation using aerial robots,” in IEEE InternationalConference on Robotics and Automation, pp. 1–6, Karlsruhe,Germany, May 2013.

[23] A. Faust, Reinforcement Learning and Planning for PreferenceBalancing Tasks, University of New Mexico, 2014.

[24] J. Ko, D. J. Klein, D. Fox, and D. Haehnel, “Gaussian processesand reinforcement learning for identification and control of anautonomous blimp,” in 2007 IEEE International Conference onRobotics and Automation, pp. 742–747, Roma, Italy, April2007.

[25] A. Rottmann, C. Plagemann, P. Hilgers, and W. Burgard,“Autonomous blimp control using model-free reinforcementlearning in a continuous state and action space,” in 2007IEEE/RSJ International Conference on Intelligent Robots andSystems, pp. 1895–1900, San Diego, CA, USA, October-No-vember 2007.

[26] J. Hwangbo, I. Sa, R. Siegwart, and M. Hutter, “Control of aquadrotor with reinforcement learning,” IEEE Robotics andAutomation Letters, vol. 2, no. 4, pp. 2096–2103, 2017.

[27] Z. Zheng, Y. Huang, L. Xie, and B. Zhu, “Adaptive trajectorytracking control of a fully actuated surface vessel with asym-metrically constrained input and output,” IEEE Transactionson Control Systems Technology, vol. 26, no. 5, pp. 1851–1859,2018.

[28] Z. Zheng, W. Huo, and Z. Wu, “Autonomous airship path fol-lowing control: theory and experiments,” Control EngineeringPractice, vol. 21, no. 6, pp. 769–788, 2013.

[29] Z. Zheng, L. Sun, and L. Xie, “Error-constrained LOS path fol-lowing of a surface vessel with actuator saturation and faults,”IEEE Transactions on Systems, Man and Cybernetics: Systems,vol. 48, no. 10, pp. 1794–1805, 2018.

12 International Journal of Aerospace Engineering

Page 13: Three-Dimensional Path-Following Control of a Robotic

International Journal of

AerospaceEngineeringHindawiwww.hindawi.com Volume 2018

RoboticsJournal of

Hindawiwww.hindawi.com Volume 2018

Hindawiwww.hindawi.com Volume 2018

Active and Passive Electronic Components

VLSI Design

Hindawiwww.hindawi.com Volume 2018

Hindawiwww.hindawi.com Volume 2018

Shock and Vibration

Hindawiwww.hindawi.com Volume 2018

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawiwww.hindawi.com Volume 2018

Hindawiwww.hindawi.com Volume 2018

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawiwww.hindawi.com

Volume 2018

Hindawi Publishing Corporation http://www.hindawi.com Volume 2013Hindawiwww.hindawi.com

The Scientific World Journal

Volume 2018

Control Scienceand Engineering

Journal of

Hindawiwww.hindawi.com Volume 2018

Hindawiwww.hindawi.com

Journal ofEngineeringVolume 2018

SensorsJournal of

Hindawiwww.hindawi.com Volume 2018

International Journal of

RotatingMachinery

Hindawiwww.hindawi.com Volume 2018

Modelling &Simulationin EngineeringHindawiwww.hindawi.com Volume 2018

Hindawiwww.hindawi.com Volume 2018

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawiwww.hindawi.com Volume 2018

Hindawiwww.hindawi.com Volume 2018

Navigation and Observation

International Journal of

Hindawi

www.hindawi.com Volume 2018

Advances in

Multimedia

Submit your manuscripts atwww.hindawi.com