
In the control of the self-driving vehicles, PID controllers are widely used due to their simple structure and good stability. However, in complex self-driving scenarios such as curvature curves, car following, overtaking, etc., it is necessary to ensure the stable control accuracy of the vehicles. Some researchers used fuzzy PID to dynamically change the parameters of PID to ensure that the vehicle control remains in a stable state. It is difficult to ensure the control effect of the fuzzy controller when the size of the domain is not selected properly. This paper designs a variable-domain fuzzy PID intelligent control method based on Q-Learning to make the system robust and adaptable, which is dynamically changed the size of the domain to further ensure the control effect of the vehicle. The variable-domain fuzzy PID algorithm based on Q-Learning takes the error and the error rate of change as input and uses the Q-Learning method to learn the scaling factor online so as to achieve online PID parameters adjustment. The proposed method is verified on the Panosim simulation platform.The experiment shows that the accuracy is improved by 15% compared with the traditional fuzzy PID, which reflects the effectiveness of the algorithm.
Citation: Yongqiang Yao, Nan Ma, Cheng Wang, Zhixuan Wu, Cheng Xu, Jin Zhang. Research and implementation of variable-domain fuzzy PID intelligent control method based on Q-Learning for self-driving in complex scenarios[J]. Mathematical Biosciences and Engineering, 2023, 20(3): 6016-6029. doi: 10.3934/mbe.2023260
[1] | Jin Zhang, Nan Ma, Zhixuan Wu, Cheng Wang, Yongqiang Yao . Intelligent control of self-driving vehicles based on adaptive sampling supervised actor-critic and human driving experience. Mathematical Biosciences and Engineering, 2024, 21(5): 6077-6096. doi: 10.3934/mbe.2024267 |
[2] | Yong Xiong, Lin Pan, Min Xiao, Han Xiao . Motion control and path optimization of intelligent AUV using fuzzy adaptive PID and improved genetic algorithm. Mathematical Biosciences and Engineering, 2023, 20(5): 9208-9245. doi: 10.3934/mbe.2023404 |
[3] | Jianzhong Shi, Ying Song . Mathematical analysis of a simplified general type-2 fuzzy PID controller. Mathematical Biosciences and Engineering, 2020, 17(6): 7994-8036. doi: 10.3934/mbe.2020406 |
[4] | Kangsen Huang, Zimin Wang . Research on robust fuzzy logic sliding mode control of Two-DOF intelligent underwater manipulators. Mathematical Biosciences and Engineering, 2023, 20(9): 16279-16303. doi: 10.3934/mbe.2023727 |
[5] | Hongjie Deng, Lingxi Peng, Jiajing Zhang, Chunming Tang, Haoliang Fang, Haohuai Liu . An intelligent aerator algorithm inspired-by deep learning. Mathematical Biosciences and Engineering, 2019, 16(4): 2990-3002. doi: 10.3934/mbe.2019148 |
[6] | Liqiong Pu, Zhigui Lin . A diffusive SIS epidemic model in a heterogeneous and periodically evolvingenvironment. Mathematical Biosciences and Engineering, 2019, 16(4): 3094-3110. doi: 10.3934/mbe.2019153 |
[7] | Wenjie Liang . Research on a vehicle and pedestrian detection algorithm based on improved attention and feature fusion. Mathematical Biosciences and Engineering, 2024, 21(4): 5782-5802. doi: 10.3934/mbe.2024255 |
[8] | Zhanying Tong, Yingying Zhou, Ke Xu . An intelligent scheduling control method for smart grid based on deep learning. Mathematical Biosciences and Engineering, 2023, 20(5): 7679-7695. doi: 10.3934/mbe.2023331 |
[9] | Anlu Yuan, Tieyi Zhang, Lingcong Xiong, Zhipeng Zhang . Torque control strategy of electric racing car based on acceleration intention recognition. Mathematical Biosciences and Engineering, 2024, 21(2): 2879-2900. doi: 10.3934/mbe.2024128 |
[10] | Muhammad Akram, Ayesha Khan, Uzma Ahmad, José Carlos R. Alcantud, Mohammed M. Ali Al-Shamiri . A new group decision-making framework based on 2-tuple linguistic complex q-rung picture fuzzy sets. Mathematical Biosciences and Engineering, 2022, 19(11): 11281-11323. doi: 10.3934/mbe.2022526 |
In the control of the self-driving vehicles, PID controllers are widely used due to their simple structure and good stability. However, in complex self-driving scenarios such as curvature curves, car following, overtaking, etc., it is necessary to ensure the stable control accuracy of the vehicles. Some researchers used fuzzy PID to dynamically change the parameters of PID to ensure that the vehicle control remains in a stable state. It is difficult to ensure the control effect of the fuzzy controller when the size of the domain is not selected properly. This paper designs a variable-domain fuzzy PID intelligent control method based on Q-Learning to make the system robust and adaptable, which is dynamically changed the size of the domain to further ensure the control effect of the vehicle. The variable-domain fuzzy PID algorithm based on Q-Learning takes the error and the error rate of change as input and uses the Q-Learning method to learn the scaling factor online so as to achieve online PID parameters adjustment. The proposed method is verified on the Panosim simulation platform.The experiment shows that the accuracy is improved by 15% compared with the traditional fuzzy PID, which reflects the effectiveness of the algorithm.
PID controller and fuzzy logic-based controller dominate the field of intelligent automation control by virtue of their scalability and structural simplicity[1]. They have been widely used in vehicle corner speed control[2,3]. However, these conventional controllers are highly parameter-dependent. In practice, the adjustment of parameters relies mainly on expert experiences, which is not only time-consuming, but once unable to automaticallly adjust when the control parameters are determined. And It cannot be executed well in a single scene. Generally, if the system is used for other tasks, the control effect often change, and the control parameters need to be adjusted frequently[4]. There are many diversity, time-varying and uncertainty problems in typical driving scenes, such as overtaking, meeting on narrow road, etc.[5,6]. Therefore, tuning the parameters of conventional controllers to achieve optimal performance has become a popular research direction[7]. To handle complicated processes and combine the benefits of traditional controllers with human operator knowledge, fuzzy control[8] has recently emerged as an alternative to conventional control algorithms. Sang Hyuk Park[9] proposed a method for fuzzy self-tuning PID controller and a method for online tuning of PID controller gain values. Priyam[10] proposed PID and fuzzy PID control for DC motors and compares the results of two methods, concluding that fuzzy PID is more adaptable than classical PID. Quan et al.[11] used fuzzy PID control to study the thermal degradation behavior and kinetics of biomass microwave pyrolysis, which significantly improved the response speed of the system. Jie et al.[12] proposed a Type-2 fuzzy PID controller for improving the drive performance of autonomous mobile robots in static obstacle environments. Neerendra[13] proposed that satisfactory results can be achieved using fuzzy logic and laser sensors for navigation in unknown environments. Muhammad[14] proposed a method for designing robust Fuzzy tuned PD (FPD) controller with better tracking and minimal ball oscillations under different lighting conditions.
Reinforcement learning[15] is a machine learning algorithm that generates optimal policies by interacting with the environment. Various approaches such as Q-Learning and deep deterministic policy gradient algorithm (DDPG) have been developed for the particular self-driving decision making and control tasks, driving policy generation[16], the autonomous mobile robot obstacle avoidance[17], robot manipulator control[18], path planning[19], and path tracking[20], etc. Traditional control methods with reinforcement learning have proven to be very effective in solving problems in high-dimensional action spaces and state spaces[21]. Many researches use reinforcement learning algorithms to completely replace traditional controllers and apply them to control strategies. One such approach is to use reinforcement learning algorithms to completely replace traditional controllers. Ramanathan[22] used Q-Learning to control the level in a non-linear conical tank with satisfactory results. It is worth noting, however, that the selection of state space, action space, and rewards all affect the performance of the algorithm in a reinforcement learning algorithm, and designing a controller by reinforcement learning algorithms alone becomes challenging when the number of dimensions increases[23].
Other methods are to combine traditional controllers with reinforcement learning methods to compensate for the shortcomings of traditional controllers. Lakhani et al.[24] proposed an automatic PID tuning framework based on reinforcement learning (RL), particularly the deterministic policy gradient (DPG) method to address traditional PID tuning methods rely on trial and error for complex processes where insights about the system is limited and may not yield the optimal PID parameters. Dogru et al.[25] combined the recent developments in computer sciences and control theory to address the tuning problem. It formulates the PID tuning problem as a reinforcement learning task with constraints.Yu et al.[26] proposed a self-adaptive model-free SAC-PID control method based on reinforcement learning for automatic control of mobile robots. The upper controller based on soft actor-critic (SAC), one of the most competitive continuous control algorithms, and the lower controller based on incremental PID controller.
Q-learning algorithm has been proved as one of the most efficient model-free RL algorithms by directly parameterizing and updating value functions or policies without explicitly modeling the environment. Therefore, we design a variable-domain fuzzy PID controller based on Q-Learning. Specifically, we use a variable-domain fuzzy PID controller to change the initial domain of the fuzzy PID controller by scaling its factors, which can change with the deviation and the rate of change of the deviation to achieve the effect of intelligent adjustment of the PID parameters. However, the control function of a variable-domain fuzzy PID controller can become distorted, which decreases control accuracy. Therefore, this paper combines reinforcement learning algorithm to improve the variable-domain fuzzy PID controller, so that it has the ability of online optimization. To improve the effectiveness of adjusting the PID control settings, these two effects are combined and added.
PID uses differential, integral, and proportional control to calculate the system's error, as is shown in Figure 1.
The PID algorithm uses a linear mix of proportional, integral, and differential error information e(t) to choose the desired control parameters:
u(t)=kpe(t)+ki∫t0e(t)+kdde(t)dt | (2.1) |
where kp, ki, and kd are the proportional, integral and differential coefficients respectively. In self-driving vehicle control, the deviation between e(t)=r(t)−c(t), c(t) is the actual trajectory and r(t) is the preset trajectory. P (proportional) control is the basis of PID control and proportionally reflects the deviation signal of the control system, which is immediately controlled to reduce the deviation as soon as it occurs. When only P control is available, steady-state error or overshoot is generated. I (integral) control is used to eliminate steady-state error (steady-state error is the very small error that still exists between the vehicle's stable driving and the preset trajectory under PD control), and the error is integrated as long as it exists, so that the output continues to increase or decrease until the error is zero, the integration stops and the output no longer changes. D (differential) control reflects the trend of the deviation signal and can introduce an effective early correction signal in the system before the deviation signal becomes too large, thus speeding up the action of the system and reducing the regulation time. In terms of time, the P control adjusts for current errors, the I control adjusts for historical errors and the D control adjusts for future errors.
Although PID control has many advantages, such as simple structure, good stability and so on, it is difficult to perform well in the face of the complex environment of self-driving vehicles and time-varying non-linear systems like speed and corner control using a set of PID parameters.
Fuzzy PID control[28] is a method for optimising the parameters of a PID in real time using fuzzy logic and according to certain fuzzy rules, in order to overcome the shortcomings of traditional PID which cannot adjust the PID parameters in real time. Fuzzy PID control includes components such as fuzzification, determination of fuzzy rules and defuzzification.
The fuzzy PID is in the form of a two-input, three-output controller that takes the error e(t) and the rate of change of the error ec(t)=de(t)/dt as inputs xi(t)(i=1,2) and x1(t)=e(t),x2(t)=ec(t), the three parameters of the PID are proportional, integral and differential corrections Δkp,Δki,Δkd as outputs yj(t)(j=1,2,3) and y1(t)=Δkp,y2(t)=Δki,y3(t)=Δkd and set the initial domain to be respectively Xi=[−Ei,Ei], and Yj=[−UJ,UJ], where Ei,anUJ is the boundary of the theoretical domain. The theoretical domains of both input and output variables are divided into 7 fuzzy subsets: NB (positive large), NM (positive medium), NS (positive small), ZO (zero), PS (negative small), PM (negative medium), PB (negative large), and determine the form of the affiliation function, then, the output variables are obtained by 3 processes of input fuzzification, fuzzy inference and defuzzification Δkp,Δki,Δkd The final control quantity is determined according to the Eq (2.2).
u(t)=(kp0+Δkp)e(t)+(ki0+Δki)∫t0e(t)+(kd0+Δkd)de(t)dt | (2.2) |
where kp0,ki0, and kd0 are the initial design value of the PID parameters, designed by the conventional PID controller parameters rectification method.Δkp,Δki,Δkd are the three outputs of the fuzzy controller, which can automatically adjust the values of the three PID control parameters according to the state of the controlled object.
When the size of the theoretical domain in a fuzzy PID controller is not chosen properly, it is more difficult to guarantee the control effect of the fuzzy controller, so variable-domain fuzzy PID controller[29] is born.
As is shown in Figure 3, the variable theoretical domain fuzzy PID adjusts the theoretical domain range of the input and output in the fuzzy controller online by introducing a scaling factor. According to e(t) and ec(t) calculate the scaling factor α(e(t)),α(ec(t)),β(e(t),ec(t)), where α(e(t)),α(ec(t)) are the scaling factors of the input variables e(t) and ec(t) and β(e(t),ec(t)) are the scaling factors of the three output variables Δkp,Δki,Δkd the common scaling factor. Then, the initial theoretical domains of the input and output variables are adjusted for scaling, taking the ith input as an example, the new theoretical domain obtained after the adjustment is [−α(xi(t))Ei,α(xi(t))Ei] The new domain is obtained after the adjustment. [30] designed a new scaling factor, as shown in Eq (2.3).
{α(e(t))=(|e(t)|E1)τ1+εα(ec(t))=(|ec(t)|E2)τ2+εβ(e(t),ec(t))=α(e(t))+α(ec(t))2+ε | (2.3) |
where ε is a sufficiently small positive number, the E1 and E2 are the initial domain boundaries of the input variables, respectively, and τi(i=1,2) is the scaling factor design parameter, and τi∈[0,1].The scaling factor should be stable. [30] verified the stability of the new scaling factor from five aspects: duality, zero avoidance, monotonicity, coordination and normality.
Reinforcement learning is a subfield of machine learning, as is shown in Figure 4, and intelligence robots learn a strategy for maximizing expected future rewards by interacting with its environment π(s)[31] which defines which action a should be taken in each state s. When an action is performed and the environment shifts to a new state s, the intelligence receives a reward r.
The reinforcement learning process can be modelled as a Markov decision process (MDP)[32], the Markov decision process is defined by the tuple (S,A,T,R,γ). S represents the state space, A represents the action space, T represents the state transfer function, R represents the reward function, and γ represents the discount factor. At each time step t, the maximise reward is
Rt=∞∑k=0γkrt+k | (2.4) |
Reinforcement learning algorithms are divided into policy-based algorithms and value-based algorithms. In value-based reinforcement learning, the value function is used to estimate the value of being in each state. The state value function is derived from the states given under the strategy π.
vπ(s)=Eπ[Rt∣St=s]=Eπ[∫∞k=0γkrt+k∣St=s] | (2.5) |
where Eπ denotes the expectation under strategy π. Similarly, the state action value function for taking action a in state s under strategy π. qπ can be defined as
qπ(s,a)=Eπ[Rt∣St=s,At=a]=Eπ[∫∞k=0γkrt+k∣St=s,At=a] | (2.6) |
Q-Learning[33] is an offline value-based algorithm with a value function based on state actions and an iterative formula that can be described as
Qt+1(st,at)←Qt(st,at)+α[rt+1+γmaxQt(st+1,a)−Qt(st,at)] | (2.7) |
where α∈(0,1] is the learning rate, and γ∈(0,1] is the discount factor describing the importance of weighting between immediate and future long-term rewards[34].
Traditional PID control has its own limitations in that the parameters cannot change with the external environment, adding fuzzy PID control allows for improved robustness. In order to increase the accuracy of fuzzy control, the theoretical domain can be changed by adding a scaling factor to optimise PID control, i.e., variable theoretical domain fuzzy PID control. The control function in a variable-domain fuzzy PID controller is self-adjusting through constant self-adjustment, making the PID parameters self-tuning. The control function is reduced or enlarged over time, making the updated control function less accurate and generating distortion. As is shown in Figure 5, the Q-Learning algorithm is added to make the scaling factor have the ability to find the optimum online, thus adjusting the PID parameters more accurately and seeking the optimum PID parameters for the current operating conditions.
This section analyses the key scaling factors in variable theoretical domain fuzzy PID control. Taking Eq (2.3) as an example, what determines the size of the theoretical domain are its two parameters τi (i=1,2) When these two parameters are changed, the domain is changed. By invoking Q-learning control of these two parameters, a better control effect can be achieved, making the controller capable of learning and online correction, and the system is more resistant to disturbances and more robust.
In the process of Q-learning, the change in parameters Δτi (i = 1, 2) as the action set, i.e.,A={−0.075,−0.05,−0.025,0,0.025,0.05,0.075}, and the state set is the deviation e. The Q matrix is built. The reward function is related to the rate of change of error. A negative rate of change of error indicates that this learning direction is the reward direction and can continue to be adjusted in this direction, a positive rate of change of error indicates that this learning direction is the penalty direction and should be adjusted in the opposite direction at this time.
r(t)={1ω|e(t)|+g, if ec (t)<01ω|e(t)|−g, if ec (t)>0 | (2.8) |
where r(t) is the immediate reward at moment t, e(t) is the deviation between the actual trajectory c(t) and the preset trajectory r(t), i.e.,e(t)=r(t)−c(t), and ec(t) is the rate of change of the error at moment t, i.e.,ec(t)=e(t)−e(t−1), denotes the direction of learning. g is the penalty term, and ω is the weight, and the iteration of the Q-value table is carried out through Eq (2.7).
Algorithm 1 Variable-domain fuzzy PID based on Q-Learning (Q-fuzzyPID) algorithm |
Require: ϵ,P,γ, the maximum simulation step N. |
Initialize: Q,S,τi,t. |
Loop |
If S=0 with probability ϵ select a random action to get out of the local |
minimum |
Else |
Select at=argmaxa′Q(si,a′;θt∣t−1) from the selection criterion |
of the current parameters τi |
End |
If t<N |
State Update the Q-value function Q(s,a) with updated action a. |
End |
t=t+1 |
Outputs: The parameters τ1,τ2. |
To verify the effectiveness of the proposed method, we trained a Q-Learning based variable-domain fuzzy PID controller and built urban road scenarios, including large curvature scenarios, overtaking scenarios, and following scenarios, in a Panosim simulation environment, as is shown in Figure 6.
Before stating the experimental results, the parameter values are presented. The stretching factor was designed according to Eq (2.3), initialized ε to 0.01, and adjusting the parameters of the scaling factor online using Q-Learning τi(i=1,2), τi∈ [0, 1], initialized τi to 0.5. Initialize the theoretical domain of E as [-350,350], Ec as [-100,100], Δkp as [-20, 20], Δki as [-5, 5] and Δkd as [-10, 10], and divide the input and output into 8 parts with affiliation values of NB,NM,NS,ZO,PS,PM,PB. The input E The corresponding intervals of affiliation values are [-350, -262.5], [-262.5, -175], [-175, -87.5], [-87.5, 0], [0, 87.5], [87.5,175], [175,262.5], [262.5,350]. According to Eq (2.3), the E1 is 350 and E2 is 100. When the error and the rate of change of the error are greater than the domain boundary, the value is taken as the boundary. The reward function is measured according to the real-time trajectory error, so initialize ω to 1 and initialized g to 20.
The whole experimental process is divided into training and testing, with training conducted in the following and overtaking scenarios, where the overtaking trajectories are generated from Bessel curves, and testing conducted in the cornering scenario. The proposed method is compared with the fuzzy PID and variable-domain fuzzy PID (vdfuzzyPID) using the error between the actual trajectory and the preset trajectory as an indicator.
As is shown in Figures 7–9, we comprise between fuzzy PID control, variable-domain fuzzy PID control and variable-domain fuzzy PID control based on Q-Learning. And the red line is the expected vehicle trajectory, the blue line is the vehicle trajectory under fuzzy PID control, the black line is the vehicle trajectory under variable-domain fuzzy PID control, and the green line is the vehicle trajectory under variable-domain fuzzy PID control based on Q-Learning. Figures 7(a)–9(a) shows the target trajectory and the actual trajectory of the three scenarios respectively, in which A indicates the starting point, B indicates the ending point. Figure 7(a) shows the control effect in the following scenario. Since the vehicle speed is slow and the corner change is small in the following scenario, the control effect is no different between the three methods. Figure 8(a) shows the control effect in the overtaking scenario, where there is a slow moving vehicle in front of the original trajectory obstructing the car from moving forward, so it make a left lane change decision, and it can be seen from the steering path that the steering is smoother than fuzzy PID under variable-domain fuzzy PID control based on Q-Learning. Figure 9(a) shows the control effect in the curvature scenario, where variable-domain fuzzy PID control based on Q-Learning results in a trajectory that is more closely aligned with the preset trajectory, and fuzzy PID results in a larger actual trajectory arc during cornering due to the slow response time and small change in steering angle.
As shown in Figures 7(b)–9(b), the average reward curve for each path showed a steady upward trend with no significant drop or fluctuation, which means that the training process was very stable.
The control error is used as a measure of the stability of the proposed method, and the error varies with the movement of the vehicle. As shown in Figure 10, it is a real time error and local zoom of the car following scene. As shown in Figure 10(a), the control errors of both methods fluctuate around 0. The enlarged area in Figure 10(b) corresponds to the curve in Figure 7(a). It can be seen from both the trajectory and the error diagram that the fuzzy PID control effect and variable-domain fuzzy PID effect are poor. Figure 10(c) the error change in the enlarged area is relatively dense. It can be seen from the enlarged view that Q-fuzzyPID control error is small.
As shown in Figure 11, it is real-time control trajectory error of overtaking and curvature. As shown in Figure 11(a), after lane changing and overtaking, the error between the actual control track and the preset track of the three methods is large, and the control effect of the proposed method will be better by comparison. Figure 11(b) is the control error of the three methods at the curvature curve. Due to the slow response time of fuzzy PID at the turning angle, the small change of steering angle leads to large error. Table 1 shows the average error of the three methods in each of the three scenarios. From the data in the table, it can be seen that the average error of the proposed method in all three scenarios is smaller than the fuzzy PID control error and variable-domain fuzzy PID control error.
Scene | Method | Average Error |
Follow | Q-Fuzzy PID | 9.41 |
Vdfuzzy PID | 13.25 | |
Fuzzy PID | 15.33 | |
Overtaking | Q-Fuzzy PID | 244.16 |
Vdfuzzy PID | 250.64 | |
Fuzzy PID | 262.04 | |
Curvature | Q-Fuzzy PID | 15.31 |
Vdfuzzy PID | 35.76 | |
Fuzzy PID | 117.73 |
Self-driving vehicles require different control accuracies when faced with different complex scenarios. Traditional PID control uses a set of parameters that are difficult to adapt to changes in the scenario. The control function in a variational domain fuzzy PID controller will make the PID parameters self-tuning by constantly adjusting itself. However, the control function is reduced or enlarged with time, which can make the updated control function no longer accurate and generating distortion. In this paper, the Q-Learning algorithm is added to make the scaling factor have the ability to find the best online, so that the PID parameters can be adjusted more accurately and the optimal PID parameters can be found under the current working conditions. The effectiveness of the method is also verified on the Panosim simulation platform.
This work was supported by Beijing Natural Science Foundation (No.4222025), the National Natural Science Foundation of China (Nos. 61871038 and 61931012), Beijing Municipal Science & Technology Commission, Administrative Commission of Zhongguancun Science Park.
The authors declare there are no conflict of interest.
[1] |
R. K. Khadanga, A. Kumar, S. Panda, Frequency control in hybrid distributed power systems via type-2 fuzzy pid controller, IET Renewable Power Gener., 15 (2021), 1706–1723. https://doi.org/10.1049/rpg2.12140 doi: 10.1049/rpg2.12140
![]() |
[2] | M. K. Diab, H. H. Ammar, R. E. Shalaby, Self-driving car lane-keeping assist using pid and pure pursuit control, in 2020 International Conference on Innovation and Intelligence for Informatics, Computing and Technologies (3ICT), IEEE, (2020), 1–6. https://doi.org/10.1109/3ICT51146.2020.9311987 |
[3] |
H. Maghfiroh, M. Ahmad, A. Ramelan, F. Adriyanto, Fuzzy-pid in bldc motor speed control using matlab/simulink, J. Rob. Control (JRC), 3 (2022), 8–13. https://doi.org/10.18196/jrc.v3i1.10964 doi: 10.18196/jrc.v3i1.10964
![]() |
[4] |
J. R. Nayak, B. Shaw, B. K. Sahu, K. A. Naidu, Application of optimized adaptive crow search algorithm based two degree of freedom optimal fuzzy pid controller for agc system, Eng. Sci. Technol. Int. J., 32 (2022), 101061. https://doi.org/10.1016/j.jestch.2021.09.007 doi: 10.1016/j.jestch.2021.09.007
![]() |
[5] |
N. Ma, D. Li, W. He, Y. Deng, J. Li, Y. Gao, et al., Future vehicles: interactive wheeled robots, Sci. China Inf. Sci., 64 (2021), 1–3. https://doi.org/10.1007/s11432-020-3171-4 doi: 10.1007/s11432-020-3171-4
![]() |
[6] | N. Ma, Y. Gao, J. Li, D. Li, Interactive cognition in self-driving, Chin. Sci.: Inf. Sci., 48 (2018), 1083–1096. |
[7] |
D. Li, N. Ma, Y. Gao, Future vehicles: learnable wheeled robots, Sci. China Inf. Sci., 63 (2020), 1–8. https://doi.org/10.1007/s11432-019-2787-2 doi: 10.1007/s11432-019-2787-2
![]() |
[8] |
T. Yang, N. Sun, Y. Fang, Adaptive fuzzy control for a class of mimo underactuated systems with plant uncertainties and actuator deadzones: Design and experiments, IEEE Trans. Cybern., 52 (2022), 8213–8226. https://doi.org/10.1109/TCYB.2021.3050475 doi: 10.1109/TCYB.2021.3050475
![]() |
[9] |
S. H. Park, K. W. Kim, W. H. Choi, M. S. Jie, Y. Kim, The autonomous performance improvement of mobile robot using type-2 fuzzy self-tuning PID controller, Adv. Sci. Technol. Lett., 138 (2016), 182–187. https://doi.org/10.14257/astl.2016.138.37 doi: 10.14257/astl.2016.138.37
![]() |
[10] |
P. Parikh, S. Sheth, R. Vasani, J. K. Gohil, Implementing fuzzy logic controller and pid controller to a dc encoder motor–-"a case of an automated guided vehicle", Procedia Manuf., 20 (2018), 219–226. https://doi.org/10.1016/j.promfg.2018.02.032 doi: 10.1016/j.promfg.2018.02.032
![]() |
[11] |
Q. Bu, J. Cai, Y. Liu, M. Cao, L. Dong, R. Ruan, et al., The effect of fuzzy pid temperature control on thermal behavior analysis and kinetics study of biomass microwave pyrolysis, J. Anal. Appl. Pyrolysis, 158 (2021), 105176. https://doi.org/10.1016/j.jaap.2021.105176 doi: 10.1016/j.jaap.2021.105176
![]() |
[12] | M. S. Jie, W. H. Choi, Type-2 fuzzy pid controller design for mobile robot, Int. J. Control Autom., 9 (2016), 203–214. |
[13] | N. Kumar, M. Takács, Z. Vámossy, Robot navigation in unknown environment using fuzzy logic, in 2017 IEEE 15th International Symposium on Applied Machine Intelligence and Informatics (SAMI), IEEE, (2017), 279–284. https://doi.org/10.1109/SAMI.2017.7880317 |
[14] | T. Muhammad, Y. Guo, Y. Wu, W. Yao, A. Zeeshan, Ccd camera-based ball balancer system with fuzzy pd control in varying light conditions, in 2019 IEEE 16th International Conference on Networking, Sensing and Control (ICNSC), IEEE, (2019), 305–310. https://doi.org/10.1109/ICNSC.2019.8743305 |
[15] |
A. Wong, T. Back, A. V. Kononova, A. Plaat, Deep multiagent reinforcement learning: Challenges and directions, Artif. Intell. Rev., 2022 (2022). https://doi.org/10.1007/s10462-022-10299-x doi: 10.1007/s10462-022-10299-x
![]() |
[16] |
Z. Cao, S. Xu, H. Peng, D. Yang, R. Zidek, Confidence-aware reinforcement learning for self-driving cars, IEEE Trans. Intell. Transp. Syst., 23 (2022), 7419–7430. https://doi.org/10.1109/TITS.2021.3069497 doi: 10.1109/TITS.2021.3069497
![]() |
[17] | T. Ribeiro, F. Gonçalves, I. Garcia, G. Lopes, A. F. Ribeiro, Q-learning for autonomous mobile robot obstacle avoidance, in 2019 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC), IEEE, (2019), 1–7. https://doi.org/10.1109/ICARSC.2019.8733621 |
[18] | S. Danthala, S. Rao, K. Mannepalli, D. Shilpa, Robotic manipulator control by using machine learning algorithms: A review, Int. J. Mech. Prod. Eng. Res. Dev., 8 (2018), 305–310. |
[19] |
X. Lei, Z. Zhang, P. Dong, Dynamic path planning of unknown environment based on deep reinforcement learning, J. Rob., 2018 (2018). https://doi.org/10.1155/2018/5781591 doi: 10.1155/2018/5781591
![]() |
[20] |
Y. Shan, B. Zheng, L. Chen, L. Chen, D. Chen, A reinforcement learning-based adaptive path tracking approach for autonomous driving, IEEE Trans. Veh. Technol., 69 (2020), 10581–10595. https://doi.org/10.1109/TVT.2020.3014628 doi: 10.1109/TVT.2020.3014628
![]() |
[21] | T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, et al., Continuous control with deep reinforcement learning, preprint, arXiv: 1509.02971. https://doi.org/10.48550/arXiv.1509.02971 |
[22] |
P. Ramanathan, K. K. Mangla, S. Satpathy, Smart controller for conical tank system using reinforcement learning algorithm, Measurement, 116 (2018), 422–428. https://doi.org/10.1016/j.measurement.2017.11.007 doi: 10.1016/j.measurement.2017.11.007
![]() |
[23] |
L. Brunke, M. Greeff, A. W. Hall, Z. Yuan, S. Zhou, J. Panerati, et al., Safe learning in robotics: From learning-based control to safe reinforcement learning, Annu. Rev. Control Rob. Auton. Syst., 5 (2022), 411–444. https://doi.org/10.1146/annurev-control-042920-020211 doi: 10.1146/annurev-control-042920-020211
![]() |
[24] | A. I. Lakhani, M. A. Chowdhury, Q. Lu, Stability-preserving automatic tuning of PID control with reinforcement learning, preprint, arXiv: 2112.15187. https://doi.org/10.20517/ces.2021.15 |
[25] |
O. Dogru, K. Velswamy, F. Ibrahim, Y. Wu, A. S. Sundaramoorthy, B. Huang, et al., Reinforcement learning approach to autonomous pid tuning, Comput. Chem. Eng., 161 (2022), 107760. https://doi.org/10.1016/j.compchemeng.2022.107760 doi: 10.1016/j.compchemeng.2022.107760
![]() |
[26] |
X. Yu, Y. Fan, S. Xu, L. Ou, A self-adaptive sac-pid control approach based on reinforcement learning for mobile robots, Int. J. Robust Nonlinear Control, 32 (2022), 9625–9643. https://doi.org/10.1002/rnc.5662 doi: 10.1002/rnc.5662
![]() |
[27] |
B. Guo, Z. Zhuang, J. S. Pan, S. C. Chu, Optimal design and simulation for pid controller using fractional-order fish migration optimization algorithm, IEEE Access, 9 (2021), 8808–8819. https://doi.org/10.1109/ACCESS.2021.3049421 doi: 10.1109/ACCESS.2021.3049421
![]() |
[28] | M. Praharaj, D. Sain, B. Mohan, Development, experimental validation, and comparison of interval type-2 mamdani fuzzy pid controllers with different footprints of uncertainty, Inf. Sci., 601 (2022), 374–402. |
[29] |
Y. Jia, R. Zhang, X. Lv, T. Zhang, Z. Fan, Research on temperature control of fuel-cell cooling system based on variable domain fuzzy pid, Processes, 10 (2022), 534. https://doi.org/10.3390/pr10030534 doi: 10.3390/pr10030534
![]() |
[30] |
J. Wei, L. Gang, W. Tao, G. Kai, Variable universe fuzzy pid control based on adaptive contracting-expanding factors, Eng. Mech., 38 (2021), 23–32. https://doi.org/10.6052/j.issn.1000-4750.2020.11.0786 doi: 10.6052/j.issn.1000-4750.2020.11.0786
![]() |
[31] | R. S. Sutton, A. G. Barto, Reinforcement Learning: An Introduction, MIT press, 2018. |
[32] |
P. R. Montague, Reinforcement learning: an introduction, by Sutton, RS and Barto, AG, Trends Cognit. Sci., 3 (1999), 360. https://doi.org/10.1016/S1364-6613(99)01331-5 doi: 10.1016/S1364-6613(99)01331-5
![]() |
[33] | D. Wang, R. Walters, X. Zhu, R. Platt, Equivariant q learning in spatial action spaces, in Conference on Robot Learning, PMLR, (2022), 1713–1723. |
[34] |
E. Anderlini, D. I. Forehand, P. Stansell, Q. Xiao, M. Abusara, Control of a point absorber using reinforcement learning, IEEE Trans. Sustainable Energy, 7 (2016), 1681–1690. https://doi.org/10.1109/TSTE.2016.2568754 doi: 10.1109/TSTE.2016.2568754
![]() |
1. | Jialu Ma, Pingping Zhang, Yixian Li, Yuhang Gao, Jiandong Zhao, Peng Hang, Longitudinal Hierarchical Control of Autonomous Vehicle Based on Deep Reinforcement Learning and PID Algorithm, 2024, 2024, 0197-6729, 10.1155/2024/2179275 | |
2. | Yu Jiang, Zeyu Sun, Ruochen Wang, Renkai Ding, Qing Ye, Design and control of a new omnidirectional levelling system for hilly crawler work machines, 2024, 218, 01681699, 108661, 10.1016/j.compag.2024.108661 | |
3. | Xiaoxiao Yao, Zhiyu Shao, 2023, Underwater Thrust Vector Control System Based on DSP28335, 979-8-3503-0937-9, 949, 10.1109/IFEEA60725.2023.10429115 | |
4. | Jin Zhang, Nan Ma, Zhixuan Wu, Cheng Wang, Yongqiang Yao, Intelligent control of self-driving vehicles based on adaptive sampling supervised actor-critic and human driving experience, 2024, 21, 1551-0018, 6077, 10.3934/mbe.2024267 | |
5. | Ibrahim Altarjami, Yassir Alhazmi, Studying the Optimal Frequency Control Condition for Electric Vehicle Fast Charging Stations as a Dynamic Load Using Reinforcement Learning Algorithms in Different Photovoltaic Penetration Levels, 2024, 17, 1996-1073, 2593, 10.3390/en17112593 | |
6. | Kun Feng, Xu Li, Wenli Li, Adaptive MPC path-tracking controller based on reinforcement learning and preview-based PID controller, 2024, 0954-4070, 10.1177/09544070241287965 | |
7. | Caichang Ding, Chao Li, Zenggang Xiong, Zhimin Li, Qiyang Liang, Intelligent Identification of Moving Trajectory of Autonomous Vehicle Based on Friction Nano-Generator, 2024, 25, 1524-9050, 3090, 10.1109/TITS.2023.3303267 | |
8. | Yanjuan Hu, Qingling Liu, You Zhou, Changhua Yin, Tracking performance optimization of balancing machine turntable servo system based on deep deterministic policy gradient fractional order proportional integral derivative control strategy, 2025, 242, 02632241, 116256, 10.1016/j.measurement.2024.116256 | |
9. | Guobin Si, Ruijie Zhang, Xiaofeng Jin, Path Planning of Factory Handling Robot Integrating Fuzzy Logic-PID Control Technology, 2025, 27729419, 200188, 10.1016/j.sasc.2025.200188 | |
10. | Rajendra Machavaram, Intelligent path planning for autonomous ground vehicles in dynamic environments utilizing adaptive Neuro-Fuzzy control, 2025, 144, 09521976, 110119, 10.1016/j.engappai.2025.110119 | |
11. | Li Wang, Yijiu Zhao, Yifan Wang, Haitao Zhou, A High-Accuracy Decoupling Balance Control Method for an Auto-Balancing Bridge Based on a Variable-Domain Fuzzy-PID Controller, 2025, 17, 2073-8994, 354, 10.3390/sym17030354 | |
12. | Qiong Wu, Hua Chen, Baolong Liu, Intelligent Robot Trajectory Tracking Control Combining Mamdani Theory and Q-Learning Algorithm, 2025, 13, 2169-3536, 91276, 10.1109/ACCESS.2025.3572461 |
Scene | Method | Average Error |
Follow | Q-Fuzzy PID | 9.41 |
Vdfuzzy PID | 13.25 | |
Fuzzy PID | 15.33 | |
Overtaking | Q-Fuzzy PID | 244.16 |
Vdfuzzy PID | 250.64 | |
Fuzzy PID | 262.04 | |
Curvature | Q-Fuzzy PID | 15.31 |
Vdfuzzy PID | 35.76 | |
Fuzzy PID | 117.73 |