A Neural Network-Based Navigation Approach
A Neural Network-Based Navigation Approach
sciences
Article
A Neural Network-Based Navigation Approach for
Autonomous Mobile Robot Systems
Yiyang Chen 1 , Chuanxin Cheng 1 , Yueyuan Zhang 1, *, Xinlin Li 2, * and Lining Sun 1
1 School of Mechanical and Electrical Engineering, Soochow University, Suzhou 215031, China;
[email protected] (Y.C.); [email protected] (C.C.); [email protected] (L.S.)
2 Department of Digital Media, Soochow University, Suzhou 215031, China
* Correspondence: [email protected] (Y.Z.); [email protected] (X.L.)
Abstract: A mobile robot is a futuristic technology that is changing the industry of automobiles as
well as boosting the operations of on-demand services and applications. The navigation capability
of mobile robots is a crucial task and one of the complex processes that guarantees moving from a
starting position to a destination. To prevent any potential incidents or accidents, navigation must
focus on the obstacle avoidance issue. This paper considers the navigation scenario of a mobile
robot with a finite number of motion types without global environmental information. In addition,
appropriate human decisions on motion types were collected in situations involving various obstacle
features, and the corresponding environmental information was also recorded with the human
decisions to establish a database. Further, an algorithm is proposed to train a neural network model
via supervising learning using the collected data to replicate the human decision-making process
under the same navigation scenario. The performance of the neural network-based decision-making
method was cross-validated using both training and testing data to show an accuracy level close to
90%. In addition, the trained neural network model was installed on a virtual mobile robot within a
Citation: Chen, Y.; Cheng, C.; Zhang,
mobile robot navigation simulator to interact with the environment and to make the decisions, and
Y.; Li, X.; Sun, L. A Neural
the results showed the effectiveness and efficacy of the proposed algorithm.
Network-Based Navigation
Approach for Autonomous Mobile
Keywords: neural network; mobile robot navigation; obstacle avoidance
Robot Systems. Appl. Sci. 2022, 12,
7796. https://doi.org/10.3390/
app12157796
in the same route, which is not efficient. Although the work in [10–12] has attempted to
overcome these concerns, the proposed algorithms were adhoc, and their applications were
restricted to the specific navigation tasks with convex shaped static obstacles or global
environmental information. An artificial potential field is considered an efficient method
for motion planning, but it runs the risk of falling into a local optimum.
This paper intends to deal with the class of navigation problems using obtained human
decision data from the aspect of artificial intelligence. As pointed out in [13], artificial
intelligence is considered as a technique aiming at replicating human behavior by making
inferences according to historical learned data. It usually establishes a neural network
model and tunes the parameters within this model using the obtained realistic data, so that
the trained model can provide useful information for the users [14]. As explained in [15],
this technique is particularly strong at making intelligent decisions for certain design
problems with variability, uncertainty, complexity, and fuzziness [16]. This advantage
enables its applications to many research fields, such as image recognition [17–19], data
analysis [20], system identification [21–23], and mechanical fault diagnosis [24]. Note that
most of the applications mentioned above deal with data using artificial intelligence, whose
structures and labels are both sparse. However, the data labels of the navigation problem
are not purely sparse, since one specific occasion can have multiple potential labels.
To solve this specific class of navigation problems using artificial intelligence, some
work has already been completed. The work in [25] employed the Q-learning method to
interact with a dynamic environment and limited the state number of its state space to
fulfill a navigation algorithm. In addition, it was reported in [26] that neural network-
based reinforcement learning was used to handle the navigation tasks with both static
and dynamic obstacles in an uncertain workspace. Deep reinforcement learning to solve
navigation problems is increasingly favored by researchers and has progressed since
2016 [27]. The exploration process is decomposed into decision-making, planning, and
mapping modules to increase the modularity of the robot system. In each framework, a
decision algorithm based on variant deep Q-learning (DQN) was proposed to learn and
explore strategies from local maps in [28]. An actor–critic model was trained to learn
navigation policy, and expert knowledge was utilized to generate subgoals in [29]. In [30],
a heuristic dynamic programming method was proposed to demonstrate a Markov decision
process during a maze navigation task, which included a goal network to model the reward.
The research in [31] developed an obstacle avoidance method using a fuzzy neural network
for wheeled mobile robots, which reduced the oscillation during obstacle avoidance and
achieved a smooth path. In [32], symbolic spatial information was employed to assist
robots to move in an unseen scene, which achieved good performance. Note that the
aforementioned methods all incorporated a relatively heavy real-time computation load
as a key component in their design framework, and it was difficult to set new plans when
receiving instantaneously a large amount of information. In [33], artificial and natural
landmark recognition-based simultaneous localization and mapping (SLAM) provided
absolute position feedback in real time to complete the task of robot localization and
modeling. For motion planning, a novel quadrupole potential field (QPF) approach was
used to realize collision-free path planning. However, the modeling environment based on
SLAM was time-consuming and had no ability to cope with real-time changes in scenarios,
which increased the difficulty of navigation without global environmental information.
To overcome the real-time computation issue in navigation problems, this paper
formulates a mobile robot navigation problem within a 2D plane with various shapes of
obstacles. It divides the motion behaviors of a given autonomous mobile robot system
into a finite number of motion types, and collects the environmental information (system
inputs) as well as appropriate decisions (labels) of these motion types made by human users
at various scenarios with random generated obstacles. These collected data are further
exploited to train a neural network model that is employed to replicate the decision-making
behavior of a human being. The trained model only responds to the real-time detected
environmental information and does not have to perform any optimization procedures. To
Appl. Sci. 2022, 12, 7796 3 of 17
determine the feasibility of this neural network based method, a simulator was established
in Python to present an exemplary mobile robot navigation task. The tuning procedures
using supervised learning were made on the structure of the neural network and the
parameters of the training algorithm to guarantee the high accuracy of this method with
respect to the training and testing data sets. To validate its performance, the trained model
was applied to the control system of the mobile robots in the simulator multiple times.
The resulting paths confirmed its ability to handle navigation tasks and achieve obstacle
avoidance as well. The scientific and technical contributions of this paper are highlighted
as follows:
• A autonomous mobile robot navigation problem was formulated with a finite number
of motion types, requirements on obstacle avoidance in 2D space, and no global
environmental information.
• An implementation algorithm is provided on how to process the obtained data and
train a neural network model using a supervised learning method to approach the
human decision-making process.
• A simulator was established in the Python programming environment to model
each step of an exemplary mobile robot navigation task with visualization outcomes,
and the human decision data were collected under the same navigation scenario of
the simulator.
• The proposed algorithm used the data collected from the human decision-making
process under the same navigation scenario to train a neural network model to achieve
high estimation accuracy rates.
• The trained neural network model was validated in the simulator to perform the required
task with the desired performance, which solves the real-time computation issue.
The nomenclature used in this paper is listed in Abbreviations part.
2. Problem Formulation
In this section, an autonomous mobile robot navigation task is first described, and then its
design objectives are given. Together, these form a navigation problem with obstacle avoidance.
at each sample decision-making time index t is considered in this paper, where x (t) and
y(t) are the horizontal and vertical coordinates of the mobile robot, and θ (t) is the heading
direction of the mobile robot. The motion of this vehicle is conducted by a finite number
of motion types Mi , i = 1, . . . , n. Here, each motion type should be linked to an imple-
mentable action of the mobile robot, e.g., go ahead, turn left or turn right, and associate
with a particular state change function
Since the motion function is simplified to motion types, it enables this task to be carried
out by sequentially making appropriate decisions of motion types at each sample time
index t. To help the achievement of the latter design, two assumptions are made as below.
Assumption 1. The motion types of this mobile robot guarantee it to reach any feasible position
within the given 2D plane.
Assumption 2. The motion types must differ significantly from each other, i.e., for a given state s,
there exists
Fi (s) − F j (s) > σ, for any i, j, (3)
where σ is a nonnegative scalar.
Assumption 1 makes sense in practice, since it guarantees the mobile robot task
achievement, as there is no need to design any autonomous mobile robot navigation
approach for an impossible task. Assumption 2 ensures all motion types will be different,
which increases the success rate of classifying the appropriate motion type using a neural
network in Section 3.
Figure 2. A visual example of the detected local environmental information of the mobile robot.
Note that the green lines in the figure are the laser beams, which will reflect after
touching an obstacle. These laser beams can detect the obstacles within a maximum distance
of dmax ; hence, there exists 0 6 di (t) 6 dmax . Furthermore, a location sensor is incorporated
to calculate the bias of the heading direction, which is the red arrow in the figure, and is
denoted as φ(t) satisfying −180◦ < φ(t) 6 180◦ .
In this sense, the design objective of the autonomous mobile robot navigation task is
to develop a mechanism, which makes the vehicle respond to the real time environmental
information in order to complete the task. The navigation problem discussed in this paper
is explained in the next definition.
Definition 1. The autonomous mobile robot navigation problem with obstacle avoidance
aims to make appropriate decisions of the given motion types based on the real-time detected
information di (t), i = 1, . . . , m and φ(t) at each sample time index t, such that the vehicle moves
from its start position ( x0 , y0 ) to a position ( x ∗ , y∗ ) close to its destination position ( x d , yd )
without colliding with any single obstacle, i.e.,
( x ∗ , y∗ ) − ( x d , yd ) 6 e, (4)
where e is a small positive scalar determining an acceptable distance from the destination to complete
the navigation task.
j
where j is the occasion index number, di , i = 1, . . . , m, are the obstacle distances detected by
the laser distance sensors, φ j is the bias of the heading direction measured by the location
sensor, and l j is the label of human made decision Mi , i.e.,
To ensure the collected data were meaningful, an assumption was made as follows:
Assumption 3. The human users tried their best to achieve the given navigation task, and they
were capable of successively completing this task, i.e., they did not aim to make the mobile robot
collide with an obstacle or move toward the wrong destination.
is minimized for all N labeled datasets, where ẑi ∈ Rn is the expected output defined by l i as
(
1, j = l i ,
ẑi ( j) = (8)
0, else.
The cross-entropy loss function (7) was utilized to evaluate the loss between the output
of the neural network and the human behavior label. The weights of the neural network
were continuously adjusted to reduce the loss value through the error back-propagation
algorithm. When the cross entropy loss became small enough, the neural network model
was considered to completely fit the mapping relationship between input and output.
Appl. Sci. 2022, 12, 7796 7 of 17
Since the data values may vary in rather different ranges, a rescaling step was per-
formed at the input layer such that the output of the input layer was computed as the vector
This rescaling step adjusted the input values to a relatively similar level, overcoming
the potential slow convergence problem, which would occur at the large-scale dimension.
As shown in Figure 3, the model included k hidden layers in the dashed red box, and
the i-th hidden layer had ri units. The output of the i-th layer was a vector ziout ∈ Rri and
was transferred to the input zin
i +1 ∈ R
ri+1 of the (i + 1)-th layer as
zin out
i +1 = Wi +1 zi , (11)
where Wi+1 ∈ Rri+1 ×ri is a weight matrix. An activation function f i,j (·) was applied before
the network output of each layer, which is defined as
zout in
k +1 = g ( z k +1 ), (14)
Appl. Sci. 2022, 12, 7796 8 of 17
e x (i )
g ( x )i = . (15)
∑nj=1 e x( j)
To evaluate the training effect of the neural network, the output accuracy of the
training network on Λtrain and Λtest were computed as
where λi is a mark used to determine whether the character corresponding to the maximum
value of neural network output was consistent with the label as
(
acc 1, arg max zi = l i ,
λi = (17)
0, else.
Remark 1. Since the elements in testing set Λtest and the training set Λtrain are completely
independent of each other, the evaluation results are able to examine the training performance in
terms of overfitting problems.
with random obstacles as that shown in Figure 1. It enables the vehicle to perform three
motion types, such as go straight ahead, rotate to the left, and rotate to the right, i.e., m = 3,
and the state change functions are defined as
where vm is the motion speed, and vr is the rotation speed. Note that the values of vm and
vr can be customized in the configuration window of the simulator.
This simulator can generate various occasions with random obstacles and collect the
decisions made by the human test candidates, and an exemplary situation is shown in
Figure 4. As can be seen from this figure, there were 17 laser detectors, i.e., m = 17, installed
on the front half plane of the mobile robot. This simulator replicated a radar with 5 m
measuring distance, and each pixel in the screen denoted 0.01 m. Therefore, the maximum
detection distance was set as dmax = 500 in units of pixels. Since the laser detectors can
only measure the information in the front 180◦ direction, the bottom half of the circular
sight in this figure was covered when recording the human user’s decisions to provide a
fair navigation environment. The color ci of the i-th laser beam followed the mechanism
Green, 0.6dmax < di 6 dmax ,
ci = Yellow, 0.3dmax < di 6 0.6dmax , (19)
Red, 0 < di 6 0.3dmax ,
which alerted the human test candidates to potential nearby obstacles. Note that only a
circular sight is shown in Figure 4, and it represents the exact region the mobile robot could
detect using its laser sensors. The obtained decisions were recorded as the label l i , using
the reference equation
0, Go ahead,
li = 1, Rotate to left, (20)
2, Rotate to right.
The obtained labels were saved in the database together with the corresponding environ-
mental information di , i = 1, . . . , 17 and φi , which formed a complete labeled element.
Appl. Sci. 2022, 12, 7796 10 of 17
Remark 2. The generated occasions only provide the human test candidates with the exact in-
formation that the mobile robot can have. Therefore, with both having the same scope of obtained
environmental information, a neural network model can be trained to replicate the human decision-
making process in a fair manner.
Remark 3. This simulator is capable of accurately modeling a navigation task in a 2D plane with
random obstacles, and each human labeling process can be completed within 1 s. However, while
using the automated algorithms, e.g., reinforcement learning, on a similar experimental established
test platform in practice, it took time to build a limited range of real navigation scenarios. The
mobile robot should not move too fast, i.e., less than 0.2 m/s, to avoid safety issues, which means
they might take more than 10 s to obtain useful data. Therefore, the human labeling approach using
this simulator has the advantages of data collection speed and data diversity.
Remark 4. Note that the algorithm performance was affected by the number m of laser sensors, as
the knowledge of the surrounding environmental information definitely increases with more sensors.
In this sense, more detailed information helps the neural network to make more appropriate decisions.
However, the implementation space of sensors is restricted on a remote robot, which means the
number m cannot be too large.
respectively. Theoretically, deepening the neural network improves the training effect, but
it consumes more computing resources and increases the risk of overfitting. As shown
in Figure 5a, net2 and net3 deepened the number of layers and increased the number of
neurons in a certain layer on the basis of net1, but their accuracy in the test set was not
improved or even reduced, and the time required for training increased. Therefore, it was
considered appropriate to set the neural network to 5 layers. The learning rate determines
the convergence speed of neural network, which has a great impact on the training of
neural network. Too small a learning rate leads to the slow convergence speed of the neural
network and increased training time while too large a learning rate results in missing the
optimal solution. As shown in Figure 5b, setting the learning rate to 0.0008 was considered
to be a suitable scheme that accounted for both the training speed and the optimal solution.
Furthermore, the batch size and the total number of iterations were selected as
which affected the total time spent in the whole training process. During the data collection
phase, 800 labeled elements were collected into data set Λ, and 640 were randomly extracted
to generate the training set Λtrain , i.e., η = 640.
Appl. Sci. 2022, 12, 7796 11 of 17
(a) (b)
Figure 5. Different layers, neurons, and learning rates affect the accuracy of the neural network
replicating human decision-making process. (a) neural network; (b) learning rate.
100
90
Estimation Accuracy Rate (%)
80
70
60
50
40
0 200 400 600 800 1000
Figure 6. The estimation accuracy rates of Algorithm 1 with respect to Λtrain and Λtest along the
training epochs.
To determine why the accuracy rate could not further increase, the dimensionality
reduction method called t-Distributed Stochastic Neighbor Embedding (t-SNE) was applied
to convert the high dimensional elements λi in the database Λ into a 2D plane for a
straightforward point of view. The results are shown in Figure 7, and it is clear that the data
points in some particular regions had more than one label. Therefore, the data structure
was not sparse enough to allow the trained model to reach a high estimation accuracy, and
it was difficult fo the model to identify the data type in those regions. This is due to the fact
that there inevitably exist some occasions in which the human test candidates cannot make
Appl. Sci. 2022, 12, 7796 12 of 17
(a) (b)
(c) (d)
Figure 8. The paths of the mobile robot while performing navigation tasks using the trained neural
network model in the simulator. (a) Case 1; (b) Case 2; (c) Case 3; (d) Case 4.
The paths of the mobile robot in some exemplary cases are shown in Figure 8. The
results demonstrate the appealing properties of the neural network based navigation
method proposed in this paper, in that the mobile robot could arrive at the destination
without colliding with any obstacles. The navigation performance in other cases was similar
to those in Figure 8. Note that all the robot knows at a certain position is sensor information,
so that it has no ability to make the optimal decision as in other global planning methods.
However, compared with other approaches, the proposed method did not consume high
computing resources to obtain the global environment information and was capable of
responding to dynamic changes in time. As shown in Figure 8a,c,d, although the path is
not the shortest, the robot safely moved from the starting point to the destination position
through its own adjustment ability.
The neural network was employed to replicate human behavior, but the planned paths
shown in Figure 8 seemed not to conform to human thinking in theory. This is mainly due
to two reasons: on the one hand, people usually combine the entire scene map to judge
the integral path from a global perspective. However, when the information available to
humans is reduced to a local area, the decision made is similar to the neural network. On
the other hand, humans utilize the brain’s memory to assist in decision making, which
is unavailable to the neural network proposed in this paper. This limitation may cause
the robot to spin around in a certain area. Note that the spinning of the robot as shown in
Figure 8c is a local optimum caused by the proposed method. However, since the choice of
action conforms to the probability distribution output by the neural network, the robot may
choose another action at the same location, which helps it out of the current predicament.
5.4. Discussion
To show the effectiveness of the proposed algorithm in autonomous mobile robot
navigation, it was compared with the artificial potential field (APF) method commonly
used in a 2D environment, which usually forms a shorter path. As shown in Figure 9a,b,
Appl. Sci. 2022, 12, 7796 14 of 17
the path that is formed by Algorithm 1 proposed in this paper was only slightly longer
than the path obtained by the artificial potential field method in the second half without
knowing the global environment information, but the generated path was smoother than
that of the APF method. Furthermore, the APF method has the probability to fall into
the local minimum solution as shown in Figure 9c, while the method based on the neural
network is capable of bypassing obstacles of various shapes ahead, as shown in Figure 9d.
(a) (b)
(c) (d)
Figure 9. The paths of the mobile robot while performing navigation tasks using the trained neural
network model in the simulator. (a) APF; (b) proposed; (c) APF; (d) proposed.
Meanwhile, the dynamic window approach (DWA) is a classic local path planning
algorithm. When the robot reaches a certain position, DWA predicts the trajectory of each
set of speeds in a limited dynamic window and gives a speed evaluation, and the set of
speeds with the highest evaluation is selected as the current speed for execution. Finally,
the path planning is transformed into the problem of maximizing the objective function,
but real-time computing consumes high computing resources, around 100 times more than
the proposed algorithm. Compared to the DWA, the advantage of the method proposed
in this paper is that the computation required for neural network training is offloaded to
the computer to reduce the requirement on the robotic hardware. When the network is
transplanted to the robot, it only needs to perform a few matrix operations to obtain the
corresponding action.
human decision process using recorded data. To validate the performance of the proposed
algorithm and the trained model, a mobile robot simulator was developed to collect the
data from the human decision process as well as perform a representative navigation task in
a 2D plane. The simulation results illustrated that a high estimation accuracy rate of nearly
90% with respect to the human decision-making process was obtained using the proposed
algorithm, and the trained model successfully completed the given tasks. Compared with
the APF method, the neural network approach was able to handle the navigation task with
obstacles without the local minimum problem, which proves the effectiveness of neural
network decision.
For future work, the trained model will be applied to an experimental test platform
to further examine its practical performance, and the iterative learning control method
in [35–37] can be implemented to guarantee the practical performance of the repetitive
motion types. It is undeniable that the neural network only considers the current state of the
robot for decision making, which leads to the certainty of global judgment. A navigation
algorithm with a short memory will be considered to be applied to further improve its
performance to handle the circular motion problem as shown in Figure 8c. Meanwhile,
reinforcement learning methods can be employed to let the robot train itself by gaining a
reward when approaching the destination or receiving punishment when colliding with an
obstacle. Last but not least, the proposed method can be extended to a distributed version
to handle the navigation task consisting of a number of mobile robots.
Author Contributions: Conceptualization, Y.C., Y.Z. and X.L.; methodology, Y.C. and C.C.; software,
Y.C. and C.C.; validation, Y.C. and C.C.; formal analysis, Y.C., Y.Z. and X.L.; investigation, C.C.;
resources, Y.C., Y.Z. and X.L.; data curation, Y.C.; writing—original draft preparation, Y.C.; writing—
review and editing, Y.Z. and X.L.; visualization, C.C.; supervision, L.S.; project administration, Y.C.;
funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.
Funding: This research was funded by the National Natural Science Foundation of China grant
number 62103293, the Natural Science Foundation of Jiangsu Province grant number BK20210709, the
Suzhou Municipal Science and Technology Bureau grant number SYG202138, and the Entrepreneur-
ship and Innovation Plan of Jiangsu Province grant number JSSCBS20210641.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Conflicts of Interest: The authors declare no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
Abbreviations Description
CUDA Compute unified device architecture
GPU Graphics processing unit
t-SNE t-Distributed Stochastic Neighbor Embedding
Symbols Description
Rn is an n-dimensional Euclidean vector space
Rm × n is an m × n real matrix space
Mi is the i-th motion type of a mobile robot
Fi (·) is the i-th state change function of a mobile robot
L(·) is the cross entropy loss function
t is the sample decision-making time index
Λtrain is the training data set
Λtest is the testing data set
Appl. Sci. 2022, 12, 7796 16 of 17
References
1. Gong, Y.; Zhang, L.; Liu, R.P.; Yu, K.; Srivastava, G. Nonlinear MIMO for Industrial Internet of Things in Cyber-Physical Systems.
IEEE Trans. Ind. Inform. 2021, 17, 5533–5541. [CrossRef]
2. Schulman, J.; Duan, Y.; Ho, J.; Lee, A.; Awwal, I.; Bradlow, H.; Pan, J.; Patil, S.; Goldberg, K.; Abbeel, P. Motion planning with
sequential convex optimization and convex collision checking. Int. J. Robot. Res. 2014, 33, 1251–1270. [CrossRef]
3. Rezaee, H.; Abdollahi, F. A Decentralized Cooperative Control Scheme with Obstacle Avoidance for a Team of Mobile Robots.
IEEE Trans. Ind. Electron. 2014, 61, 347–354. [CrossRef]
4. Ji, J.; Khajepour, A.; Melek, W.W.; Huang, Y. Path Planning and Tracking for Vehicle Collision Avoidance Based on Model
Predictive Control With Multiconstraints. IEEE Trans. Veh. Technol. 2017, 66, 952–964. [CrossRef]
5. Cheng, C.; Chen, Y. A neural network based mobile robot navigation approach using reinforcement learning parameter tuning
mechanism. In Proceedings of the IEEE 2021 China Automation Congress (CAC), Beijing, China, 22–24 October 2021.
6. Goerzen, C.; Kong, Z.; Mettler, B. A Survey of Motion Planning Algorithms from the Perspective of Autonomous UAV Guidance.
J. Intell. Robot. Syst. 2010, 57, 65–100. [CrossRef]
7. Campbell, S.; Naeem, W.; Irwin, G.W. A review on improving the autonomy of unmanned surface vehicles through intelligent
collision avoidance manoeuvres. Annu. Rev. Control 2012, 36, 267–283. [CrossRef]
8. Galceran, E.; Carreras, M. A survey on coverage path planning for robotics. Robot. Auton. Syst. 2013, 61, 1258–1276. [CrossRef]
9. Meng, D.; Xiao, Y.; Guo, Z.; Jolfaei, A.; Qin, L.; Lu, X.; Xiang, Q. A Data-driven Intelligent Planning Model for UAVs Routing
Networks in Mobile Internet of Things. Comput. Commun. 2021, 179, 231–241. [CrossRef]
10. Khansari-Zadeh, S.M.; Billard, A. A dynamical system approach to realtime obstacle avoidance. Auton. Robot. 2012, 32, 433–454.
[CrossRef]
11. Chu, K.; Lee, M.; Sunwoo, M. Local Path Planning for Off-Road Autonomous Driving with Avoidance of Static Obstacles.
IEEE Trans. Intell. Transp. Syst. 2012, 13, 1599–1616. [CrossRef]
12. Alonso-Mora, J.; Naegeli, T.; Siegwart, R.; Beardsley, P. Collision avoidance for aerial vehicles in multi-agent scenarios. Auton.
Robot. 2015, 39, 101–121. [CrossRef]
13. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [CrossRef] [PubMed]
14. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning; Springer: Berlin, Germany, 2001.
15. Nasrabadi, N.M. Pattern recognition and machine learning. J. Electron. Imaging 2007, 16, 049901.
16. Sun, Y.; Liu, J.; Yu, K.; Alazab, M.; Lin, K. PMRSS: Privacy-preserving Medical Record Searching Scheme for Intelligent Diagnosis
in IoT Healthcare. IEEE Trans. Ind. Inform. 2021, 18, 1. [CrossRef]
17. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of
the 2012 International Conference on Neural Information Processing Systems (NIPS), Lake Tahoe, NV, USA , 3–6 December 2012;
pp. 1097–1105.
18. Weng, G.; Dong, B.; Lei, Y. A level set method based on additive bias correction for image segmentation. Expert Syst. Appl. 2021,
185, 115633. [CrossRef]
19. Ge, P.; Chen, Y.; Wang, G.; Weng, G. A hybrid active contour model based on pre-fitting energy and adaptive functions for fast
image segmentation. Pattern Recognit. Lett. 2022, 158, 71–79. [CrossRef]
20. Zhang, X.; Yang, L.; Ding, Z.; Song, J.; Zhai, Y.; Zhang, D. Sparse Vector Coding-based Multi-Carrier NOMA for In-Home Health
Networks. IEEE J. Sel. Areas Commun. 2021, 39, 325–337. [CrossRef]
21. Chen, Y.; Zhou, Y. Machine learning based decision making for time varying systems: Parameter estimation and performance
optimization. Knowl.-Based Syst. 2020, 190, 105479. [CrossRef]
22. Chen, Y.; Zhou, Y.; Zhang, Y. Machine Learning-Based Model Predictive Control for Collaborative Production Planning Problem
with Unknown Information. Electronics 2021, 10, 1818. [CrossRef]
23. Chen, Y.; Jiang, W.; Charalambous, T. Machine learning based iterative learning control for non-repetitive time-varying systems.
Int. J. Robust Nonlinear Control 2022, Early View. [CrossRef]
24. Tao, H.; Wang, P.; Chen, Y.; Stojanovic, V.; Yang, H. An unsupervised fault diagnosis method for rolling bearing using STFT and
generative neural networks. J. Franklin Inst. 2020, 357, 7286–7307. [CrossRef]
25. Jaradat, M.A.K.; Al-Rousan, M.; Quadan, L. Reinforcement based mobile robot navigation in dynamic environment. Robot.
Comput.-Integr. Manuf. 2011, 27, 135–149. [CrossRef]
26. Duguleana, M.; Mogan, G. Neural networks based reinforcement learning for mobile robots obstacle avoidance. Expert Syst.
Appl. 2016, 62, 104–115. [CrossRef]
27. Zhu, K.; Zhang, T. Deep reinforcement learning based mobile robot navigation: A review. Tsinghua Sci. Technol. 2021, 26, 674–691.
[CrossRef]
28. Li, H.; Zhang, Q.; Zhao, D. Deep Reinforcement Learning-Based Automatic Exploration for Navigation in Unknown Environment.
IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 2064–2076. [CrossRef] [PubMed]
29. Xiao, W.; Yuan, L.; He, L.; Ran, T.; Zhang, J.; Cui, J. Multigoal Visual Navigation with Collision Avoidance via Deep Reinforcement
Learning. IEEE Trans. Instrum. Meas. 2022, 71, 1–9. [CrossRef]
30. Ni, Z.; He, H.; Wen, J.; Xu, X. Goal Representation Heuristic Dynamic Programming on Maze Navigation. IEEE Trans. Neural
Netw. Learn. Syst. 2013, 24, 2038–2050. [PubMed]
Appl. Sci. 2022, 12, 7796 17 of 17
31. Kim, C.J.; Chwa, D. Obstacle Avoidance Method for Wheeled MobileRobots Using Interval Type-2 FuzzyNeural Network. IEEE
Trans. Fuzzy Syst. 2015, 23, 677–687. [CrossRef]
32. Talbot, B.; Dayoub, F.; Corke, P.; Wyeth, G. Robot Navigation in Unseen Spaces Using an Abstract Map. IEEE Trans. Cogn. Dev.
Syst. 2021, 13, 791–805. [CrossRef]
33. Yuan, W.; Li, Z.; Su, C.Y. Multisensor-Based Navigation and Control of a Mobile Service Robot. IEEE Trans. Syst. Man Cybern.
Syst. 2021, 51, 2624–2634. [CrossRef]
34. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning; Springer: Berlin, Germany, 2013.
35. Chen, Y.; Chu, B.; Freeman, C.T. Generalized Iterative Learning Control using Successive Projection: Algorithm, Convergence
and Experimental Verification. IEEE Trans. Control Syst. Technol. 2020, 28, 2079–2091. [CrossRef]
36. Chen, Y.; Chu, B.; Freeman, C.T. Iterative Learning Control for Path-Following Tasks with Performance Optimization. IEEE Trans.
Control Syst. Technol. 2022, 30, 234–246. [CrossRef]
37. Chen, Y.; Chu, B.; Freeman, C.T. Iterative Learning Control for Robotic Path Following with Trial-Varying Motion Profiles.
IEEE/ASME Trans. Mechatron. 2022, Unpublished work. [CrossRef]