AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning
In a complex underwater environment, finding a viable, collision-free path for an autonomous underwater vehicle (AUV) is a challenging task. The purpose of this paper is to establish a safe, real-time, and robust method of collision avoidance that improves the autonomy of AUVs. We propose a method b...
Guardado en:
Autores principales: | , , , , , |
---|---|
Formato: | article |
Lenguaje: | EN |
Publicado: |
MDPI AG
2021
|
Materias: | |
Acceso en línea: | https://doaj.org/article/aa4f8fea66634cbb8579ebfc57708835 |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
id |
oai:doaj.org-article:aa4f8fea66634cbb8579ebfc57708835 |
---|---|
record_format |
dspace |
spelling |
oai:doaj.org-article:aa4f8fea66634cbb8579ebfc577088352021-11-25T18:03:49ZAUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning10.3390/jmse91111662077-1312https://doaj.org/article/aa4f8fea66634cbb8579ebfc577088352021-10-01T00:00:00Zhttps://www.mdpi.com/2077-1312/9/11/1166https://doaj.org/toc/2077-1312In a complex underwater environment, finding a viable, collision-free path for an autonomous underwater vehicle (AUV) is a challenging task. The purpose of this paper is to establish a safe, real-time, and robust method of collision avoidance that improves the autonomy of AUVs. We propose a method based on active sonar, which utilizes a deep reinforcement learning algorithm to learn the processed sonar information to navigate the AUV in an uncertain environment. We compare the performance of double deep Q-network algorithms with that of a genetic algorithm and deep learning. We propose a line-of-sight guidance method to mitigate abrupt changes in the yaw direction and smooth the heading changes when the AUV switches trajectory. The different experimental results show that the double deep Q-network algorithms ensure excellent collision avoidance performance. The effectiveness of the algorithm proposed in this paper was verified in three environments: random static, mixed static, and complex dynamic. The results show that the proposed algorithm has significant advantages over other algorithms in terms of success rate, collision avoidance performance, and generalization ability. The double deep Q-network algorithm proposed in this paper is superior to the genetic algorithm and deep learning in terms of the running time, total path, performance in avoiding collisions with moving obstacles, and planning time for each step. After the algorithm is trained in a simulated environment, it can still perform online learning according to the information of the environment after deployment and adjust the weight of the network in real-time. These results demonstrate that the proposed approach has significant potential for practical applications.Jianya YuanHongjian WangHonghan ZhangChangjian LinDan YuChengfeng LiMDPI AGarticleautonomous underwater vehicle (AUV)collision avoidance planningdeep reinforcement learning (DRL)double-DQN (D-DQN)Naval architecture. Shipbuilding. Marine engineeringVM1-989OceanographyGC1-1581ENJournal of Marine Science and Engineering, Vol 9, Iss 1166, p 1166 (2021) |
institution |
DOAJ |
collection |
DOAJ |
language |
EN |
topic |
autonomous underwater vehicle (AUV) collision avoidance planning deep reinforcement learning (DRL) double-DQN (D-DQN) Naval architecture. Shipbuilding. Marine engineering VM1-989 Oceanography GC1-1581 |
spellingShingle |
autonomous underwater vehicle (AUV) collision avoidance planning deep reinforcement learning (DRL) double-DQN (D-DQN) Naval architecture. Shipbuilding. Marine engineering VM1-989 Oceanography GC1-1581 Jianya Yuan Hongjian Wang Honghan Zhang Changjian Lin Dan Yu Chengfeng Li AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning |
description |
In a complex underwater environment, finding a viable, collision-free path for an autonomous underwater vehicle (AUV) is a challenging task. The purpose of this paper is to establish a safe, real-time, and robust method of collision avoidance that improves the autonomy of AUVs. We propose a method based on active sonar, which utilizes a deep reinforcement learning algorithm to learn the processed sonar information to navigate the AUV in an uncertain environment. We compare the performance of double deep Q-network algorithms with that of a genetic algorithm and deep learning. We propose a line-of-sight guidance method to mitigate abrupt changes in the yaw direction and smooth the heading changes when the AUV switches trajectory. The different experimental results show that the double deep Q-network algorithms ensure excellent collision avoidance performance. The effectiveness of the algorithm proposed in this paper was verified in three environments: random static, mixed static, and complex dynamic. The results show that the proposed algorithm has significant advantages over other algorithms in terms of success rate, collision avoidance performance, and generalization ability. The double deep Q-network algorithm proposed in this paper is superior to the genetic algorithm and deep learning in terms of the running time, total path, performance in avoiding collisions with moving obstacles, and planning time for each step. After the algorithm is trained in a simulated environment, it can still perform online learning according to the information of the environment after deployment and adjust the weight of the network in real-time. These results demonstrate that the proposed approach has significant potential for practical applications. |
format |
article |
author |
Jianya Yuan Hongjian Wang Honghan Zhang Changjian Lin Dan Yu Chengfeng Li |
author_facet |
Jianya Yuan Hongjian Wang Honghan Zhang Changjian Lin Dan Yu Chengfeng Li |
author_sort |
Jianya Yuan |
title |
AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning |
title_short |
AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning |
title_full |
AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning |
title_fullStr |
AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning |
title_full_unstemmed |
AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning |
title_sort |
auv obstacle avoidance planning based on deep reinforcement learning |
publisher |
MDPI AG |
publishDate |
2021 |
url |
https://doaj.org/article/aa4f8fea66634cbb8579ebfc57708835 |
work_keys_str_mv |
AT jianyayuan auvobstacleavoidanceplanningbasedondeepreinforcementlearning AT hongjianwang auvobstacleavoidanceplanningbasedondeepreinforcementlearning AT honghanzhang auvobstacleavoidanceplanningbasedondeepreinforcementlearning AT changjianlin auvobstacleavoidanceplanningbasedondeepreinforcementlearning AT danyu auvobstacleavoidanceplanningbasedondeepreinforcementlearning AT chengfengli auvobstacleavoidanceplanningbasedondeepreinforcementlearning |
_version_ |
1718411699738902528 |