
Sign up to save your podcasts
Or
Article Link | Author's Personal Website
Abstract: Path planning is one of the most essential parts of autonomous navigation. Most existing works suppose that the environment is static and fixed. However, path planning is widely used in random and dynamic environments (such as search and rescue, surveillance, and other scenarios). In this paper, we propose a Deep Reinforcement Learning (DRL)-based method that enables unmanned aerial vehicles (UAVs) to execute navigation tasks in multi-obstacle environments with randomness and dynamics. The method is based on the Twin Delayed Deep Deterministic Policy Gradients (TD3) algorithm. In order to predict the impact of the environment on UAV, the change of environment observations is added to the Actor-Critic network input, and the two-stream Actor-Critic network structure is proposed to extract features of environment observations. Simulations are carried out to evaluate the performance of the algorithm and experiment results show that our method can enable the UAV to complete autonomous navigation tasks safely in multi-obstacle environments, which reflects the efficiency of our method. Moreover, compared to DDPG and the conventional TD3, our method has better generalization ability.
⭐️ Note: All episodes are automatically generated by NotebookLM, and there may be some discrepancies between the content and the original texts.
Article Link | Author's Personal Website
Abstract: Path planning is one of the most essential parts of autonomous navigation. Most existing works suppose that the environment is static and fixed. However, path planning is widely used in random and dynamic environments (such as search and rescue, surveillance, and other scenarios). In this paper, we propose a Deep Reinforcement Learning (DRL)-based method that enables unmanned aerial vehicles (UAVs) to execute navigation tasks in multi-obstacle environments with randomness and dynamics. The method is based on the Twin Delayed Deep Deterministic Policy Gradients (TD3) algorithm. In order to predict the impact of the environment on UAV, the change of environment observations is added to the Actor-Critic network input, and the two-stream Actor-Critic network structure is proposed to extract features of environment observations. Simulations are carried out to evaluate the performance of the algorithm and experiment results show that our method can enable the UAV to complete autonomous navigation tasks safely in multi-obstacle environments, which reflects the efficiency of our method. Moreover, compared to DDPG and the conventional TD3, our method has better generalization ability.
⭐️ Note: All episodes are automatically generated by NotebookLM, and there may be some discrepancies between the content and the original texts.