Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic
The challenge of mobile robot navigation system is how to find the optimal path and be able to avoid obstacles. The obstacles could be stationary or moving objects, which make the environment of mobile robot would be changed dynamically. Regarding this problem, the conventional method should update...
Published in: | Proceedings - 2024 FORTEI-International Conference on Electrical Engineering: Empowering Innovations : Navigating The Future Of Semiconductor Industry, FORTEI-ICEE 2024 |
---|---|
Main Author: | |
Format: | Conference paper |
Language: | English |
Published: |
Institute of Electrical and Electronics Engineers Inc.
2024
|
Online Access: | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85217560118&doi=10.1109%2fFORTEI-ICEE64706.2024.10824360&partnerID=40&md5=1b15218e2331fd0090df4cea3056bbfd |
id |
2-s2.0-85217560118 |
---|---|
spelling |
2-s2.0-85217560118 Sendari S.; Muladi; Ardiyansyah F.; Zaeni I.A.E.; Widyaningtyas T.; Wibowo D.A.; Yuztar M.; Setumin S. Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic 2024 Proceedings - 2024 FORTEI-International Conference on Electrical Engineering: Empowering Innovations : Navigating The Future Of Semiconductor Industry, FORTEI-ICEE 2024 10.1109/FORTEI-ICEE64706.2024.10824360 https://www.scopus.com/inward/record.uri?eid=2-s2.0-85217560118&doi=10.1109%2fFORTEI-ICEE64706.2024.10824360&partnerID=40&md5=1b15218e2331fd0090df4cea3056bbfd The challenge of mobile robot navigation system is how to find the optimal path and be able to avoid obstacles. The obstacles could be stationary or moving objects, which make the environment of mobile robot would be changed dynamically. Regarding this problem, the conventional method should update the path everytime when there are any changes in environment, so it causes the computational burden. This paper aims to solve the problem by developing Deep Reinforcement Learning with Actor-Critic (DRL-AC) as navigation method to find the optimal path. DRL-AC determines robot movement by selecting action which have highest reward. DRL-AC is developed with actor and critic neural networks which are trained in dynamic environment. To prevent hardware damages during the training process, DRL-AC is developed in simulator. This simulator is designed to adopt real world condition. After training, DRL-AC can be implemented to the real mobile robot. The performance of DRL-AC is analyzed through neuron dimension and how the mobile robot navigated successfully in dynamic environment. The training results show that neural network with 2048 x 1024 hidden layer started to be convergent and mobile robot navigated successfully to avoid obstacle. © 2024 IEEE. Institute of Electrical and Electronics Engineers Inc. English Conference paper |
author |
Sendari S.; Muladi; Ardiyansyah F.; Zaeni I.A.E.; Widyaningtyas T.; Wibowo D.A.; Yuztar M.; Setumin S. |
spellingShingle |
Sendari S.; Muladi; Ardiyansyah F.; Zaeni I.A.E.; Widyaningtyas T.; Wibowo D.A.; Yuztar M.; Setumin S. Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
author_facet |
Sendari S.; Muladi; Ardiyansyah F.; Zaeni I.A.E.; Widyaningtyas T.; Wibowo D.A.; Yuztar M.; Setumin S. |
author_sort |
Sendari S.; Muladi; Ardiyansyah F.; Zaeni I.A.E.; Widyaningtyas T.; Wibowo D.A.; Yuztar M.; Setumin S. |
title |
Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
title_short |
Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
title_full |
Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
title_fullStr |
Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
title_full_unstemmed |
Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
title_sort |
Navigation System in Dynamic Indoor Environment Using Deep Reinforcement Learning with Actor-Critic |
publishDate |
2024 |
container_title |
Proceedings - 2024 FORTEI-International Conference on Electrical Engineering: Empowering Innovations : Navigating The Future Of Semiconductor Industry, FORTEI-ICEE 2024 |
container_volume |
|
container_issue |
|
doi_str_mv |
10.1109/FORTEI-ICEE64706.2024.10824360 |
url |
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85217560118&doi=10.1109%2fFORTEI-ICEE64706.2024.10824360&partnerID=40&md5=1b15218e2331fd0090df4cea3056bbfd |
description |
The challenge of mobile robot navigation system is how to find the optimal path and be able to avoid obstacles. The obstacles could be stationary or moving objects, which make the environment of mobile robot would be changed dynamically. Regarding this problem, the conventional method should update the path everytime when there are any changes in environment, so it causes the computational burden. This paper aims to solve the problem by developing Deep Reinforcement Learning with Actor-Critic (DRL-AC) as navigation method to find the optimal path. DRL-AC determines robot movement by selecting action which have highest reward. DRL-AC is developed with actor and critic neural networks which are trained in dynamic environment. To prevent hardware damages during the training process, DRL-AC is developed in simulator. This simulator is designed to adopt real world condition. After training, DRL-AC can be implemented to the real mobile robot. The performance of DRL-AC is analyzed through neuron dimension and how the mobile robot navigated successfully in dynamic environment. The training results show that neural network with 2048 x 1024 hidden layer started to be convergent and mobile robot navigated successfully to avoid obstacle. © 2024 IEEE. |
publisher |
Institute of Electrical and Electronics Engineers Inc. |
issn |
|
language |
English |
format |
Conference paper |
accesstype |
|
record_format |
scopus |
collection |
Scopus |
_version_ |
1825722578497437696 |