Off-Policy Adversarial Inverse Reinforcement Learning in Mobile Robot Navigation Task

Subeno, Muhammad Rizqi and Ataka, Ahmad and Ardiyanto, Igi and Cahyadi, Adha Imam (2024) Off-Policy Adversarial Inverse Reinforcement Learning in Mobile Robot Navigation Task. In: 2024 IEEE International Conference on Artificial Intelligence and Mechatronics Systems (AIMS), 21 - 23 Februari 2024, Bandung.

[thumbnail of Off-Policy_Adversarial_Inverse_Reinforcement_Learning_in_Mobile_Robot_Navigation_Task.pdf] Text
Off-Policy_Adversarial_Inverse_Reinforcement_Learning_in_Mobile_Robot_Navigation_Task.pdf - Published Version
Restricted to Registered users only

Download (1MB) | Request a copy

Abstract

Navigation tasks on mobile robots are getting more important to assist human tasks. However, navigation on mobile robots generally has several challenges, such as performing the task of navigating an unknown environment and being able to avoid obstacles. Conventional methods generally rely only on obstacle maps and cannot learn autonomously. Mapless navigation methods based on Reinforcement Learning (RL) are widely used to learn autonomous mobile robot navigation systems. However, the RL method requires a suitable reward function setting technique, and a simple reward function may not be easy to determine for the robot to perform navigation tasks. Therefore, this research proposed Off-Policy Adversarial Inverse Reinforcement Learning (AIRL) on mobile robot navigation tasks. We choose Off-Policy AIRL because it has strong advantages against dynamic changes so that mobile robots can perform navigation tasks efficiently. We compared the cumulative reward value, success rate, and trajectory length to reach the destination point with a reward designed via the reward shaping method and the reward estimated by Off-Policy AIRL. By applying RL using reward estimation from Off-Policy AIRL, the success rate of the robot reaching the destination point is 14 out of 15 scenarios and the proposed algorithm can outperform the reward produced by manually-reward shaping in 12 out of 15 scenarios. © 2024 IEEE.

Item Type: Conference or Workshop Item (Paper)
Additional Information: Cited by: 0
Uncontrolled Keywords: Inverse problems; Learning systems; Navigation; Navigation systems; Reinforcement learning; Adversarial inverse reinforcement learning; Autonomous navigation; Destination points; Human tasks; Inverse reinforcement learning; Learn+; Mobile Robot Navigation; Navigation tasks; Reinforcement learnings; Reward function; Mobile robots
Subjects: T Technology > TK Electrical engineering. Electronics Nuclear engineering
Divisions: Faculty of Engineering > Electrical and Information Technology Department
Depositing User: Rita Yulianti Yulianti
Date Deposited: 18 Feb 2025 04:47
Last Modified: 18 Feb 2025 04:47
URI: https://ir.lib.ugm.ac.id/id/eprint/13689

Actions (login required)

View Item
View Item