Dynamic Rewards in Reinforcement Learning for Robotic Navigation

Authors

  • Ahmed Badi Alshammari Department of Computer Sciences, Faculty of Computing and Information Technology, Northern Border University, Rafha, Saudi Arabia
Volume: 15 | Issue: 4 | Pages: 25766-25771 | August 2025 | https://doi.org/10.48084/etasr.11986

Abstract

The paper presents a new reinforcement learning method, called Adaptive Q-learning with Dynamic Reward (AQDR), for efficient route planning of mobile robots operating in a partially known and unknown environment. Traditional Q-learning techniques are often limited in their adaptability due to slow convergence in dynamic environments. To overcome these limitations, AQDR combines an adaptive reward mechanism that adjusts in real time based on the distance of the robot to the obstacle and the target position. This dynamic feedback allows for better informed decision making and reduces unnecessary exploration. The proposed algorithm was evaluated against the Q-learning method based on static rewards by performing simulated experiments in different environmental configurations. The results show that AQDR consistently exceeds the baseline in terms of convergence speed, path efficiency, and adaptability. These results highlight the potential of dynamic reward design to improve learning performance and robustness of reinforcement learning navigation systems.

Keywords:

reinforcement learning, Q-learning, dynamic reward, mobile robot navigation, path planning, adaptive algorithms

Downloads

Download data is not yet available.

References

L. Liu, X. Wang, X. Yang, H. Liu, J. Li, and P. Wang, "Path planning techniques for mobile robots: Review and prospect," Expert Systems with Applications, vol. 227, Oct. 2023, Art. no. 120254. DOI: https://doi.org/10.1016/j.eswa.2023.120254

U. Orozco-Rosas, K. Picos, J. J. Pantrigo, A. S. Montemayor, and A. Cuesta-Infante, "Mobile Robot Path Planning Using a QAPF Learning Algorithm for Known and Unknown Environments," IEEE Access, vol. 10, pp. 84648–84663, 2022. DOI: https://doi.org/10.1109/ACCESS.2022.3197628

Y. Wang, X. Li, J. Zhang, S. Li, Z. Xu, and X. Zhou, "Review of wheeled mobile robot collision avoidance under unknown environment," Science Progress, vol. 104, no. 3, Jul. 2021, Art. no. 00368504211037771. DOI: https://doi.org/10.1177/00368504211037771

O. A. A. Salama, M. E. H. Eltaib, H. A. Mohamed, and O. Salah, "RCD: Radial Cell Decomposition Algorithm for Mobile Robot Path Planning," IEEE Access, vol. 9, pp. 149982–149992, 2021. DOI: https://doi.org/10.1109/ACCESS.2021.3125105

A. N. A. Rafai, N. Adzhar, and N. I. Jaini, "A Review on Path Planning and Obstacle Avoidance Algorithms for Autonomous Mobile Robots," Journal of Robotics, vol. 2022, no. 1, 2022, Art. no. 2538220. DOI: https://doi.org/10.1155/2022/2538220

J. Luo, Z. X. Wang, and K. L. Pan, "Reliable Path Planning Algorithm Based on Improved Artificial Potential Field Method," IEEE Access, vol. 10, pp. 108276–108284, 2022. DOI: https://doi.org/10.1109/ACCESS.2022.3212741

Y. Xu, Q. Li, X. Xu, J. Yang, and Y. Chen, "Research Progress of Nature-Inspired Metaheuristic Algorithms in Mobile Robot Path Planning," Electronics, vol. 12, no. 15, Jan. 2023, Art. no. 3263. DOI: https://doi.org/10.3390/electronics12153263

M. N. Ab Wahab et al., "Improved genetic algorithm for mobile robot path planning in static environments," Expert Systems with Applications, vol. 249, Sep. 2024, Art. no. 123762. DOI: https://doi.org/10.1016/j.eswa.2024.123762

K. Shi, Z. Wu, B. Jiang, and H. R. Karimi, "Dynamic path planning of mobile robot based on improved simulated annealing algorithm," Journal of the Franklin Institute, vol. 360, no. 6, pp. 4378–4398, Apr. 2023. DOI: https://doi.org/10.1016/j.jfranklin.2023.01.033

L. Qin and Z. Fan, "Robot Path Planning Based on Tabu Particle Swarm Optimization Integrating Cauchy Mutation," in 2024 14th International Conference on Information Science and Technology (ICIST), Chengdu, China, Dec. 2024, pp. 244–253. DOI: https://doi.org/10.1109/ICIST63249.2024.10805364

B. Tao and J. H. Kim, "Mobile robot path planning based on bi-population particle swarm optimization with random perturbation strategy," Journal of King Saud University - Computer and Information Sciences, vol. 36, no. 2, Feb. 2024, Art. no. 101974. DOI: https://doi.org/10.1016/j.jksuci.2024.101974

H. Huang, G. Tan, and L. Jiang, "Robot Path Planning Using Improved Ant Colony Algorithm in the Environment of Internet of Things," Journal of Robotics, vol. 2022, no. 1, 2022, Art. no. 1739884. DOI: https://doi.org/10.1155/2022/1739884

Z. Wang, J. Peng, S. Ding, and School of Electrical Engineering, Zhengzhou University, No. 100 of Science Avenue, Zhengzhou 450001, China, "A Bio-inspired trajectory planning method for robotic manipulators based on improved bacteria foraging optimization algorithm and tau theory," Mathematical Biosciences and Engineering, vol. 19, no. 1, pp. 643–662, 2021. DOI: https://doi.org/10.3934/mbe.2022029

Y. Cui, W. Hu, and A. Rahmani, "A reinforcement learning based artificial bee colony algorithm with application in robot path planning," Expert Systems with Applications, vol. 203, Oct. 2022, Art. no. 117389. DOI: https://doi.org/10.1016/j.eswa.2022.117389

A. Hentout, A. Maoudj, and M. Aouache, "A review of the literature on fuzzy-logic approaches for collision-free path planning of manipulator robots," Artificial Intelligence Review, vol. 56, no. 4, pp. 3369–3444, Apr. 2023. DOI: https://doi.org/10.1007/s10462-022-10257-7

J. Wang, J. Liu, W. Chen, W. Chi, and M. Q. H. Meng, "Robot Path Planning via Neural-Network-Driven Prediction," IEEE Transactions on Artificial Intelligence, vol. 3, no. 3, pp. 451–460, Jun. 2022. DOI: https://doi.org/10.1109/TAI.2021.3119890

C. Han and B. Li, "Mobile Robot Path Planning Based on Improved A* Algorithm," in 2023 IEEE 11th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), Chongqing, China, Dec. 2023, pp. 672–676. DOI: https://doi.org/10.1109/ITAIC58329.2023.10408799

L. Dong, Z. He, C. Song, and C. Sun, "A review of mobile robot motion planning methods: from classical motion planning workflows to reinforcement learning-based architectures," Journal of Systems Engineering and Electronics, vol. 34, no. 2, pp. 439–459, Apr. 2023. DOI: https://doi.org/10.23919/JSEE.2023.000051

Q. Zhou, Y. Lian, J. Wu, M. Zhu, H. Wang, and J. Cao, "An optimized Q-Learning algorithm for mobile robot local path planning," Knowledge-Based Systems, vol. 286, Feb. 2024, Art. no. 111400. DOI: https://doi.org/10.1016/j.knosys.2024.111400

A. N. Kadhim and M. S. Salim, "Enhancing Navigation Efficiency in Robotics with PRM-DDPG," Engineering, Technology & Applied Science Research, vol. 15, no. 3, pp. 24077–24086, Jun. 2025. DOI: https://doi.org/10.48084/etasr.11213

Downloads

How to Cite

[1]
A. B. Alshammari, “Dynamic Rewards in Reinforcement Learning for Robotic Navigation”, Eng. Technol. Appl. Sci. Res., vol. 15, no. 4, pp. 25766–25771, Aug. 2025.

Metrics

Abstract Views: 112
PDF Downloads: 241

Metrics Information