Acta Scientific Computer Sciences

Review Article Volume 5 Issue 5

Advances in Reinforcement Learning: A Comprehensive Review of Real-World Applications in Industry

K Ussenko* and VI Goncharov

Division for Automation and Robotics, School of Computer Science and Robotics, Tomsk Polytechnic University, Russian Federation

*Corresponding Author: K Ussenko, Division for Automation and Robotics, School of Computer Science and Robotics, Tomsk Polytechnic University, Russian Federation.

Received: March 15, 2023; Published: April 11, 2023


This paper investigates the current feasibility of utilizing reinforcement learning algorithms in the industrial sector. Although many studies have showcased the success of these algorithms in simulations or on isolated real-world objects, there is a paucity of research examining their wider implementation in real-world systems. In this study, we identify the obstacles that must be surmounted to fully leverage the potential benefits of reinforcement learning algorithms in practical applications. Moreover, we present a thorough overview of existing literature aimed at tackling these challenges.

Keywords: Reinforcement Learning; Deep Learning; Sim-to-real; Engineering; Artificial Intelligence; Control; Robotics; Autonomous Control


  1. Sutton Richard S and Andrew G Barto. “Reinforcement Learning: An Introduction”. The MIT Press, (1998).
  2. Nian Rui., et al. “A Review On Reinforcement Learning: Introduction and Applications in Industrial Process Control”. Computers and Chemical Engineering 139 (2020): 106886.
  3. Mnih, Volodymyr., et al. “Human-Level Control through Deep Reinforcement Learning”. Nature7540 (2015): 529-533.
  4. van Hasselt Hado., et al. “Deep Reinforcement Learning with Double Q-Learning”. arXiv:1509.06461, arXiv, 8 Dec. (2015).
  5. Barto Andrew G., et al. “Neuronlike Adaptive Elements That Can Solve Difficult Learning Control Problems”. IEEE Transactions on Systems, Man, and Cybernetics SMC-13.5 (1983): 834-846.
  6. Lillicrap Timothy P., et al. “Continuous Control with Deep Reinforcement Learning”. 1, arXiv:1509.02971, arXiv (2015).
  7. Dulac-Arnold Gabriel., et al. “Challenges of Real-World Reinforcement Learning: Definitions, Benchmarks and Analysis”. Machine Learning9 (2021): 2419-2468.
  8. Hwangbo Jemin., et al. “Control of a Quadrotor With Reinforcement Learning”. IEEE Robotics and Automation Letters4 (2017): 2096-2103.
  9. Haarnoja Tuomas., et al. “Soft Actor-Critic Algorithms and Applications”. arXiv:1812.05905, arXiv, 29 Jan. (2019).
  10. Singh Avi., et al. “End-to-End Robotic Reinforcement Learning without Reward Engineering”. arXiv:1904.07854, arXiv, 15 May (2019).
  11. McClement Daniel G., et al. “Meta-Reinforcement Learning for the Tuning of PI Controllers: An Offline Approach”. Journal of Process Control 118 (2022): 139-152.
  12. Hui Jonathan. “RL — Model-Based Reinforcement Learning”. Medium (2019).
  13. Moerland Thomas M., et al. “Model-Based Reinforcement Learning: A Survey”. arXiv:2006.16712, arXiv, 31 Mar. (2022).
  14. Sünderhauf Niko., et al. “The Limits and Potentials of Deep Learning for Robotics”. The International Journal of Robotics Research4-5 (2018): 405-420.
  15. Thuruthel Thomas George., et al. “Model-Based Reinforcement Learning for Closed-Loop Dynamic Control of Soft Robotic Manipulators”. IEEE Transactions on Robotics1 (2019): 124-134.
  16. Ju Hao., et al. “Transferring Policy of Deep Reinforcement Learning from Simulation to Reality for Robotics”. Nature Machine Intelligence12 (2022): 1077-1087.
  17. Zhao Wenshuai., et al. “Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: A Survey”. 2020 IEEE Symposium Series on Computational Intelligence (SSCI) (2020): 737-744.
  18. Yang Chenhao., et al. “Relative Camera Pose Estimation Using Synthetic Data with Domain Adaptation via Cycle-Consistent Adversarial Networks”. Journal of Intelligent and Robotic Systems4 (2021): 79.
  19. Tobin Josh., et al. “Domain Randomization for Transferring Deep Neural Networks from Simulation to the Real World”. 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (2017): 23-30.
  20. Mankowitz Daniel J., et al. “Situational Awareness by Risk-Conscious Skills”. arXiv:1610.02847, arXiv, 10 Oct. (2016).
  21. Christiano Paul., et al. “Transfer from Simulation to Real World through Learning Deep Inverse Dynamics Model”. arXiv:1610.03518, arXiv, 11 Oct. (2016).
  22. Hanna Josiah P., et al. “Grounded Action Transformation for Sim-to-Real Reinforcement Learning”. Machine Learning9 (2021): 2469-2499.
  23. Tzeng Eric., et al. “Adapting Deep Visuomotor Representations with Weak Pairwise Constraints”. Algorithmic Foundations of Robotics XII: Proceedings of the Twelfth Workshop on the Algorithmic Foundations of Robotics, edited by Ken Goldberg et al., Springer International Publishing, (2020): 688-703.
  24. Andrychowicz OpenAI: Marcin., et al. “Learning Dexterous In-Hand Manipulation”. The International Journal of Robotics Research1 (2020): 3-20.
  25. Rusu Andrei A., et al. “Sim-to-Real Robot Learning from Pixels with Progressive Nets”. Proceedings of the 1st Annual Conference on Robot Learning, PMLR, (2017): 262-270.
  26. Nagabandi Anusha., et al. “Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning”. arXiv:1803.11347, arXiv, 27 Feb. (2019).
  27. Arndt Karol., et al. “Meta Reinforcement Learning for Sim-to-Real Domain Adaptation”. 2020 IEEE International Conference on Robotics and Automation (ICRA) (2020): 2725-2731.
  28. Wang Pin., et al. “A Reinforcement Learning Based Approach for Automated Lane Change Maneuvers”. 2018 IEEE Intelligent Vehicles Symposium (IV), (2018): 1379-1384.
  29. Gu Shixiang., et al. “Deep Reinforcement Learning for Robotic Manipulation with Asynchronous Off-Policy Updates”. 2017 IEEE International Conference on Robotics and Automation (ICRA), (2017): 3389-3396.
  30. Bakker B., et al. “Quasi-Online Reinforcement Learning for Robots”. Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006., (2006): 2997-3002.
  31. Hester Todd., et al. “RTMBA: A Real-Time Model-Based Reinforcement Learning Architecture for Robot Control”. 2012 IEEE International Conference on Robotics and Automation, (2012): 85-90.
  32. Riedmiller Martin., et al. “Reinforcement Learning for Robot Soccer”. Autonomous Robots1 (2009): 55-73.
  33. Fang Guoxin., et al. “Efficient Jacobian-Based Inverse Kinematics With Sim-to-Real Transfer of Soft Robots by Learning”. IEEE/ASME Transactions on Mechatronics6 ( 2022): 5296-5306.
  34. Chaplot Devendra Singh., et al. “Active Neural Localization”. arXiv:1801.08214, arXiv, 24 Jan. (2018).
  35. Wu Chunxue., et al. “UAV Autonomous Target Search Based on Deep Reinforcement Learning in Complex Disaster Scene”. IEEE Access 7 (2019): 117227-117245.
  36. Lei Lei., et al. “Deep Reinforcement Learning for Autonomous Internet of Things: Model, Applications and Challenges”. IEEE Communications Surveys and Tutorials3 (2020): 1722-1760.
  37. Saravanan M., et al. “IoT Enabled Indoor Autonomous Mobile Robot Using CNN and Q-Learning”. 2019 IEEE International Conference on Industry 4.0, Artificial Intelligence, and Communications Technology (IAICT), (2019): 7-13.
  38. Liu Xiaolan., et al. “Resource Allocation for Edge Computing in IoT Networks via Reinforcement Learning”. ICC 2019 - 2019 IEEE International Conference on Communications (ICC), (2019): 1-6.
  39. Foruzan Elham., et al. “Reinforcement Learning Approach for Optimal Distributed Energy Management in a Microgrid”. IEEE Transactions on Power Systems5 (2018): 5749-5758.
  40. Zhang Yameng., et al. “A Deep Reinforcement Learning Approach for Online Computation Offloading in Mobile Edge Computing”. 2020 IEEE/ACM 28th International Symposium on Quality of Service (IWQoS), (2020): 1-10.
  41. Rolf Benjamin., et al. “A Review on Reinforcement Learning Algorithms and Applications in Supply Chain Management”. International Journal of Production Research (2022): 1-29.
  42. Wang Fang and Lin Lin. ‘Spare Parts Supply Chain Network Modeling Based on a Novel Scale-Free Network and Replenishment Path Optimization with Q Learning”. Computers and Industrial Engineering 157 (2021): 107312.
  43. Li Xijun., et al. “Learning to Optimize Industry-Scale Dynamic Pickup and Delivery Problems”. 2021 IEEE 37th International Conference on Data Engineering (ICDE), (2021): 2511-2522.
  44. Zhu Zheng., et al. “A Mean-Field Markov Decision Process Model for Spatial-Temporal Subsidies in Ride-Sourcing Markets”. Transportation Research Part B: Methodological 150 (2021): 540-565.
  45. Kegenbekov Zhandos and Ilya Jackson. ‘Adaptive Supply Chain: Demand-Supply Synchronization Using Deep Reinforcement Learning”. Algorithms8 (2021): 240.
  46. Zhao Zhijia., et al. “Reinforcement Learning Control for a 2-DOF Helicopter With State Constraints: Theory and Experiments”. IEEE Transactions on Automation Science and Engineering (2022): 1-11.
  47. Yang Tao., et al. “A Soft Artificial Muscle Driven Robot with Reinforcement Learning”. Scientific Reports1 (2018): 14518.
  48. Zhu Wei., et al. “A Survey of Sim-to-Real Transfer Techniques Applied to Reinforcement Learning for Bioinspired Robots”. IEEE Transactions on Neural Networks and Learning Systems (2021): 1-16.
  49. Ha Sehoon. “Quadrupedal Robots Trot into the Wild”. Science Robotics47 (2020): eabe5218.
  50. Naug Avisek., et al. “Deep Reinforcement Learning Control for Non-Stationary Building Energy Management”. Energy and Buildings 277 (2022): 112584.


Citation: K Ussenko and VI Goncharov. “Advances in Reinforcement Learning: A Comprehensive Review of Real-World Applications in Industry". Acta Scientific Computer Sciences 5.5 (2023): 32-38.


Copyright: © 2023 K Ussenko and VI Goncharov. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.


Acceptance rate35%
Acceptance to publication20-30 days

Indexed In

News and Events

  • Certification for Review
    Acta Scientific certifies the Editors/reviewers for their review done towards the assigned articles of the respective journals.
  • Submission Timeline for Upcoming Issue
    The last date for submission of articles for regular Issues is June 25, 2024.
  • Publication Certificate
    Authors will be issued a "Publication Certificate" as a mark of appreciation for publishing their work.
  • Best Article of the Issue
    The Editors will elect one Best Article after each issue release. The authors of this article will be provided with a certificate of "Best Article of the Issue"
  • Welcoming Article Submission
    Acta Scientific delightfully welcomes active researchers for submission of articles towards the upcoming issue of respective journals.

Contact US