TY - GEN
T1 - A Mini Review on the utilization of Reinforcement Learning with OPC UA
AU - Schindler, S.
AU - Uray, M.
AU - Huber, S.
N1 - Conference code: 192026
Export Date: 14 December 2023
Correspondence Address: Schindler, S.; Salzburg University of Applied Sciences, Austria; email: [email protected]
References: OPC UA Online Reference-Released Specifications, , https://reference.opcfoundation.org/; Unified Architecture-OPC Foundation, , https://opcfoundation.org/about/opc-technologies/opc-ua/; Abdoune, F., Nouiri, M., Cardin, O., Castagna, P., Integration of artificial intelligence in the life cycle of industrial digital twins (2022) IFACPapersOnLine, 55 (1), pp. 2545-2550; Bakakeu, J., Bauer, J., An artificial intelligence approach for online energy optimization of flexible manufacturing systems (2018) Article in Applied Mechanics and Materials; Bakakeu, J., Schafer, F., Franke, J., Baer, S., Klos, H.H., Peschke, J., Reasoning over opc ua information models using graph embedding and reinforcement learning (2020) Proceedings-2020 3rd International Conference on Artificial Intelligence for Industries, AI4I 2020, 9, pp. 40-47; Blasi, S.D., Klser, S., Mller, A., Reuben, R., Sturm, F., Zerrer, T., Kicker: An industrial drive and control foosball system automated with deep reinforcement learning (2021) Journal of Intelligent and Robotic Systems: Theory and Applications, 102 (5), pp. 1-18; Burggrf, P., Steinberg, F., Heinbach, B., Bamberg, M., Reinforcement learning for process time optimization in an assembly process utilizing an industry 4. 0 demonstration cell (2022) Procedia CIRP, 107 (1), pp. 1095-1100; Csiszar, A., Krimstein, V., Bogner, J., Verl, A., Generating reinforcement learning environments for industrial communication protocols (2021) Proceedings-2021 4th International Conference on Artificial Intelligence for Industries, AI4I 2021, pp. 57-60; Dobrescu, R., Chenaru, O., Florea, G., Geampalia, G., Mocanu, S., Hardware-in-loop assessment of control architectures (2020) 2020 24th International Conference on System Theory, Control and Computing, ICSTCC 2020-Proceedings, 10, pp. 880-885; Dogru, O., Velswamy, K., Ibrahim, F., Wu, Y., Sundaramoorthy, A.S., Huang, B., Xu, S., Bell, N., Reinforcement learning approach to autonomous pid tuning (2022) Computers and Chemical Engineering, 161 (5), p. 107760; Gracia, J.B., Leber, F., Aburaia, M., Wber, W., A configurable skill oriented architecture based on opc ua (2022) 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 7317-7322; Grothoff, J., Kleinert, T., (2021) Mapping of standardized state machines to utilize machine learning models in process control environments, pp. 39-53; Hermann, M., Pentek, T., Otto, B., Design principles for industrie 4. 0 scenarios (2016) Proceedings of the Annual Hawaii International Conference on System Sciences, 2016, pp. 3928-3937. , March, 3; Khaydarov, V., Neuendorf, L., Kock, T., Kockmann, N., Urbas, L., Mtppy: Open-source ai-friendly modular automation (2022) 2022 IEEE 27th International Conference on Emerging Technologies and Factory Automation (ETFA), pp. 1-7; Kober, J., Bagnell, J.A., Peters, J., Reinforcement learning in robotics: A survey (2013) International Journal of Robotics Research, 32; Luketina, J., Nardelli, N., Farquhar, G., Foerster, J., Andreas, J., Grefenstette, E., Whiteson, S., Rocktschel, T., A Survey of Reinforcement Learning Informed by Natural Language (2019) IJCAI International Joint Conference on Artificial Intelligence, 2019 (6), pp. 6309-6317. , August; Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller, M., (2013) Playing Atari with Deep Reinforcement Learning, , 12; Nian, R., Liu, J., Huang, B., A review on reinforcement learning: Introduction and applications in industrial process control (2020) Computers and Chemical Engineering, 139, p. 106886; Randolph, J., A guide to writing the dissertation literature review (2019) Practical Assessment, Research, and Evaluation, 14, p. 13. , 11; Rohrer, T., Samuel, L., Gashi, A., Grieser, G., Hergenröther, E., Foosball table goalkeeper automation using reinforcement learning (2021) LWDA, pp. 173-182; Schäfer, G., Kozlica, R., Wegenkittl, S., Huber, S., An architecture for deploying reinforcement learning in industrial environments (2022) Computer Aided Systems Theory-EUROCAST 2022, pp. 569-576; Schleipen, M., Gilani, S.-S., Bischoff, T., Pfrommer, J., Opc ua and industrie 4. 0-enabling technology with high diversity and variability (2016) Procedia CIRP, 57, pp. 315-320. , factories of the Future in the digital environment-Proceedings of the 49th CIRP Conference on Manufacturing Systems; Schmidl, E., Fischer, E., Steindl, J., Wenk, M., Franke, J., Reinforcement learning for energy reduction of conveying and handling systems (2021) Procedia CIRP, 97 (1), pp. 290-295; Schmidl, E., Fischer, E., Wenk, M., Franke, J., Knowledge-based generation of a plant-specific reinforcement learning framework for energy reduction of production plants (2020) 2020 25th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), 1, pp. 1-4; Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Hassabis, D., Mastering the game of go without human knowledge (2017) Nature, 550, pp. 354-359. , 10 2017 550: 7676; Sutton, R.S., Barto, A.G., (2018) Reinforcement Learning: An Introduction, , MIT press; Tschuchnig, M.E., Gadermayr, M., Anomaly detection in medical imaging-a mini review (2022) Data Science Analytics and Applications, 8, pp. 33-38; Uc-Cetina, V., Navarro-Guerrero, N., Martin-Gonzalez, A., Weber, C., Wermter, S., Survey on reinforcement learning for language processing (2022) Artificial Intelligence Review, 6, pp. 1-33; Xia, K., Sacco, C., Kirkpatrick, M., Saidy, C., Nguyen, L., Kircaliali, A., Harik, R., A digital twin to train deep reinforcement learning agent for smart manufacturing plants: Environment, interfaces and intelligence (2021) Journal of Manufacturing Systems, 58 (1), pp. 210-230; Zhao, W., Queralta, J.P., Westerlund, T., Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: A Survey 2020 IEEE Symposium Series on Computational Intelligence (SSCI). IEEE, pp. 737-744; Zheng, P., Xia, L., Li, C., Li, X., Liu, B., Towards self-x cognitive manufacturing network: An industrial knowledge graph-based multiagent reinforcement learning approach (2021) Journal of Manufacturing Systems, 61 (10), pp. 16-26
PY - 2023
Y1 - 2023
N2 - Reinforcement Learning (RL) is a powerful machine learning paradigm that has been applied in various fields such as robotics, natural language processing and game playing achieving state-of-the-art results. Targeted to solve sequential decision making problems, it is by design able to learn from experience and therefore adapt to changing dynamic environments. These capabilities make it a prime candidate for controlling and optimizing complex processes in industry. The key to fully exploiting this potential is the seamless integration of RL into existing industrial systems. The industrial communication standard Open Platform Communications Unified Architecture (OPC UA) could bridge this gap.However, since RL and OPC UA are from different fields, there is a need for researchers to bridge the gap between the two technologies. This work serves to bridge this gap by providing a brief technical overview of both technologies and carrying out a semi-exhaustive literature review to gain insights on how RL and OPC UA are applied in combination.With this survey, three main research topics have been identified, following the intersection of RL with OPC UA. The results of the literature review show that RL is a promising technology for the control and optimization of industrial processes, but does not yet have the necessary standardized interfaces to be deployed in real-world scenarios with reasonably low effort. © 2023 IEEE.
AB - Reinforcement Learning (RL) is a powerful machine learning paradigm that has been applied in various fields such as robotics, natural language processing and game playing achieving state-of-the-art results. Targeted to solve sequential decision making problems, it is by design able to learn from experience and therefore adapt to changing dynamic environments. These capabilities make it a prime candidate for controlling and optimizing complex processes in industry. The key to fully exploiting this potential is the seamless integration of RL into existing industrial systems. The industrial communication standard Open Platform Communications Unified Architecture (OPC UA) could bridge this gap.However, since RL and OPC UA are from different fields, there is a need for researchers to bridge the gap between the two technologies. This work serves to bridge this gap by providing a brief technical overview of both technologies and carrying out a semi-exhaustive literature review to gain insights on how RL and OPC UA are applied in combination.With this survey, three main research topics have been identified, following the intersection of RL with OPC UA. The results of the literature review show that RL is a promising technology for the control and optimization of industrial processes, but does not yet have the necessary standardized interfaces to be deployed in real-world scenarios with reasonably low effort. © 2023 IEEE.
KW - OPC UA
KW - Reinforcement Learning
KW - Survey
KW - Computer architecture
KW - Decision making
KW - Learning algorithms
KW - Natural language processing systems
KW - Process control
KW - Language games
KW - Learning paradigms
KW - Learning platform
KW - Literature reviews
KW - Machine-learning
KW - Natural languages
KW - Open platform communication unified architecture
KW - Open platforms
KW - Reinforcement learnings
KW - Unified architecture
KW - Reinforcement learning
U2 - 10.1109/INDIN51400.2023.10218289
DO - 10.1109/INDIN51400.2023.10218289
M3 - Conference contribution
SN - 978-1-6654-9314-7
BT - 2023 IEEE 21st International Conference on Industrial Informatics (INDIN)
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 21st IEEE International Conference on Industrial Informatics, INDIN 2023
Y2 - 18 July 2023 through 20 July 2023
ER -