{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T01:36:37Z","timestamp":1760232997272,"version":"build-2065373602"},"reference-count":26,"publisher":"MDPI AG","issue":"6","license":[{"start":{"date-parts":[[2022,12,9]],"date-time":"2022-12-09T00:00:00Z","timestamp":1670544000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Robotics"],"abstract":"<jats:p>Many robotics systems carrying certain payloads are employed in manufacturing industries for pick and place tasks. The system experiences inefficiency if more or less weight is introduced. If a different payload is introduced (either due to a change in the load or a change in the parameters of the robot system), the robot must be re-trained with the new weight\/parameters and the new network must be trained. Parameters such as the robot weight, length of limbs, or new payload may vary for an agent depending on the circumstance. Parameter changes pose a problem to the agent in achieving the same goal it is expected to achieve with the original parameters. Hence, it becomes mandatory to re-train the agent with the new parameters in order for it to achieve its goal. This research proposes a novel framework for the adaption of varying conditions on a robot agent in a given simulated environment without any retraining. Utilizing the properties of Generative Adversarial Network (GAN), the agent is able to train only once with reinforcement learning and by tweaking the noise vector of the generator in the GAN network, the agent can adapt to new conditions accordingly and demonstrate similar performance as if it were trained with the new physical attributes using reinforcement learning. A simple CartPole environment is considered for the experimentation, and it is shown that with the propose approached the agent remains stable for more iterations. The approach can be extended to the real world in the future.<\/jats:p>","DOI":"10.3390\/robotics11060150","type":"journal-article","created":{"date-parts":[[2022,12,12]],"date-time":"2022-12-12T01:42:27Z","timestamp":1670809347000},"page":"150","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Integrating the Generative Adversarial Network for Decision Making in Reinforcement Learning for Industrial Robot Agents"],"prefix":"10.3390","volume":"11","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9258-0855","authenticated-orcid":false,"given":"Neelabh","family":"Paul","sequence":"first","affiliation":[{"name":"Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune 412115, India"}]},{"given":"Vaibhav","family":"Tasgaonkar","sequence":"additional","affiliation":[{"name":"Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune 412115, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1745-5231","authenticated-orcid":false,"given":"Rahee","family":"Walambe","sequence":"additional","affiliation":[{"name":"Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune 412115, India"},{"name":"Symbiosis Centre of Applied Artificial Intelligence, Symbiosis International (Deemed University), Pune 412115, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2653-3780","authenticated-orcid":false,"given":"Ketan","family":"Kotecha","sequence":"additional","affiliation":[{"name":"Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune 412115, India"},{"name":"Symbiosis Centre of Applied Artificial Intelligence, Symbiosis International (Deemed University), Pune 412115, India"}]}],"member":"1968","published-online":{"date-parts":[[2022,12,9]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"583","DOI":"10.1108\/IR-07-2016-0194","article-title":"Growth in e-commerce boosts innovation in the warehouse robot market","volume":"43","author":"Bogue","year":"2016","journal-title":"Ind. Robot. Int. J."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1016\/j.jretconser.2014.10.010","article-title":"U.S. online shopping: Facts, fiction, hopes and dreams","volume":"23","author":"Schultz","year":"2015","journal-title":"J. Retail. Consum. Serv."},{"key":"ref_3","unstructured":"(2021, November 05). Amazon Prime and \u201cFree\u201d Shipping. Available online: https:\/\/escholarship.org\/uc\/item\/0681j9rr."},{"key":"ref_4","first-page":"63","article-title":"The Impact of Warehouse Automation in Amazon\u2019s Success","volume":"7","author":"Laber","year":"2020","journal-title":"IJISET-Int. J. Innov. Sci. Eng. Technol."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Bouman, A., Ginting, M.F., Alatur, N., Palieri, M., Fan, D.D., Touma, T., Pailevanian, T., Kim, S.K., Otsu, K., and Burdick, J. (2020\u201324, January 24). Autonomous spot: Long-range autonomous exploration of extreme environments with legged locomotion. Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA.","DOI":"10.1109\/IROS45743.2020.9341361"},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Van Hasselt, H., and Wiering, M.A. (2007, January 1\u20135). Reinforcement Learning in Continuous Action. Proceedings of the 2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning, Honolulu, HI, USA. Available online: http:\/\/ieeexplore.ieee.org\/lpdocs\/epic03\/wrapper.htm?arnumber=4220844.","DOI":"10.1109\/ADPRL.2007.368199"},{"key":"ref_7","unstructured":"Mahmood, A.R., Korenkevych, D., Vasan, G., Ma, W., and Bergstra, J. (2018, January 29\u201331). Benchmarking Reinforcement Learning Algorithms on Real-World Robots. Proceedings of the 2nd Conference on Robot Learning, Zurich, Switzerland. Available online: https:\/\/proceedings.mlr.press\/v87\/mahmood18a.html."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","article-title":"Generative adversarial networks","volume":"63","author":"Goodfellow","year":"2020","journal-title":"Commun. ACM"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Tian, Y., Wang, Q., Huang, Z., Li, W., Dai, D., Yang, M., Wang, J., Fink, O., Z\u00fcrich, E., and Europe, N. (2020, January 23\u201328). Off-Policy Reinforcement Learning for Efficient and Effective GAN Architecture Search (Supplementary Material). Proceedings of the 16th European Conference on Computer Vision (ECCV 2020), Glasgow, UK.","DOI":"10.1007\/978-3-030-58571-6_11"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Sarmad, M., Korea, S., Lee, H.J., and Kim, Y.M. (2019, January 15\u201320). RL-GAN-Net: A Reinforcement Learning Agent Controlled GAN Network for Real-Time Point Cloud Shape Completion. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00605"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"115680","DOI":"10.1016\/j.eswa.2021.115680","article-title":"Improving exploration efficiency of deep reinforcement learning through samples produced by generative model","volume":"185","author":"Xu","year":"2021","journal-title":"Expert Syst. Appl."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Rao, K., Harris, C., Irpan, A., Levine, S., Ibarz, J., and Khansari, M. (2020, January 13\u201319). RL-CycleGAN: Reinforcement Learning Aware Simulation-To-Real. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.01117"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"884","DOI":"10.1109\/TCOMM.2020.3031930","article-title":"Experienced Deep Reinforcement Learning with Generative Adversarial Networks (GANs) for Model-Free Ultra Reliable Low Latency Communication","volume":"69","author":"Kasgari","year":"2021","journal-title":"IEEE Trans. Commun."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"3545","DOI":"10.1109\/LRA.2021.3063927","article-title":"Human-guided Robot Behavior Learning: A GAN-assisted Preference-based Reinforcement Learning Approach","volume":"6","author":"Zhan","year":"2020","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_15","unstructured":"Watkins, C.J.C.H., and Dayan, P. (1992). Q-Learning, Kluwer Academic Publisher."},{"key":"ref_16","unstructured":"(2021, October 31). Image Synthesis\u2014Noise Generation. Available online: https:\/\/homepages.inf.ed.ac.uk\/rbf\/HIPR2\/noise.htm."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"1721","DOI":"10.1016\/j.eswa.2012.09.010","article-title":"Neural network Reinforcement Learning for visual control of robot manipulators","volume":"40","year":"2013","journal-title":"Expert Syst. Appl."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"104","DOI":"10.1016\/j.eswa.2016.06.021","article-title":"Neural networks based reinforcement learning for mobile robots obstacle avoidance","volume":"62","author":"Duguleana","year":"2016","journal-title":"Expert Syst. Appl."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"9849","DOI":"10.1109\/TVT.2020.3002983","article-title":"Multi-Agent Deep Reinforcement Learning-Based Flexible Satellite Payload for Mobile Terminals","volume":"69","author":"Hu","year":"2020","journal-title":"IEEE Trans. Veh. Technol."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Kim, P. (2017). Convolutional Neural Network. MATLAB Deep Learning, Springer.","DOI":"10.1007\/978-1-4842-2845-6"},{"key":"ref_21","unstructured":"Radford, A., Metz, L., and Chintala, S. (2016). Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv."},{"key":"ref_22","unstructured":"Brockman, G., Cheung, V., Pettersson, L., Schneider, J., Schulman, J., Tang, J., and Zaremba, W. (2016). OpenAI Gym. arXiv."},{"key":"ref_23","unstructured":"Bjorck, J., Gomes, C., Selman, B., and Weinberger, K.Q. (2018, January 3\u20138). Understanding Batch Normalization. Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada."},{"key":"ref_24","unstructured":"Fred Agarap, A.M. (2021, July 19). Deep Learning using Rectified Linear Units (ReLU). Available online: https:\/\/github.com\/AFAgarap\/relu-classifier."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"873","DOI":"10.1007\/978-981-13-6772-4_76","article-title":"Comparative Study of Convolution Neural Network\u2019s Relu and Leaky-Relu Activation Functions","volume":"Volume 553","author":"Dubey","year":"2019","journal-title":"Applications of Computing, Automation and Wireless Systems in Electrical Engineering"},{"key":"ref_26","unstructured":"(2021, May 30). Statistical Analysis Based on a Certain Multivariate Complex Gaussian Distribution (An Introduction) on JSTOR. Available online: https:\/\/www.jstor.org\/stable\/2991290?seq=1."}],"container-title":["Robotics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2218-6581\/11\/6\/150\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T01:37:33Z","timestamp":1760146653000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2218-6581\/11\/6\/150"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,9]]},"references-count":26,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2022,12]]}},"alternative-id":["robotics11060150"],"URL":"https:\/\/doi.org\/10.3390\/robotics11060150","relation":{},"ISSN":["2218-6581"],"issn-type":[{"type":"electronic","value":"2218-6581"}],"subject":[],"published":{"date-parts":[[2022,12,9]]}}}