{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,11]],"date-time":"2026-05-11T20:52:59Z","timestamp":1778532779784,"version":"3.51.4"},"reference-count":164,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"3","license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61671089"],"award-info":[{"award-number":["61671089"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Commun. Surv. Tutorials"],"published-print":{"date-parts":[[2020]]},"DOI":"10.1109\/comst.2020.2988367","type":"journal-article","created":{"date-parts":[[2020,4,16]],"date-time":"2020-04-16T20:25:06Z","timestamp":1587068706000},"page":"1722-1760","source":"Crossref","is-referenced-by-count":254,"title":["Deep Reinforcement Learning for Autonomous Internet of Things: Model, Applications and Challenges"],"prefix":"10.1109","volume":"22","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7828-0415","authenticated-orcid":false,"given":"Lei","family":"Lei","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8369-2521","authenticated-orcid":false,"given":"Yue","family":"Tan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8531-6762","authenticated-orcid":false,"given":"Kan","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Shiwen","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4262-153X","authenticated-orcid":false,"given":"Kuan","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4140-287X","authenticated-orcid":false,"given":"Xuemin","family":"Shen","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","author":"lillicrap","year":"2015","journal-title":"Continuous control with deep reinforcement learning"},{"key":"ref38","author":"haarnoja","year":"2018","journal-title":"Soft actor-critic Off-policy maximum entropy deep reinforcement learning with a stochastic actor"},{"key":"ref33","first-page":"387","article-title":"Deterministic policy gradient algorithms","author":"silver","year":"2014","journal-title":"Proc Int Conf Mach Learn (ICML)"},{"key":"ref32","first-page":"1889","article-title":"Trust region policy optimization","author":"schulman","year":"2015","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref31","first-page":"1531","article-title":"A natural policy gradient","author":"kakade","year":"2002","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1162\/089976698300017746"},{"key":"ref37","first-page":"1928","article-title":"Asynchronous methods for deep reinforcement learning","author":"mnih","year":"2016","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref36","author":"wang","year":"2016","journal-title":"Sample efficient actor&#x2013;critic with experience replay"},{"key":"ref35","first-page":"1008","article-title":"Actor&#x2013;critic algorithms","author":"konda","year":"2000","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992696"},{"key":"ref28","author":"schaul","year":"2015","journal-title":"Prioritized experience replay"},{"key":"ref27","first-page":"2094","article-title":"Deep reinforcement learning with double Q-learning","author":"van hasselt","year":"2016","journal-title":"Proc 13th AAAI Conf Artif Intell"},{"key":"ref29","author":"wang","year":"2015","journal-title":"Dueling network architectures for deep reinforcement learning"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2916583"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2943405"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2947542"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2018.1800109"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2019.2921977"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4757-6451-2_4"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1561\/2200000071"},{"key":"ref100","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2018.8461113"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2017.2743240"},{"key":"ref50","first-page":"29","article-title":"Deep recurrent Q-learning for partially observable MDPs","author":"hausknecht","year":"2015","journal-title":"Proc AAAI Fall Symp Series"},{"key":"ref51","author":"wayne","year":"2018","journal-title":"Unsupervised Predictive Memory in a Goal-Directed Agent"},{"key":"ref154","first-page":"1446","article-title":"Strategy learning for autonomous agents in smart grid markets","author":"reddy","year":"2011","journal-title":"Proc 22nd Int Joint Conf Artif Intell"},{"key":"ref153","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2015.2495145"},{"key":"ref156","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2802704"},{"key":"ref155","doi-asserted-by":"publisher","DOI":"10.1109\/TPWRS.2018.2823641"},{"key":"ref150","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-67540-4_12"},{"key":"ref152","doi-asserted-by":"publisher","DOI":"10.1109\/INFCOMW.2014.6849306"},{"key":"ref151","author":"xiao","year":"2018","journal-title":"Reinforcement learning-based energy trading for microgrids"},{"key":"ref146","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2016.2640184"},{"key":"ref147","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2018.03.072"},{"key":"ref148","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2018.12.061"},{"key":"ref149","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2013.12.013"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2016.2615643"},{"key":"ref58","first-page":"2974","article-title":"Counterfactual multi-agent policy gradients","author":"foerster","year":"2018","journal-title":"Proc 32nd AAAI Conf Artif Intell"},{"key":"ref57","first-page":"1","article-title":"Coordinated deep reinforcement learners for traffic light control","author":"van der pol","year":"2016","journal-title":"Proc Learn Inference Multiagent Syst NIPS"},{"key":"ref56","first-page":"1146","article-title":"Stabilising experience replay for deep multi-agent reinforcement learning","volume":"70","author":"foerster","year":"2017","journal-title":"Proc 34th Int Conf Mach Learn"},{"key":"ref55","doi-asserted-by":"crossref","first-page":"156","DOI":"10.1109\/TSMCC.2007.913919","article-title":"A comprehensive survey of multiagent reinforcement learning","volume":"38","author":"bu","year":"2008","journal-title":"IEEE Trans Syst Man Cybern C Appl Rev"},{"key":"ref54","author":"foerster","year":"2016","journal-title":"Learning to communicate to solve riddles with deep distributed recurrent q-networks"},{"key":"ref53","author":"zhu","year":"2018","journal-title":"On improving deep reinforcement learning for pomdps"},{"key":"ref52","article-title":"Deep reinforcement learning with POMDPs","author":"egorov","year":"2015"},{"key":"ref40","author":"barth-maron","year":"2018","journal-title":"Distributed Distributional Deterministic Policy Gradients"},{"key":"ref164","first-page":"237","article-title":"Gradient flow in recurrent nets: The difficulty of learning long-term dependencies","author":"hochreiter","year":"2001","journal-title":"A Field Guide to Dynamical Recurrent Neural Networks"},{"key":"ref163","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref162","first-page":"2140","article-title":"Playing FPS games with deep reinforcement learning","author":"lample","year":"2017","journal-title":"Proc 31st AAAI Conf Artif Intell"},{"key":"ref161","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2010.5650345"},{"key":"ref160","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-008-9056-7"},{"key":"ref4","author":"sutton","year":"2018","journal-title":"Reinforcement Learning An Introduction"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2018.2844341"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2017.2773600"},{"key":"ref5","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","article-title":"Human-level control through deep reinforcement learning","volume":"518","author":"mnih","year":"2015","journal-title":"Nature"},{"key":"ref159","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2003.809799"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.dcan.2017.10.002"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-74690-4_71"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2019.2893866"},{"key":"ref157","doi-asserted-by":"publisher","DOI":"10.1109\/TGCN.2017.2675542"},{"key":"ref158","author":"igl","year":"2018","journal-title":"Deep variational reinforcement learning for POMDPs"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/s13042-018-0834-5"},{"key":"ref46","first-page":"5279","article-title":"Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation","author":"wu","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref45","author":"schulman","year":"2017","journal-title":"Proximal policy optimization algorithms"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2013.06.002"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1007\/s10458-012-9200-2"},{"key":"ref42","first-page":"6379","article-title":"Multi-agent actor&#x2013;critic for mixed cooperative-competitive environments","author":"lowe","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref41","author":"fujimoto","year":"2018","journal-title":"Addressing function approximation error in actor&#x2013;critic methods"},{"key":"ref44","author":"gu","year":"2016","journal-title":"Q-prop Sample-efficient policy gradient with an off-policy critic"},{"key":"ref43","author":"heess","year":"2015","journal-title":"Memory-based control with recurrent neural networks"},{"key":"ref127","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.376"},{"key":"ref126","author":"kendall","year":"2018","journal-title":"Learning to drive in a day"},{"key":"ref125","doi-asserted-by":"publisher","DOI":"10.5220\/0007520305640572"},{"key":"ref124","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2883762"},{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2019.2921159"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1109\/ICCW.2019.8756797"},{"key":"ref129","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2872122"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2872440"},{"key":"ref128","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2016.7513867"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2016.2558146"},{"key":"ref76","doi-asserted-by":"publisher","DOI":"10.1016\/j.dcan.2018.10.003"},{"key":"ref130","author":"ye","year":"2017","journal-title":"Machine learning for vehicular networks"},{"key":"ref77","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2019.2928811"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1109\/GLOCOM.2018.8647178"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/ICC.2019.8761385"},{"key":"ref133","doi-asserted-by":"publisher","DOI":"10.1016\/j.adhoc.2018.07.016"},{"key":"ref134","doi-asserted-by":"publisher","DOI":"10.1109\/MWC.2019.1700441"},{"key":"ref131","doi-asserted-by":"publisher","DOI":"10.4018\/978-1-5225-0773-4.ch015"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2019.2935543"},{"key":"ref132","doi-asserted-by":"publisher","DOI":"10.1109\/PERCOMW.2017.7917508"},{"key":"ref79","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2019.2924015"},{"key":"ref136","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2015.2461134"},{"key":"ref135","first-page":"1","article-title":"Deep reinforcement learning solutions for energy microgrids management","author":"fran\u00e7ois-lavet","year":"2016","journal-title":"Proceedings of the 8th European Workshop on Reinforcement Learning (EWRL)"},{"key":"ref138","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2018.2859821"},{"key":"ref137","first-page":"36","article-title":"Reinforcement learning-based battery energy management in a solar microgrid","volume":"2","author":"mbuwir","year":"2017","journal-title":"Energy-Open"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2019.2949997"},{"key":"ref139","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2016.2514358"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.2019.2920390"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICON.2006.302580"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2019.2925719"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2017.2759728"},{"key":"ref140","doi-asserted-by":"publisher","DOI":"10.3390\/en12122291"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/WAINA.2017.67"},{"key":"ref141","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2018.2834219"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2018.06.236"},{"key":"ref142","doi-asserted-by":"publisher","DOI":"10.1109\/TSG.2015.2396993"},{"key":"ref67","author":"leong","year":"2018","journal-title":"Deep reinforcement learning for wireless sensor scheduling in cyber-physical systems"},{"key":"ref143","doi-asserted-by":"publisher","DOI":"10.1109\/SMARTGRID.2010.5622078"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682697"},{"key":"ref144","author":"claessens","year":"2016","journal-title":"Convolutional neural networks for automatic state-time feature extraction in reinforcement learning applied to residential load control"},{"key":"ref2","year":"2018","journal-title":"Smarter Things The Autonomous IoT"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/WCNC.2018.8377263"},{"key":"ref145","doi-asserted-by":"publisher","DOI":"10.1109\/PSCC.2014.7038106"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/37.88585"},{"key":"ref109","author":"wu","year":"2017","journal-title":"Flow Architecture and benchmarking for reinforcement learning in traffic control"},{"key":"ref95","first-page":"651","article-title":"Scalable deep reinforcement learning for vision-based robotic manipulation","author":"kalashnikov","year":"2018","journal-title":"Proc Conf Robot Learn"},{"key":"ref108","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC.2018.8569448"},{"key":"ref94","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989385"},{"key":"ref107","article-title":"CARMA: A deep reinforcement learning approach to autonomous driving","author":"vitelli","year":"2016"},{"key":"ref93","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2805379"},{"key":"ref106","article-title":"Deep reinforcement learning for simulated autonomous vehicle control","author":"yu","year":"2016"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.1109\/ICRAE.2017.8291355"},{"key":"ref105","doi-asserted-by":"publisher","DOI":"10.3390\/robotics7030047"},{"key":"ref91","doi-asserted-by":"publisher","DOI":"10.1109\/BDAI.2018.8546675"},{"key":"ref104","first-page":"292","article-title":"A survey on multiagent reinforcement learning towards multi-robot systems","author":"yang","year":"2005","journal-title":"Proc IEEE Conf Comput Intell Games (CIG)"},{"key":"ref90","doi-asserted-by":"publisher","DOI":"10.1109\/ICIAICT.2019.8784847"},{"key":"ref103","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2814606"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967908"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.1109\/IVS.2018.8500675"},{"key":"ref112","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC.2018.8569484"},{"key":"ref110","first-page":"1","article-title":"Reinforcement learning based driving speed control for two vehicle scenario","author":"gamage","year":"2017","journal-title":"Proc Australian Transp Res Forum (ATRF)"},{"key":"ref98","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2009.5354406"},{"key":"ref99","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-05816-6_3"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.1016\/j.robot.2018.11.004"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.1109\/IRC.2018.00038"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3390\/fi11040094"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TETCI.2019.2907718"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2926642"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2894819"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.17775\/CSEEJPES.2018.00520"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2916177"},{"key":"ref118","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2017.2760281"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2014.2320099"},{"key":"ref82","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2019.2913162"},{"key":"ref117","doi-asserted-by":"publisher","DOI":"10.1109\/TWC.2019.2900035"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2926625"},{"key":"ref81","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2018.2890685"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TCOMM.2019.2924010"},{"key":"ref84","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2919736"},{"key":"ref119","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2018.2832219"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2018.2846401"},{"key":"ref83","doi-asserted-by":"publisher","DOI":"10.1109\/TETC.2019.2902661"},{"key":"ref114","doi-asserted-by":"publisher","DOI":"10.1109\/MMAR.2013.6669888"},{"key":"ref113","doi-asserted-by":"publisher","DOI":"10.1109\/ITSC.2017.8317735"},{"key":"ref116","doi-asserted-by":"publisher","DOI":"10.1109\/IWCMC.2018.8450518"},{"key":"ref80","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2876279"},{"key":"ref115","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2014.01.007"},{"key":"ref120","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2019.2894437"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.1109\/CAC.2017.8244061"},{"key":"ref121","author":"qi","year":"2018","journal-title":"Vehicular Edge Computing via Deep Reinforcement Learning"},{"key":"ref122","doi-asserted-by":"publisher","DOI":"10.1109\/TIE.2015.2482119"},{"key":"ref123","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2019.2935450"},{"key":"ref85","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2019.2906789"},{"key":"ref86","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2882583"},{"key":"ref87","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2018.2878435"},{"key":"ref88","doi-asserted-by":"publisher","DOI":"10.23919\/SICE.2017.8105597"}],"container-title":["IEEE Communications Surveys &amp; Tutorials"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9739\/9173582\/09069178.pdf?arnumber=9069178","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T16:58:58Z","timestamp":1651078738000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9069178\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"references-count":164,"journal-issue":{"issue":"3"},"URL":"https:\/\/doi.org\/10.1109\/comst.2020.2988367","relation":{},"ISSN":["1553-877X","2373-745X"],"issn-type":[{"value":"1553-877X","type":"electronic"},{"value":"2373-745X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]}}}