{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T04:38:14Z","timestamp":1742963894091,"version":"3.40.3"},"publisher-location":"Cham","reference-count":50,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030788100"},{"type":"electronic","value":"9783030788117"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-78811-7_35","type":"book-chapter","created":{"date-parts":[[2021,7,6]],"date-time":"2021-07-06T23:22:37Z","timestamp":1625613757000},"page":"365-377","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Exploring the Landscapes and Emerging Trends of Reinforcement Learning from 1990 to 2020: A Bibliometric Analysis"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4219-788X","authenticated-orcid":false,"given":"Li","family":"Zeng","sequence":"first","affiliation":[]},{"given":"Xiaoqing","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zili","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,7,7]]},"reference":[{"key":"35_CR1","unstructured":"Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction (1988)"},{"issue":"2","key":"35_CR2","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","volume":"128","author":"RR Selvaraju","year":"2020","unstructured":"Selvaraju, R.R., Das, A., Vedantam, R., Cogswell, M., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 128(2), 336\u2013359 (2020)","journal-title":"Int. J. Comput. Vis."},{"key":"35_CR3","unstructured":"Dai, H., Khalil, E.B., Zhang, Y., Dilkina, B., Song, L.: Learning combinatorial optimization algorithms over graphs. In: NIPS 2017 Proceedings of the 31st International Conference on Neural Information Processing Systems, vol. 30, pp. 6351\u20136361 (2017)"},{"issue":"4","key":"35_CR4","doi-asserted-by":"publisher","first-page":"3133","DOI":"10.1109\/COMST.2019.2916583","volume":"21","author":"NC Luong","year":"2019","unstructured":"Luong, N.C., et al.: Applications of deep reinforcement learning in communications and networking: a survey. IEEE Commun. Surv. Tutor. 21(4), 3133\u20133174 (2019)","journal-title":"IEEE Commun. Surv. Tutor."},{"issue":"1","key":"35_CR5","first-page":"1334","volume":"17","author":"S Levine","year":"2016","unstructured":"Levine, S., Finn, C., Darrell, T., Abbeel, P.: End-to-end training of deep visuomotor policies. J. Mach. Learn. Res. 17(1), 1334\u20131373 (2016)","journal-title":"J. Mach. Learn. Res."},{"issue":"11","key":"35_CR6","doi-asserted-by":"publisher","first-page":"1238","DOI":"10.1177\/0278364913495721","volume":"32","author":"J Kober","year":"2013","unstructured":"Kober, J., Andrew Bagnell, J., Peters, J.: Reinforcement learning in robotics: a survey. Int. J. Robot. Res. 32(11), 1238\u20131274 (2013)","journal-title":"Int. J. Robot. Res."},{"issue":"3","key":"35_CR7","doi-asserted-by":"publisher","first-page":"278","DOI":"10.1061\/(ASCE)0733-947X(2003)129:3(278)","volume":"129","author":"B Abdulhai","year":"2003","unstructured":"Abdulhai, B., Pringle, R., Karakoulas, G.J.: Reinforcement learning for true adaptive traffic signal control. J. Transp. Eng.-ASCE 129(3), 278\u2013285 (2003)","journal-title":"J. Transp. Eng.-ASCE"},{"key":"35_CR8","doi-asserted-by":"publisher","first-page":"538","DOI":"10.1016\/j.apenergy.2017.11.072","volume":"211","author":"R Xiong","year":"2018","unstructured":"Xiong, R., Cao, J., Yu, Q.: Reinforcement learning-based real-time power management for hybrid energy storage system in the plug-in hybrid electric vehicle. Appl. Energy 211, 538\u2013548 (2018)","journal-title":"Appl. Energy"},{"issue":"1","key":"35_CR9","doi-asserted-by":"publisher","first-page":"48","DOI":"10.1186\/s13321-017-0235-x","volume":"9","author":"M Olivecrona","year":"2017","unstructured":"Olivecrona, M., Blaschke, T., Engkvist, O., Chen, H.: Molecular de-Novo design through deep reinforcement learning. J. Cheminformatics 9(1), 48 (2017)","journal-title":"J. Cheminformatics"},{"issue":"12","key":"35_CR10","doi-asserted-by":"publisher","first-page":"1337","DOI":"10.1021\/acscentsci.7b00492","volume":"3","author":"Z Zhou","year":"2017","unstructured":"Zhou, Z., Li, X., Zare, R.N.: Optimizing chemical reactions with deep reinforcement learning. ACS Cent. Sci. 3(12), 1337\u20131344 (2017)","journal-title":"ACS Cent. Sci."},{"key":"35_CR11","doi-asserted-by":"publisher","first-page":"317","DOI":"10.1038\/s42256-020-0177-2","volume":"2","author":"C Fan","year":"2020","unstructured":"Fan, C., Zeng, L., Sun, Y., Liu, Y.-Y.: Finding key players in complex networks through deep reinforcement learning. Nat. Mach. Intell. 2, 317\u2013324 (2020)","journal-title":"Nat. Mach. Intell."},{"issue":"1","key":"35_CR12","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1613\/jair.301","volume":"4","author":"LP Kaelbling","year":"1996","unstructured":"Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: a survey. J. Artif. Intell. Res. 4(1), 237\u2013285 (1996)","journal-title":"J. Artif. Intell. Res."},{"issue":"2","key":"35_CR13","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1109\/TSMCC.2007.913919","volume":"38","author":"L Busoniu","year":"2008","unstructured":"Busoniu, L., Babuska, R., De. Schutter, B.: A comprehensive survey of multiagent reinforcement learning. Syst. Man Cybern. 38(2), 156\u2013172 (2008)","journal-title":"Syst. Man Cybern."},{"key":"35_CR14","first-page":"348","volume":"25","author":"A Pritchard","year":"1969","unstructured":"Pritchard, A.: Statistical bibliography or bibliometrics. J. Documentation 25, 348 (1969)","journal-title":"J. Documentation"},{"key":"35_CR15","doi-asserted-by":"crossref","unstructured":"Fan, C., Zeng, L., Ding, Y., Chen, M., Sun, Y., Liu, Z.: Learning to identify high betweenness centrality nodes from scratch: a novel graph neural network approach. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 559\u2013568 (2019)","DOI":"10.1145\/3357384.3357979"},{"issue":"3159","key":"35_CR16","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1126\/science.122.3159.108","volume":"122","author":"E Garfield","year":"1955","unstructured":"Garfield, E.: Citation indexes for science: a new dimension in documentation through association of ideas. Science 122(3159), 108\u2013111 (1955)","journal-title":"Science"},{"issue":"3","key":"35_CR17","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1002\/asi.20317","volume":"57","author":"C Chen","year":"2006","unstructured":"Chen, C.: CiteSpace II: detecting and visualizing emerging trends and transient patterns in scientific literature. J. Am. Soc. Inf. Sci. 57(3), 359\u2013377 (2006)","journal-title":"J. Am. Soc. Inf. Sci."},{"issue":"2","key":"35_CR18","doi-asserted-by":"publisher","first-page":"523","DOI":"10.1007\/s11192-009-0146-3","volume":"84","author":"NJ van Eck","year":"2010","unstructured":"van Eck, N.J., Waltman, L.: Software survey: VOSviewer, a computer program for bibliometric mapping. Scientometrics 84(2), 523\u2013538 (2010)","journal-title":"Scientometrics"},{"issue":"4","key":"35_CR19","doi-asserted-by":"publisher","first-page":"373","DOI":"10.1023\/A:1024940629314","volume":"7","author":"J Kleinberg","year":"2003","unstructured":"Kleinberg, J.: Bursty and hierarchical structure in streams. Data Min. Knowl. Disc. 7(4), 373\u2013397 (2003)","journal-title":"Data Min. Knowl. Disc."},{"issue":"1","key":"35_CR20","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1007\/s11192-006-0144-7","volume":"69","author":"L Egghe","year":"2006","unstructured":"Egghe, L.: Theory and practise of the g-index. Scientometrics 69(1), 131\u2013152 (2006)","journal-title":"Scientometrics"},{"key":"35_CR21","unstructured":"Schvaneveldt, R.W.: Pathfinder Associative Networks: Studies in Knowledge Organization (1990)"},{"issue":"2","key":"35_CR22","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1080\/0022250X.2001.9990249","volume":"25","author":"U Brandes","year":"2001","unstructured":"Brandes, U.: A faster algorithm for betweenness centrality. J. Math. Sociol. 25(2), 163\u2013177 (2001)","journal-title":"J. Math. Sociol."},{"key":"35_CR23","doi-asserted-by":"publisher","first-page":"2197","DOI":"10.3389\/fpsyg.2019.02197","volume":"10","author":"SV Aryadoust","year":"2019","unstructured":"Aryadoust, S.V., Tan, H.A.H., Ng, L.Y.: A scientometric review of Rasch measurement: the rise and progress of a specialty. Front. Psychol. 10, 2197 (2019)","journal-title":"Front. Psychol."},{"key":"35_CR24","first-page":"849","volume":"14","author":"AY Ng","year":"2001","unstructured":"Ng, A.Y., Jordan, M.I., Weiss, Y.: On spectral clustering: analysis and an algorithm. Adv. Neural. Inf. Process. Syst. 14, 849\u2013856 (2001)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"1","key":"35_CR25","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1016\/0377-0427(87)90125-7","volume":"20","author":"P Rousseeuw","year":"1987","unstructured":"Rousseeuw, P.: Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20(1), 53\u201365 (1987)","journal-title":"J. Comput. Appl. Math."},{"issue":"6","key":"35_CR26","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1109\/MCS.2012.2214134","volume":"32","author":"FL Lewis","year":"2012","unstructured":"Lewis, F.L., Vrabie, D., Vamvoudakis, K.G.: Reinforcement learning and feedback control: using natural decision methods to design optimal adaptive controllers. IEEE Control Syst. Mag. 32(6), 76\u2013105 (2012)","journal-title":"IEEE Control Syst. Mag."},{"issue":"3","key":"35_CR27","doi-asserted-by":"publisher","first-page":"621","DOI":"10.1109\/TNNLS.2013.2281663","volume":"25","author":"D Liu","year":"2014","unstructured":"Liu, D., Wei, Q.: Policy iteration adaptive dynamic programming algorithm for discrete-time nonlinear systems. IEEE Trans. Neural Netw. 25(3), 621\u2013634 (2014)","journal-title":"IEEE Trans. Neural Netw."},{"issue":"1","key":"35_CR28","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.automatica.2012.09.019","volume":"49","author":"S Bhasin","year":"2013","unstructured":"Bhasin, S., Kamalapurkar, R., Johnson, M., Vamvoudakis, K.G., Lewis, F.L., Dixon, W.E.: A novel actor-critic-identifier architecture for approximate optimal control of uncertain nonlinear systems. Automatica 49(1), 82\u201392 (2013)","journal-title":"Automatica"},{"issue":"4","key":"35_CR29","doi-asserted-by":"publisher","first-page":"1167","DOI":"10.1016\/j.automatica.2014.02.015","volume":"50","author":"A Karimpour","year":"2014","unstructured":"Karimpour, A., Naghibi-Sistani, M.B.: Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics. Automatica 50(4), 1167\u20131175 (2014)","journal-title":"Automatica"},{"issue":"8","key":"35_CR30","doi-asserted-by":"publisher","first-page":"1598","DOI":"10.1016\/j.automatica.2012.05.074","volume":"48","author":"KG Vamvoudakis","year":"2012","unstructured":"Vamvoudakis, K.G., Lewis, F.L., Hudas, G.R.: Multi-agent differential graphical games: online adaptive learning solution for synchronization with optimality. Automatica 48(8), 1598\u20131611 (2012)","journal-title":"Automatica"},{"issue":"2","key":"35_CR31","doi-asserted-by":"publisher","first-page":"311","DOI":"10.1016\/0004-3702(92)90058-6","volume":"55","author":"S Mahadevan","year":"1992","unstructured":"Mahadevan, S., Connell, J.: Automatic programming of behavior-based robots using reinforcement learning. Artif. Intell. 55(2), 311\u2013365 (1992)","journal-title":"Artif. Intell."},{"issue":"5","key":"35_CR32","doi-asserted-by":"publisher","first-page":"674","DOI":"10.1109\/9.580874","volume":"42","author":"JN Tsitsiklis","year":"1997","unstructured":"Tsitsiklis, J.N., Van Roy, B.: An analysis of temporal-difference learning with function approximation. IEEE Trans. Autom. Control 42(5), 674\u2013690 (1997)","journal-title":"IEEE Trans. Autom. Control"},{"key":"35_CR33","doi-asserted-by":"publisher","first-page":"1185","DOI":"10.1162\/neco.1994.6.6.1185","volume":"6","author":"T Jaakkola","year":"1994","unstructured":"Jaakkola, T., Jordan, M.I., Singh, S.P.: On the convergence of stochastic iterative dynamic programming algorithms. Neural Comput. 6, 1185\u20131201 (1994)","journal-title":"Neural Comput."},{"issue":"2","key":"35_CR34","first-page":"279","volume":"23","author":"M Asada","year":"1996","unstructured":"Asada, M., Noda, S., Tawaratsumida, S., Hosoda, K.: Purposive behavior acquisition for a real robot by vision-based reinforcement learning. Mach. Learn. 23(2), 279\u2013303 (1996)","journal-title":"Mach. Learn."},{"issue":"1","key":"35_CR35","first-page":"59","volume":"22","author":"JN Tsitsiklis","year":"1996","unstructured":"Tsitsiklis, J.N., van Roy, B.: Feature-based methods for large scale dynamic programming. Mach. Learn. 22(1), 59\u201394 (1996)","journal-title":"Mach. Learn."},{"key":"35_CR36","doi-asserted-by":"crossref","unstructured":"Gu, S., Holly, E., Lillicrap, T., Levine, S.: Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 3389\u20133396 (2017)","DOI":"10.1109\/ICRA.2017.7989385"},{"key":"35_CR37","doi-asserted-by":"crossref","unstructured":"Todorov, E., Erez, T., Tassa, Y.: MuJoCo: a physics engine for model-based control. In: 2012 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 5026\u20135033 (2012)","DOI":"10.1109\/IROS.2012.6386109"},{"key":"35_CR38","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Mak, B.: Speeding up softmax computations in DNN-based large vocabulary speech recognition by senone weight vector selection. In: 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5335\u20135339 (2017)","DOI":"10.1109\/ICASSP.2017.7953175"},{"issue":"4\u20135","key":"35_CR39","first-page":"421","volume":"37","author":"S Levine","year":"2017","unstructured":"Levine, S., Pastor, P., Krizhevsky, A., Ibarz, J., Quillen, D.: Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int. J. Robot. Res. 37(4\u20135), 421\u2013436 (2017)","journal-title":"Int. J. Robot. Res."},{"key":"35_CR40","doi-asserted-by":"crossref","unstructured":"Tobin, J., Fong, R., Ray, A., Schneider, J., Zaremba, W., Abbeel, P.: Domain randomization for transferring deep neural networks from simulation to the real world. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 23\u201330 (2017)","DOI":"10.1109\/IROS.2017.8202133"},{"key":"35_CR41","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1016\/j.apenergy.2018.03.104","volume":"222","author":"J Wu","year":"2018","unstructured":"Wu, J., He, H., Peng, J., Li, Y., Li, Z.: Continuous reinforcement learning of energy management with deep Q network for a power split hybrid electric bus. Appl. Energy 222, 799\u2013811 (2018)","journal-title":"Appl. Energy"},{"issue":"3","key":"35_CR42","doi-asserted-by":"publisher","first-page":"247","DOI":"10.1109\/JAS.2016.7508798","volume":"3","author":"L Li","year":"2016","unstructured":"Li, L., Lv, Y., Wang, F.-Y.: Traffic signal timing via deep reinforcement learning. IEEE\/CAA J. Automatica Sinica 3(3), 247\u2013254 (2016)","journal-title":"IEEE\/CAA J. Automatica Sinica"},{"issue":"4","key":"35_CR43","doi-asserted-by":"publisher","first-page":"3698","DOI":"10.1109\/TSG.2018.2834219","volume":"10","author":"E Mocanu","year":"2019","unstructured":"Mocanu, E., et al.: On-line building energy optimization using deep reinforcement learning. IEEE Trans. Smart Grid 10(4), 3698\u20133708 (2019)","journal-title":"IEEE Trans. Smart Grid"},{"issue":"3","key":"35_CR44","doi-asserted-by":"publisher","first-page":"1140","DOI":"10.1109\/TITS.2013.2255286","volume":"14","author":"S El-Tantawy","year":"2013","unstructured":"El-Tantawy, S., Abdulhai, B., Abdelgawad, H.: Multiagent reinforcement learning for integrated network of adaptive traffic signal controllers (MARLIN-ATSC): methodology and large-scale application on downtown Toronto. IEEE Trans. Intell. Transp. Syst. 14(3), 1140\u20131150 (2013)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"5","key":"35_CR45","doi-asserted-by":"publisher","first-page":"2149","DOI":"10.1109\/TSG.2016.2517211","volume":"8","author":"F Ruelens","year":"2017","unstructured":"Ruelens, F., Claessens, B.J., Vandael, S., De Schutter, B., Babuska, R., Belmans, R.: Residential demand response of thermostatically controlled loads using batch reinforcement learning. IEEE Trans. Smart Grid 8(5), 2149\u20132159 (2017)","journal-title":"IEEE Trans. Smart Grid"},{"issue":"3","key":"35_CR46","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1109\/MCAS.2009.933854","volume":"9","author":"FL Lewis","year":"2009","unstructured":"Lewis, F.L., Vrabie, D.: Reinforcement learning and adaptive dynamic programming for feedback control. IEEE Circuits Syst. Mag. 9(3), 32\u201350 (2009)","journal-title":"IEEE Circuits Syst. Mag."},{"issue":"4","key":"35_CR47","doi-asserted-by":"publisher","first-page":"943","DOI":"10.1109\/TSMCB.2008.926614","volume":"38","author":"A Al-Tamimi","year":"2008","unstructured":"Al-Tamimi, A., Lewis, F.L., Abu-Khalaf, M.: Discrete-time nonlinear HJB solution using approximate dynamic programming: convergence proof. Syst. Man Cybern. 38(4), 943\u2013949 (2008)","journal-title":"Syst. Man Cybern."},{"issue":"5","key":"35_CR48","doi-asserted-by":"publisher","first-page":"878","DOI":"10.1016\/j.automatica.2010.02.018","volume":"46","author":"KG Vamvoudakis","year":"2010","unstructured":"Vamvoudakis, K.G., Lewis, F.L.: Online actor-critic algorithm to solve the continuous-time infinite horizon optimal control problem. Automatica 46(5), 878\u2013888 (2010)","journal-title":"Automatica"},{"issue":"12","key":"35_CR49","doi-asserted-by":"publisher","first-page":"2226","DOI":"10.1109\/TNN.2011.2168538","volume":"22","author":"H Zhang","year":"2011","unstructured":"Zhang, H., Cui, L., Zhang, X., Luo, Y.: Data-driven robust approximate optimal tracking control for unknown general nonlinear systems using adaptive dynamic programming method. IEEE Trans. Neural Netw. 22(12), 2226\u20132236 (2011)","journal-title":"IEEE Trans. Neural Netw."},{"issue":"11","key":"35_CR50","doi-asserted-by":"publisher","first-page":"3051","DOI":"10.1109\/TAC.2014.2317301","volume":"59","author":"H Modares","year":"2014","unstructured":"Modares, H., Lewis, F.L.: Linear quadratic tracking control of partially-unknown continuous-time systems using reinforcement learning. IEEE Trans. Autom. Control 59(11), 3051\u20133056 (2014)","journal-title":"IEEE Trans. Autom. Control"}],"container-title":["Lecture Notes in Computer Science","Advances in Swarm Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-78811-7_35","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,7,6]],"date-time":"2021-07-06T23:25:57Z","timestamp":1625613957000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-78811-7_35"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030788100","9783030788117"],"references-count":50,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-78811-7_35","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"7 July 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICSI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Swarm Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Qingdao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 July 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 July 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"swarm2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.iasei.org\/icsi2021\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"177","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"104","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"59% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,5","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4-5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}