{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T04:11:28Z","timestamp":1746072688247,"version":"3.40.4"},"publisher-location":"Cham","reference-count":55,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031908996","type":"print"},{"value":"9783031909009","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"vor","delay-in-days":120,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Tabular reinforcement learning methods cannot operate directly on continuous state spaces. One solution to this problem is to partition the state space. A good partitioning enables generalization during learning and more efficient exploitation of prior experiences. Consequently, the learning process becomes faster and produces more reliable policies. However, partitioning introduces approximation, which is particularly harmful in the presence of nonlinear relations between state components. An ideal partition should be as coarse as possible, while capturing the key structure of the state space for the given problem. This work extracts partitions from the environment dynamics by symbolic execution. We show that symbolic partitioning improves state space coverage with respect to environmental behavior and allows reinforcement learning to perform better for sparse rewards. We evaluate symbolic state space partitioning with respect to precision, scalability, learning agent performance and state space coverage for the learned policies.<\/jats:p>","DOI":"10.1007\/978-3-031-90900-9_7","type":"book-chapter","created":{"date-parts":[[2025,4,30]],"date-time":"2025-04-30T04:44:44Z","timestamp":1745988284000},"page":"122-144","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Symbolic State Partitioning for Reinforcement Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1939-9053","authenticated-orcid":false,"given":"Mohsen","family":"Ghaffari","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4776-883X","authenticated-orcid":false,"given":"Mahsa","family":"Varshosaz","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5382-3949","authenticated-orcid":false,"given":"Einar Broch","family":"Johnsen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0532-2685","authenticated-orcid":false,"given":"Andrzej","family":"W\u0105sowski","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,1]]},"reference":[{"key":"7_CR1","doi-asserted-by":"publisher","unstructured":"Adelt, J., Herber, P., Niehage, M., Remke, A.: Towards safe and resilient hybrid systems in the presence of learning and uncertainty. In: Proc. 11th Intl. Symposium on Leveraging Applications of Formal Methods, Verification and Validation. Verification Principles (ISoLA 2022). Lecture Notes in Computer Science, vol. 13701, pp. 299\u2013319. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-19849-6_18","DOI":"10.1007\/978-3-031-19849-6_18"},{"key":"7_CR2","doi-asserted-by":"publisher","unstructured":"Ahrendt, W., Beckert, B., Bubel, R., H\u00e4hnle, R., Schmitt, P.H., Ulbrich, M. (eds.): Deductive Software Verification - The KeY Book - From Theory to Practice, Lecture Notes in Computer Science, vol. 10001. Springer (2016). https:\/\/doi.org\/10.1007\/978-3-319-49812-6","DOI":"10.1007\/978-3-319-49812-6"},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Akrour, R., Veiga, F., Peters, J., Neumann, G.: Regularizing reinforcement learning with state abstraction. In: Proc. Intl. Conf. on Intelligent Robots and Systems (IROS). pp. 534\u2013539. IEEE (2018)","DOI":"10.1109\/IROS.2018.8594201"},{"key":"7_CR4","unstructured":"Albus, J.S.: Brains, behavior, and robotics. BYTE Books (1981)"},{"key":"7_CR5","doi-asserted-by":"publisher","unstructured":"de\u00a0Boer, F.S., Bonsangue, M.M.: Symbolic execution formally explained. Formal Aspects Comput. 33(4-5), 617\u2013636 (2021). https:\/\/doi.org\/10.1007\/S00165-020-00527-Y","DOI":"10.1007\/S00165-020-00527-Y"},{"key":"7_CR6","unstructured":"Cadar, C., Dunbar, D., Engler, D.R.: KLEE: unassisted and automatic generation of high-coverage tests for complex systems programs. In: Proc. 8th Symposium on Operating Systems Design and Implementation (OSDI 2008). pp. 209\u2013224. USENIX Association (2008), http:\/\/www.usenix.org\/events\/osdi08\/tech\/full_papers\/cadar\/cadar.pdf"},{"key":"7_CR7","doi-asserted-by":"crossref","unstructured":"Chen, X., Wang, H., Li, Z., Ding, W., Dang, F., Wu, C., Chen, X.: Deliversense: Efficient delivery drone scheduling for crowdsensing with deep reinforcement learning. In: Adjunct Proceedings of the 2022 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the 2022 ACM International Symposium on Wearable Computers. pp. 403\u2013408 (2022)","DOI":"10.1145\/3544793.3560412"},{"key":"7_CR8","doi-asserted-by":"publisher","unstructured":"Clarke, L.A.: A program testing system. In: Proc. 1976 Annual Conf. pp. 488\u2013491. ACM (1976). https:\/\/doi.org\/10.1145\/800191.805647","DOI":"10.1145\/800191.805647"},{"key":"7_CR9","unstructured":"Dadvar, M., Nayyar, R.K., Srivastava, S.: Conditional abstraction trees for sample-efficient reinforcement learning. In: Proc. 39th Conf. on Uncertainty in Artificial Intelligence. Proc. Machine Learning Research, vol.\u00a0216, pp. 485\u2013495. PMLR (2023)"},{"key":"7_CR10","unstructured":"Davey, B.A., Priestley, H.A.: Introduction to lattices and order. Cambridge University Press, Cambridge (1990), http:\/\/www.worldcat.org\/search?qt=worldcat_org_all&q=0521367662"},{"key":"7_CR11","unstructured":"Ferns, N., Panangaden, P., Precup, D.: Metrics for finite Markov decision processes. In: UAI. vol.\u00a04, pp. 162\u2013169 (2004)"},{"key":"7_CR12","doi-asserted-by":"crossref","unstructured":"Ferns, N., Panangaden, P., Precup, D.: Bisimulation metrics for continuous Markov decision processes. SIAM Journal on Computing 40(6), 1662\u20131714 (2011)","DOI":"10.1137\/10080484X"},{"key":"7_CR13","doi-asserted-by":"publisher","unstructured":"Fulton, N., Platzer, A.: Safe reinforcement learning via formal methods: Toward safe control through proof and learning. In: Proc. 32nd Conf. on Artificial Intelligence (AAAI-18). pp. 6485\u20136492. AAAI Press (2018). https:\/\/doi.org\/10.1609\/AAAI.V32I1.12107","DOI":"10.1609\/AAAI.V32I1.12107"},{"key":"7_CR14","doi-asserted-by":"publisher","unstructured":"Gao, S., Kong, S., Clarke, E.M.: dReal: An SMT solver for nonlinear theories over the reals. In: Proc. 24th Intl. Conf. on Automated Deduction (CADE-24). Lecture Notes in Computer Science, vol.\u00a07898, pp. 208\u2013214. Springer (2013). https:\/\/doi.org\/10.1007\/978-3-642-38574-2_14","DOI":"10.1007\/978-3-642-38574-2_14"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"Ghaffari, M., Afsharchi, M.: Learning to shift load under uncertain production in the smart grid. Intl. Transactions on Electrical Energy Systems 31(2), e12748 (2021)","DOI":"10.1002\/2050-7038.12748"},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"Giunchiglia, F., Walsh, T.: A theory of abstraction. Artificial intelligence 57(2-3), 323\u2013389 (1992)","DOI":"10.1016\/0004-3702(92)90021-O"},{"key":"7_CR17","doi-asserted-by":"publisher","unstructured":"Jaeger, M., Jensen, P.G., Larsen, K.G., Legay, A., Sedwards, S., Taankvist, J.H.: Teaching Stratego to play ball: Optimal synthesis for continuous space MDPs. In: Proc. 17th Intl. Symposium on Automated Technology for Verification and Analysis (ATVA 2019). Lecture Notes in Computer Science, vol. 11781, pp. 81\u201397. Springer (2019). https:\/\/doi.org\/10.1007\/978-3-030-31784-3_5","DOI":"10.1007\/978-3-030-31784-3_5"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Jansson, A.D.: Discretization and representation of a complex environment for on-policy reinforcement learning for obstacle avoidance for simulated autonomous mobile agents. In: Proc. 7th Intl. Congress on Information and Communication Technology. Lecture Notes in Networks and Systems, vol.\u00a0464, pp. 461\u2013476. Springer (2023)","DOI":"10.1007\/978-981-19-2394-4_42"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Jevti\u0107, \u0110., Miljkovi\u0107, Z., Petrovi\u0107, M., Joki\u0107, A.: Reinforcement learning-based collision avoidance for uav. In: 2023 10th International Conference on Electrical, Electronic and Computing Engineering (IcETRAN). pp.\u00a01\u20136. IEEE (2023)","DOI":"10.1109\/IcETRAN59631.2023.10192168"},{"key":"7_CR20","doi-asserted-by":"publisher","unstructured":"Jin, P., Tian, J., Zhi, D., Wen, X., Zhang, M.: Trainify: A CEGAR-driven training and verification framework for safe deep reinforcement learning. In: Proc. 34th Intl. Conf. on Computer Aided Verification (CAV 2022). Lecture Notes in Computer Science, vol. 13371, pp. 193\u2013218. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-13185-1_10","DOI":"10.1007\/978-3-031-13185-1_10"},{"key":"7_CR21","doi-asserted-by":"crossref","unstructured":"King, J.C.: Symbolic execution and program testing. Communications of the ACM 19(7), 385\u2013394 (1976)","DOI":"10.1145\/360248.360252"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Kober, J., Bagnell, J.A., Peters, J.: Reinforcement learning in robotics: A survey. The Intl. Journal of Robotics Research 32(11), 1238\u20131274 (2013)","DOI":"10.1177\/0278364913495721"},{"key":"7_CR23","doi-asserted-by":"publisher","unstructured":"Kozen, D.: Semantics of probabilistic programs. In: Proc. 20th Annual Symposium on Foundations of Computer Science (SFCS 1979). pp. 101\u2013114. IEEE Computer Society (1979). https:\/\/doi.org\/10.1109\/SFCS.1979.38","DOI":"10.1109\/SFCS.1979.38"},{"key":"7_CR24","doi-asserted-by":"crossref","unstructured":"Kretchmara, R.M., Young, P.M., Anderson, C.W., Hittle, D.C., Anderson, M.L., Delnero, C.C.: Robust reinforcement learning control. In: Proceedings of the 2001 American Control Conference.(Cat. No. 01CH37148). vol.\u00a02, pp. 902\u2013907. IEEE (2001)","DOI":"10.1109\/ACC.2001.945833"},{"key":"7_CR25","doi-asserted-by":"publisher","unstructured":"Lanzi, P.L., Loiacono, D., Wilson, S.W., Goldberg, D.E.: Classifier prediction based on tile coding. In: Proc. Genetic and Evolutionary Computation Conf. (GECCO 2006). pp. 1497\u20131504. ACM (2006). https:\/\/doi.org\/10.1145\/1143997.1144242","DOI":"10.1145\/1143997.1144242"},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Lee, I.S., Lau, H.Y.: Adaptive state space partitioning for reinforcement learning. Engineering applications of artificial intelligence 17(6), 577\u2013588 (2004)","DOI":"10.1016\/j.engappai.2004.08.005"},{"key":"7_CR27","doi-asserted-by":"publisher","unstructured":"Madumal, P., Miller, T., Sonenberg, L., Vetere, F.: Explainable reinforcement learning through a causal lens. In: Proc. 34th Conf. on Artificial Intelligence (AAAI 2020). pp. 2493\u20132500. AAAI Press (2020). https:\/\/doi.org\/10.1609\/AAAI.V34I03.5631","DOI":"10.1609\/AAAI.V34I03.5631"},{"key":"7_CR28","doi-asserted-by":"crossref","unstructured":"Mavridis, C.N., Baras, J.S.: Vector quantization for adaptive state aggregation in reinforcement learning. In: 2021 American Control Conf. (ACC). pp. 2187\u20132192. IEEE (2021)","DOI":"10.23919\/ACC50511.2021.9483052"},{"key":"7_CR29","unstructured":"Michie, D., Chambers, R.A.: Boxes: An experiment in adaptive control. Machine intelligence 2(2), 137\u2013152 (1968)"},{"key":"7_CR30","unstructured":"Mnih, V., Badia, A.P., Mirza, M., Graves, A., Lillicrap, T.P., Harley, T., Silver, D., Kavukcuoglu, K.: Asynchronous methods for deep reinforcement learning. In: Proc. 33nd Intl. Conf. on Machine Learning (ICML 2016). JMLR Workshop and Conf. Proceedings, vol.\u00a048, pp. 1928\u20131937. JMLR.org (2016), http:\/\/proceedings.mlr.press\/v48\/mniha16.html"},{"key":"7_CR31","unstructured":"Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller, M.: Playing Atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602 (2013)"},{"key":"7_CR32","unstructured":"Moore, A.W.: Efficient memory-based learning for robot control. Ph.D. thesis, University of Cambridge, UK (1990). https:\/\/doi.org\/10.1.1.17.2654"},{"key":"7_CR33","doi-asserted-by":"crossref","unstructured":"Moore, A.W.: Variable resolution dynamic programming: Efficiently learning action maps in multivariate real-valued state-spaces. In: Machine Learning Proceedings 1991, pp. 333\u2013337. Elsevier (1991)","DOI":"10.1016\/B978-1-55860-200-7.50069-6"},{"key":"7_CR34","doi-asserted-by":"publisher","unstructured":"de\u00a0Moura, L.M., Bj\u00f8rner, N.S.: Z3: an efficient SMT solver. In: Proc. 14th Intl. Conf. on Tools and Algorithms for the Construction and Analysis of Systems (TACAS 2008). Lecture Notes in Computer Science, vol.\u00a04963, pp. 337\u2013340. Springer (2008). https:\/\/doi.org\/10.1007\/978-3-540-78800-3_24","DOI":"10.1007\/978-3-540-78800-3_24"},{"key":"7_CR35","doi-asserted-by":"crossref","unstructured":"Nicol, S., Chad\u00e8s, I.: Which states matter? an application of an intelligent discretization method to solve a continuous POMDP in conservation biology. PloS one 7(2), e28993 (2012)","DOI":"10.1371\/journal.pone.0028993"},{"key":"7_CR36","doi-asserted-by":"publisher","unstructured":"Pasareanu, C.S., Visser, W., Bushnell, D.H., Geldenhuys, J., Mehlitz, P.C., Rungta, N.: Symbolic PathFinder: integrating symbolic execution with model checking for Java bytecode analysis. Autom. Softw. Eng. 20(3), 391\u2013425 (2013). https:\/\/doi.org\/10.1007\/S10515-013-0122-2","DOI":"10.1007\/S10515-013-0122-2"},{"key":"7_CR37","doi-asserted-by":"publisher","unstructured":"Puiutta, E., Veith, E.M.S.P.: Explainable reinforcement learning: A survey. In: Proc. 4th Intl. Cross-Domain Conf. (CD-MAKE 2020). Lecture Notes in Computer Science, vol. 12279, pp. 77\u201395. Springer (2020). https:\/\/doi.org\/10.1007\/978-3-030-57321-8_5","DOI":"10.1007\/978-3-030-57321-8_5"},{"key":"7_CR38","unstructured":"Raffin, A., Hill, A., Ernestus, M., Gleave, A., Kanervisto, A., Dormann, N.: Stable baselines3 (2019), https:\/\/stable-baselines3.readthedocs.io\/"},{"key":"7_CR39","unstructured":"Russell, S.J., Norvig, P.: Artificial intelligence a modern approach. London (2010)"},{"key":"7_CR40","unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347 (2017)"},{"key":"7_CR41","doi-asserted-by":"crossref","unstructured":"Seipp, J., Helmert, M.: Counterexample-guided cartesian abstraction refinement for classical planning. Journal of Artificial Intelligence Research 62, 535\u2013577 (2018)","DOI":"10.1613\/jair.1.11217"},{"key":"7_CR42","doi-asserted-by":"crossref","unstructured":"Sharon, G., Stern, R., Felner, A., Sturtevant, N.R.: Conflict-based search for optimal multi-agent pathfinding. Artificial Intelligence 219, 40\u201366 (2015)","DOI":"10.1016\/j.artint.2014.11.006"},{"key":"7_CR43","unstructured":"Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. The MIT Press, 2nd edn. (2018)"},{"key":"7_CR44","doi-asserted-by":"publisher","unstructured":"Szita, I.: Reinforcement learning in games. In: Reinforcement Learning, Adaptation, Learning, and Optimization, vol.\u00a012, pp. 539\u2013577. Springer (2012). https:\/\/doi.org\/10.1007\/978-3-642-27645-3_17","DOI":"10.1007\/978-3-642-27645-3_17"},{"key":"7_CR45","doi-asserted-by":"crossref","unstructured":"Tran, H.D., Cai, F., Diego, M.L., Musau, P., Johnson, T.T., Koutsoukos, X.: Safety verification of cyber-physical systems with reinforcement learning control. ACM Transactions on Embedded Computing Systems (TECS) 18(5s), 1\u201322 (2019)","DOI":"10.1145\/3358230"},{"key":"7_CR46","unstructured":"Uther, W.T.B., Veloso, M.M.: Tree based discretization for continuous state space reinforcement learning. In: Proc. 15th National Conf. on Artificial Intelligence and Tenth Innovative Applications of Artificial Intelligence Conf. (AAAI 98, IAAI 98). pp. 769\u2013774. AAAI Press \/ The MIT Press (1998), http:\/\/www.aaai.org\/Library\/AAAI\/1998\/aaai98-109.php"},{"key":"7_CR47","doi-asserted-by":"publisher","unstructured":"Varshosaz, M., Ghaffari, M., Johnsen, E.B., W\u0105sowski, A.: Formal specification and testing for reinforcement learning. Proc. ACM Program. Lang. 7(ICFP) (aug 2023). https:\/\/doi.org\/10.1145\/3607835","DOI":"10.1145\/3607835"},{"key":"7_CR48","doi-asserted-by":"publisher","unstructured":"Verdier, C.F., Babu\u0161ka, R., Shyrokau, B., Mazo, M.: Near optimal control with reachability and safety guarantees. IFAC-PapersOnLine 52(11), 230\u2013235 (2019). https:\/\/doi.org\/10.1016\/j.ifacol.2019.09.146","DOI":"10.1016\/j.ifacol.2019.09.146"},{"key":"7_CR49","doi-asserted-by":"publisher","unstructured":"Visser, W., Pasareanu, C.S., Pel\u00e1nek, R.: Test input generation for java containers using state matching. In: Pollock, L.L., Pezz\u00e8, M. (eds.) Proceedings of the ACM\/SIGSOFT International Symposium on Software Testing and Analysis, ISSTA 2006, Portland, Maine, USA, July 17-20, 2006. pp. 37\u201348. ACM (2006). https:\/\/doi.org\/10.1145\/1146238.1146243","DOI":"10.1145\/1146238.1146243"},{"key":"7_CR50","doi-asserted-by":"publisher","unstructured":"Voogd, E., Johnsen, E.B., Silva, A., Susag, Z.J., W\u0105sowski, A.: Symbolic semantics for probabilistic programs. In: Proc. 20th Intl. Conf. on Quantitative Evaluation of Systems (QEST 2023). Lecture Notes in Computer Science, vol. 14287, pp. 329\u2013345. Springer (2023). https:\/\/doi.org\/10.1007\/978-3-031-43835-6_23","DOI":"10.1007\/978-3-031-43835-6_23"},{"key":"7_CR51","doi-asserted-by":"publisher","unstructured":"Vyetrenko, S., Xu, S.: Risk-sensitive compact decision trees for autonomous execution in presence of simulated market response. arXiv preprint arXiv:1906.02312 (2019). https:\/\/doi.org\/10.48550\/ARXIV.1906.02312","DOI":"10.48550\/ARXIV.1906.02312"},{"key":"7_CR52","doi-asserted-by":"crossref","unstructured":"Wei, H., Corder, K., Decker, K.: Q-learning acceleration via state-space partitioning. In: Proc. 17th Intl. Conf. on Machine Learning and Applications (ICMLA 2018). pp. 293\u2013298. IEEE (2018)","DOI":"10.1109\/ICMLA.2018.00050"},{"key":"7_CR53","doi-asserted-by":"publisher","unstructured":"Whiteson, S.: Adaptive Representations for Reinforcement Learning, Studies in Computational Intelligence, vol.\u00a0291. Springer (2010). https:\/\/doi.org\/10.1007\/978-3-642-13932-1","DOI":"10.1007\/978-3-642-13932-1"},{"key":"7_CR54","doi-asserted-by":"crossref","unstructured":"Yu, C., Liu, J., Nemati, S., Yin, G.: Reinforcement learning in healthcare: A survey. ACM Computing Surveys (CSUR) 55(1), 1\u201336 (2021)","DOI":"10.1145\/3477600"},{"key":"7_CR55","doi-asserted-by":"publisher","unstructured":"Zelvelder, A.E., Westberg, M., Fr\u00e4mling, K.: Assessing explainability in reinforcement learning. In: Proc. Third Intl. Workshop on Explainable and Transparent AI and Multi-Agent Systems (EXTRAAMAS 2021). Lecture Notes in Computer Science, vol. 12688, pp. 223\u2013240. Springer (2021). https:\/\/doi.org\/10.1007\/978-3-030-82017-6_14","DOI":"10.1007\/978-3-030-82017-6_14"}],"container-title":["Lecture Notes in Computer Science","Fundamental Approaches to Software Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-90900-9_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,30]],"date-time":"2025-04-30T04:45:18Z","timestamp":1745988318000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-90900-9_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031908996","9783031909009"],"references-count":55,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-90900-9_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"1 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"FASE","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Fundamental Approaches to Software Engineering","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hamilton, ON","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 May 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 May 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"fase2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/etaps.org\/2025\/conferences\/fase\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}