{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T07:50:34Z","timestamp":1740124234116,"version":"3.37.3"},"reference-count":53,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2019,5,17]],"date-time":"2019-05-17T00:00:00Z","timestamp":1558051200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2019,5,17]],"date-time":"2019-05-17T00:00:00Z","timestamp":1558051200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100005881","name":"Emberi Eroforr\u00e1sok Miniszt\u00e9riuma","doi-asserted-by":"publisher","award":["BME FIKP-MI\/SC"],"award-info":[{"award-number":["BME FIKP-MI\/SC"]}],"id":[{"id":"10.13039\/501100005881","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Pro Progressio Foundation"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Software Qual J"],"published-print":{"date-parts":[[2019,9]]},"DOI":"10.1007\/s11219-019-09446-5","type":"journal-article","created":{"date-parts":[[2019,5,17]],"date-time":"2019-05-17T00:03:50Z","timestamp":1558051430000},"page":"1339-1380","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Classifying generated white-box tests: an exploratory study"],"prefix":"10.1007","volume":"27","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5217-828X","authenticated-orcid":false,"given":"D\u00e1vid","family":"Honfi","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1846-261X","authenticated-orcid":false,"given":"Zolt\u00e1n","family":"Micskei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,5,17]]},"reference":[{"issue":"3","key":"9446_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11219-016-9339-1","volume":"25","author":"R Almaghairbe","year":"2016","unstructured":"Almaghairbe, R., & Roper, M. (2016). Separating passing and failing test executions by clustering anomalies. Software Quality Journal, 25(3), 1\u201338. \n                    https:\/\/doi.org\/10.1007\/s11219-016-9339-1\n                    \n                  .","journal-title":"Software Quality Journal"},{"issue":"8","key":"9446_CR2","doi-asserted-by":"publisher","first-page":"1978","DOI":"10.1016\/j.jss.2013.02.061","volume":"86","author":"S Anand","year":"2013","unstructured":"Anand, S., Burke, E.K., Chen, T.Y., Clark, J., Cohen, M.B., Grieskamp, W., Harman, M., Harrold, M.J., McMinn, P. (2013). An orchestrated survey of methodologies for automated software test case generation. Journal of Systems and Software, 86(8), 1978\u20132001. \n                    https:\/\/doi.org\/10.1016\/j.jss.2013.02.061\n                    \n                  .","journal-title":"Journal of Systems and Software"},{"issue":"5","key":"9446_CR3","doi-asserted-by":"publisher","first-page":"507","DOI":"10.1109\/TSE.2014.2372785","volume":"41","author":"ET Barr","year":"2015","unstructured":"Barr, E.T., Harman, M., McMinn, P., Shahbaz, M., Yoo, S. (2015). The oracle problem in software testing: A survey. IEEE Transactions on Software Engineering, 41(5), 507\u2013525. \n                    https:\/\/doi.org\/10.1109\/TSE.2014.2372785\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"issue":"12","key":"9446_CR4","doi-asserted-by":"publisher","first-page":"1278","DOI":"10.1109\/TSE.1987.232881","volume":"SE-13","author":"V Basili","year":"1987","unstructured":"Basili, V., & Selby, R. (1987). Comparing the effectiveness of software testing strategies. IEEE Transactions on Software Engineering, SE-13(12), 1278\u20131296. \n                    https:\/\/doi.org\/10.1109\/TSE.1987.232881\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"issue":"4","key":"9446_CR5","doi-asserted-by":"publisher","first-page":"456","DOI":"10.1109\/32.799939","volume":"25","author":"V Basili","year":"1999","unstructured":"Basili, V., Shull, F., Lanubile, F. (1999). Building knowledge through families of experiments. IEEE Transactions on Software Engineering, 25(4), 456\u2013473. \n                    https:\/\/doi.org\/10.1109\/32.799939\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"key":"9446_CR6","doi-asserted-by":"publisher","unstructured":"Bounimova, E., Godefroid, P., Molnar D. (2013). Billions and billions of constraints: Whitebox fuzz testing in production. In: International conference on software engineering, IEEE, ICSE, pp. 122\u2013131. \n                    https:\/\/doi.org\/10.1109\/ICSE.2013.6606558\n                    \n                  .","DOI":"10.1109\/ICSE.2013.6606558"},{"issue":"11","key":"9446_CR7","doi-asserted-by":"publisher","first-page":"943","DOI":"10.1109\/32.177364","volume":"18","author":"R Chillarege","year":"1992","unstructured":"Chillarege, R., Bhandari, I.S., Chaar, J.K., Halliday, M.J., Moebus, D.S., Ray, B.K., Wong, M.Y. (1992). Orthogonal defect classification-a concept for in-process measurements. IEEE Transactions on Software Engineering, 18(11), 943\u2013956. \n                    https:\/\/doi.org\/10.1109\/32.177364\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"key":"9446_CR8","doi-asserted-by":"publisher","unstructured":"Claessen, K., & Hughes, J. (2000). Quickcheck: A lightweight tool for random testing of haskell programs. In: International Conference on Functional Programming, ACM, ICFP, pp. 268\u2013279. \n                    https:\/\/doi.org\/10.1145\/351240.351266\n                    \n                  .","DOI":"10.1145\/351240.351266"},{"key":"9446_CR9","doi-asserted-by":"publisher","unstructured":"Daka, E., Rojas, J.M., Fraser, G. (2017). Generating unit tests with descriptive names or: Would you name your children thing1 and thing2?. In: Int. symp on software testing and analysis, ISSTA, pp 57\u201367. \n                    https:\/\/doi.org\/10.1145\/3092703.3092727\n                    \n                  .","DOI":"10.1145\/3092703.3092727"},{"issue":"5","key":"9446_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10664-016-9471-3","volume":"22","author":"O Dieste","year":"2017","unstructured":"Dieste, O., Aranda, A.M., Uyaguari, F., Turhan, B., Tosun, A., Fucci, D., Oivo, M., Juristo, N. (2017). Empirical evaluation of the effects of experience on code quality and programmer productivity: an exploratory study. Empirical Software Engineering, 22(5), 1\u201386. \n                    https:\/\/doi.org\/10.1007\/s10664-016-9471-3\n                    \n                  .","journal-title":"Empirical Software Engineering"},{"issue":"11","key":"9446_CR11","doi-asserted-by":"publisher","first-page":"849","DOI":"10.1109\/TSE.2006.113","volume":"32","author":"JA Duraes","year":"2006","unstructured":"Duraes, J.A., & Madeira, H.S. (2006). Emulation of software faults: A field data study and a practical approach. IEEE Transactions on Software Engineering, 32(11), 849\u2013867. \n                    https:\/\/doi.org\/10.1109\/TSE.2006.113\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"key":"9446_CR12","doi-asserted-by":"publisher","unstructured":"Easterbrook, S., Singer, J., Storey, M.A., Damian, D. (2008). Selecting empirical methods for software engineering research, Springer, chap 11, pp. 285\u2013311. \n                    https:\/\/doi.org\/10.1007\/978-1-84800-044-5_11\n                    \n                  .","DOI":"10.1007\/978-1-84800-044-5_11"},{"key":"9446_CR13","doi-asserted-by":"publisher","unstructured":"Enoiu, E.P., Causevic, A., Sundmark, D., Pettersson, P. (2016). A controlled experiment in testing of safety-critical embedded software. In: Int. conf. on software testing, verification and validation, ICST, pp. 1\u201311. \n                    https:\/\/doi.org\/10.1109\/ICST.2016.15\n                    \n                  .","DOI":"10.1109\/ICST.2016.15"},{"key":"9446_CR14","doi-asserted-by":"publisher","unstructured":"Enoiu, E.P., Causevic, A., Sundmark, D., Pettersson, P. (2017). A comparative study of manual and automated testing for industrial control software. In: Int. conf. on software testing, verification and validation, ICST, pp 412\u2013417. \n                    https:\/\/doi.org\/10.1109\/ICST.2017.44\n                    \n                  .","DOI":"10.1109\/ICST.2017.44"},{"issue":"1","key":"9446_CR15","doi-asserted-by":"publisher","first-page":"452","DOI":"10.1007\/s10664-017-9523-3","volume":"23","author":"D Falessi","year":"2018","unstructured":"Falessi, D., Juristo, N., Wohlin, C., Turhan, B., M\u00fcnch, J., Jedlitschka, A., Oivo, M. (2018). Empirical software engineering experts on the use of students and professionals in experiments. Empirical Software Engineering, 23(1), 452\u2013489. \n                    https:\/\/doi.org\/10.1007\/s10664-017-9523-3\n                    \n                  .","journal-title":"Empirical Software Engineering"},{"issue":"2","key":"9446_CR16","doi-asserted-by":"publisher","first-page":"276","DOI":"10.1109\/TSE.2012.14","volume":"39","author":"G Fraser","year":"2013","unstructured":"Fraser, G., & Arcuri, A. (2013). Whole test suite generation. IEEE Transactions on Software Engineering, 39(2), 276\u2013291. \n                    https:\/\/doi.org\/10.1109\/TSE.2012.14\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"issue":"4","key":"9446_CR17","doi-asserted-by":"publisher","first-page":"23:1","DOI":"10.1145\/2699688","volume":"24","author":"G Fraser","year":"2015","unstructured":"Fraser, G., Staats, M., McMinn, P., Arcuri, A., Padberg, F. (2015). Does automated unit test generation really help software testers? a controlled empirical study. ACM Trans Softw Eng Methodol, 24(4), 23:1\u201323:49. \n                    https:\/\/doi.org\/10.1145\/2699688\n                    \n                  .","journal-title":"ACM Trans Softw Eng Methodol"},{"issue":"11","key":"9446_CR18","doi-asserted-by":"publisher","first-page":"1325","DOI":"10.1111\/2041-210X.12584","volume":"7","author":"O Friard","year":"2016","unstructured":"Friard, O., & Gamba, M. (2016). Boris: a free, versatile open-source event-logging software for video\/audio coding and live observations. Methods in Ecology and Evolution, 7(11), 1325\u20131330. \n                    https:\/\/doi.org\/10.1111\/2041-210X.12584\n                    \n                  .","journal-title":"Methods in Ecology and Evolution"},{"key":"9446_CR19","doi-asserted-by":"publisher","unstructured":"Honfi, D., & Micskei, Z. (2018). Classifying generated white-box tests: An exploratory study [Data set]. \n                    https:\/\/doi.org\/10.5281\/zenodo.2596044\n                    \n                  .","DOI":"10.5281\/zenodo.2596044"},{"issue":"1","key":"9446_CR20","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1023\/B:EMSE.0000013513.48963.1b","volume":"9","author":"N Juristo","year":"2004","unstructured":"Juristo, N., Moreno, A.M., Vegas, S. (2004). Reviewing 25 years of testing technique experiments. Empirical Software Engineering, 9(1), 7\u201344. \n                    https:\/\/doi.org\/10.1023\/B:EMSE.0000013513.48963.1b\n                    \n                  .","journal-title":"Empirical Software Engineering"},{"key":"9446_CR21","doi-asserted-by":"publisher","unstructured":"Just, R., Jalali, D., Inozemtseva, L., Ernst, M.D., Holmes, R., Fraser, G. (2014). Are mutants a valid substitute for real faults in software testing?. In: Int. symp. on foundations of software engineering, ACM, FSE, pp. 654\u2013665. \n                    https:\/\/doi.org\/10.1145\/2635868.2635929\n                    \n                  .","DOI":"10.1145\/2635868.2635929"},{"issue":"1","key":"9446_CR22","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1007\/s10664-013-9279-3","volume":"20","author":"AJ Ko","year":"2013","unstructured":"Ko, A.J., LaToza, T.D., Burnett, M.M. (2013). A practical guide to controlled experiments of software engineering tools with human participants. Empirical Software Engineering, 20(1), 110\u2013141. \n                    https:\/\/doi.org\/10.1007\/s10664-013-9279-3\n                    \n                  .","journal-title":"Empirical Software Engineering"},{"key":"9446_CR23","doi-asserted-by":"publisher","unstructured":"Kracht, J., Petrovic, J., Walcott-Justice, K. (2014). Empirically evaluating the quality of automatically generated and manually written test suites. In: Int. Conf on Quality Software, QSIC, pp. 256\u2013265. \n                    https:\/\/doi.org\/10.1109\/QSIC.2014.33\n                    \n                  .","DOI":"10.1109\/QSIC.2014.33"},{"key":"9446_CR24","unstructured":"MathNET. (2017). Math.NET numerics. \n                    https:\/\/numerics.mathdotnet.com\/\n                    \n                  ."},{"key":"9446_CR25","doi-asserted-by":"publisher","unstructured":"McMinn, P., Stevenson, M., Harman, M. (2010). Reducing qualitative human oracle costs associated with automatically generated test data. In: Int. Workshop on Software Test Output Validation, ACM, STOV \u201910, pp 1\u20134. \n                    https:\/\/doi.org\/10.1145\/1868048.1868049\n                    \n                  .","DOI":"10.1145\/1868048.1868049"},{"issue":"9","key":"9446_CR26","doi-asserted-by":"publisher","first-page":"760","DOI":"10.1145\/359588.359602","volume":"21","author":"GJ Myers","year":"1978","unstructured":"Myers, G.J. (1978). A controlled experiment in program testing and code walkthroughs\/inspections. Commun. ACM, 21(9), 760\u2013768. \n                    https:\/\/doi.org\/10.1145\/359588.359602\n                    \n                  .","journal-title":"Commun. ACM"},{"key":"9446_CR27","unstructured":"NBitcoin. (2017). NBitcoin. \n                    https:\/\/github.com\/MetacoSA\/NBitcoin\n                    \n                  ."},{"key":"9446_CR28","unstructured":"NetTopologySuite. (2018). NetTopologySuite. \n                    https:\/\/github.com\/NetTopologySuite\/NetTopologySuite\n                    \n                  ."},{"key":"9446_CR29","doi-asserted-by":"publisher","unstructured":"Nguyen, C.D., Marchetto, A., Tonella, P. (2013). Automated oracles: An empirical study on cost and effectiveness. In: Foundations of software engineering, ACM, ESEC\/FSE, pp. 136\u2013146. \n                    https:\/\/doi.org\/10.1145\/2491411.2491434\n                    \n                  .","DOI":"10.1145\/2491411.2491434"},{"key":"9446_CR30","unstructured":"NodaTime. (2018). NodaTime. \n                    https:\/\/github.com\/nodatime\/nodatime\n                    \n                  ."},{"key":"9446_CR31","doi-asserted-by":"publisher","unstructured":"Pacheco, C., Lahiri, S.K., Ernst, M.D., Ball, T. (2007). Feedback-directed random test generation. In: Int. conf on software engineering. ICSE, pp. 75\u201384. \n                    https:\/\/doi.org\/10.1109\/ICSE.2007.37\n                    \n                  .","DOI":"10.1109\/ICSE.2007.37"},{"key":"9446_CR32","doi-asserted-by":"publisher","unstructured":"Panichella, S., Panichella, A., Beller, M., Zaidman, A., Gall, H.C. (2016a). The impact of test case summaries on bug fixing performance: An empirical investigation. In: Int. conf on software engineering. ICSE, pp. 547\u2013558. \n                    https:\/\/doi.org\/10.1145\/2884781.2884847\n                    \n                  .","DOI":"10.1145\/2884781.2884847"},{"key":"9446_CR33","doi-asserted-by":"publisher","unstructured":"Panichella, S., Panichella, A., Beller, M., Zaidman, A., Gall, H.C. (2016b). The impact of test case summaries on bug fixing performance: An empirical investigation. In: Int. conf on software engineering. ICSE, pp. 547\u2013558. \n                    https:\/\/doi.org\/10.1145\/2884781.2884847\n                    \n                  .","DOI":"10.1145\/2884781.2884847"},{"key":"9446_CR34","doi-asserted-by":"publisher","unstructured":"Papadakis, M., Henard, C., Harman, M., Jia, Y., Le Traon, Y. (2016). Threats to the validity of mutation-based test assessment. In: Int. symp. on software testing and analysis, ACM, ISSTA, pp. 354\u2013365. \n                    https:\/\/doi.org\/10.1145\/2931037.2931040\n                    \n                  .","DOI":"10.1145\/2931037.2931040"},{"key":"9446_CR35","doi-asserted-by":"publisher","unstructured":"Pastore, F., & Mariani, L. (2015). Zoomin: Discovering failures by detecting wrong assertions. In: Int. conf. on software engineering. IEEE, ICSE, pp. 66\u201376. \n                    https:\/\/doi.org\/10.1109\/ICSE.2015.29\n                    \n                  .","DOI":"10.1109\/ICSE.2015.29"},{"key":"9446_CR36","doi-asserted-by":"publisher","unstructured":"Pastore, F., Mariani, L., Fraser, G. (2013). Crowdoracles: Can the crowd solve the oracle problem?. In: Int. conf. on software testing, verification and validation, ICST, pp. 342\u2013351. \n                    https:\/\/doi.org\/10.1109\/ICST.2013.13\n                    \n                  \n                           \n                    https:\/\/doi.org\/10.1109\/ICST.2013.13\n                    \n                  .","DOI":"10.1109\/ICST.2013.13 10.1109\/ICST.2013.13"},{"issue":"1","key":"9446_CR37","first-page":"37","volume":"2","author":"DMW Powers","year":"2011","unstructured":"Powers, D.M.W. (2011). Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation. International Journal of Machine Learning Technology, 2(1), 37\u201363.","journal-title":"International Journal of Machine Learning Technology"},{"key":"9446_CR38","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.jss.2018.03.052","volume":"141","author":"MP Prado","year":"2018","unstructured":"Prado, M.P., & Vincenzi, A.M.R. (2018). Towards cognitive support for unit testing: A qualitative study with practitioners. Journal of Systems and Software, 141, 66\u201384. \n                    https:\/\/doi.org\/10.1016\/j.jss.2018.03.052\n                    \n                  .","journal-title":"Journal of Systems and Software"},{"key":"9446_CR39","unstructured":"R. Core Team. (2016). R: A language and environment for statistical computing, R Foundation for Statistical Computing, Vienna. \n                    https:\/\/www.R-project.org\/\n                    \n                  ."},{"key":"9446_CR40","doi-asserted-by":"publisher","unstructured":"Ramler, R., Winkler, D., Schmidt, M. (2012). Random test case generation and manual unit testing: Substitute or complement in retrofitting tests for legacy code? In: EUROMICRO Conf. on Software Engineering and Advanced Applications (SEAA), pp. 286\u2013293. \n                    https:\/\/doi.org\/10.1109\/SEAA.2012.42\n                    \n                  .","DOI":"10.1109\/SEAA.2012.42"},{"key":"9446_CR41","doi-asserted-by":"publisher","unstructured":"Rojas, J.M., Fraser, G., Arcuri, A. (2015). Automated unit test generation during software development: A controlled experiment and think-aloud observations. In: Int. symp. on software testing and analysis, ACM, ISSTA, pp. 338\u2013349. \n                    https:\/\/doi.org\/10.1145\/2771783.2771801\n                    \n                  .","DOI":"10.1145\/2771783.2771801"},{"issue":"1","key":"9446_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.infsof.2012.06.002","volume":"55","author":"A Shahrokni","year":"2013","unstructured":"Shahrokni, A., & Feldt, R. (2013). A systematic review of software robustness. Information and Software Technology, 55(1), 1\u201317. \n                    https:\/\/doi.org\/10.1016\/j.infsof.2012.06.002\n                    \n                  .","journal-title":"Information and Software Technology"},{"key":"9446_CR43","doi-asserted-by":"publisher","unstructured":"Shamshiri, S., Just, R., Rojas, J.M., Fraser, G., McMinn, P., Arcuri, A. (2015). Do automatically generated unit tests find real faults? An empirical study of effectiveness and challenges. In: Int. conf. on automated software engineering, ACM, ASE, pp. 201\u2013211. \n                    https:\/\/doi.org\/10.1109\/ASE.2015.86\n                    \n                  .","DOI":"10.1109\/ASE.2015.86"},{"key":"9446_CR44","doi-asserted-by":"crossref","unstructured":"Shamshiri, S., Rojas, J.M., Galeotti, J.P., Walkinshaw, N., Fraser, G. (2018). How do automatically generated unit tests influence software maintenance? In: Int. conf on software testing, verification and validation, ICST.","DOI":"10.1109\/ICST.2018.00033"},{"issue":"9","key":"9446_CR45","doi-asserted-by":"publisher","first-page":"733","DOI":"10.1109\/TSE.2005.97","volume":"31","author":"D Sj\u00f8berg","year":"2005","unstructured":"Sj\u00f8berg, D., Hannay, J., Hansen, O., Kampenes, V., Karahasanovic, A., Liborg, N.K., Rekdal, A. (2005). A survey of controlled experiments in software engineering. IEEE Transactions on Software Engineering, 31(9), 733\u2013753. \n                    https:\/\/doi.org\/10.1109\/TSE.2005.97\n                    \n                  .","journal-title":"IEEE Transactions on Software Engineering"},{"key":"9446_CR46","doi-asserted-by":"publisher","unstructured":"Staats, M., Hong, S., Kim, M., Rothermel, G. (2012). Understanding user understanding: Determining correctness of generated program invariants. In: Int. symp. on software testing and analysis, ACM, ISSTA, pp. 188\u2013198. \n                    https:\/\/doi.org\/10.1145\/2338965.2336776\n                    \n                  .","DOI":"10.1145\/2338965.2336776"},{"key":"9446_CR47","doi-asserted-by":"publisher","unstructured":"Tillmann, N., & de Halleux, J. (2008). Pex\u2013White Box Test Generation for.NET. In Beckert, B., & H\u00e4hnle, R. (Eds.) Tests and Proofs, LNCS, (Vol. 4966 pp. 134\u2013153): Springer.. \n                    https:\/\/doi.org\/10.1007\/978-3-540-79124-9_10","DOI":"10.1007\/978-3-540-79124-9_10"},{"key":"9446_CR48","doi-asserted-by":"publisher","unstructured":"Tillmann, N., & Schulte, W. (2005). Parameterized unit tests. In: Foundations of Software Engineering, ESEC\/FSE, pp. 253\u2013262. \n                    https:\/\/doi.org\/10.1145\/1081706.1081749\n                    \n                  .","DOI":"10.1145\/1081706.1081749"},{"key":"9446_CR49","doi-asserted-by":"publisher","unstructured":"Tillmann, N., de Halleux, J., Xie, T. (2014). Transferring an automated test generation tool to practice: From Pex to Fakes and Code Digger. In: Int. conf on automated software engineering, ASE, pp. 385\u2013396. \n                    https:\/\/doi.org\/10.1145\/2642937.2642941\n                    \n                  .","DOI":"10.1145\/2642937.2642941"},{"key":"9446_CR50","doi-asserted-by":"publisher","unstructured":"Wang, X., Zhang, L., Tanofsky, P. (2015). Experience report: How is dynamic symbolic execution different from manual testing? A study on KLEE. In: int. symp. on software testing and analysis, ACM, ISSTA, pp 199\u2013210. \n                    https:\/\/doi.org\/10.1145\/2771783.2771818\n                    \n                  .","DOI":"10.1145\/2771783.2771818"},{"issue":"6","key":"9446_CR51","doi-asserted-by":"publisher","first-page":"1427","DOI":"10.1007\/s10664-014-9319-7","volume":"20","author":"C Wohlin","year":"2015","unstructured":"Wohlin, C., & Aurum, A. (2015). Towards a decision-making structure for selecting a research design in empirical software engineering. Empirical Software Engineering, 20 (6), 1427\u20131455. \n                    https:\/\/doi.org\/10.1007\/s10664-014-9319-7\n                    \n                  .","journal-title":"Empirical Software Engineering"},{"key":"9446_CR52","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-29044-2","volume-title":"Experimentation in software engineering","author":"C Wohlin","year":"2012","unstructured":"Wohlin, C., Runeson, P., H\u00f6st, M, Ohlsson, M.C., Regnell, B., Wessl\u0117n, A. (2012). Experimentation in software engineering. Berlin: Springer . \n                    https:\/\/doi.org\/10.1007\/978-3-642-29044-2\n                    \n                  ."},{"issue":"2","key":"9446_CR53","doi-asserted-by":"publisher","first-page":"67","DOI":"10.1002\/stv.430","volume":"22","author":"S Yoo","year":"2012","unstructured":"Yoo, S., & Harman, M. (2012). Regression testing minimization, selection and prioritization: a survey. Software Testing. Verification and Reliability, 22(2), 67\u2013120. \n                    https:\/\/doi.org\/10.1002\/stvr.430\n                    \n                  .","journal-title":"Verification and Reliability"}],"container-title":["Software Quality Journal"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11219-019-09446-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s11219-019-09446-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11219-019-09446-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,5,15]],"date-time":"2020-05-15T23:35:59Z","timestamp":1589585759000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s11219-019-09446-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,5,17]]},"references-count":53,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2019,9]]}},"alternative-id":["9446"],"URL":"https:\/\/doi.org\/10.1007\/s11219-019-09446-5","relation":{},"ISSN":["0963-9314","1573-1367"],"issn-type":[{"type":"print","value":"0963-9314"},{"type":"electronic","value":"1573-1367"}],"subject":[],"published":{"date-parts":[[2019,5,17]]},"assertion":[{"value":"17 May 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}