{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T06:39:35Z","timestamp":1772087975301,"version":"3.50.1"},"reference-count":61,"publisher":"Cambridge University Press (CUP)","license":[{"start":{"date-parts":[[2022,7,27]],"date-time":"2022-07-27T00:00:00Z","timestamp":1658880000000},"content-version":"unspecified","delay-in-days":207,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000147","name":"Division of Civil, Mechanical and Manufacturing Innovation","doi-asserted-by":"publisher","award":["2145432"],"award-info":[{"award-number":["2145432"]}],"id":[{"id":"10.13039\/100000147","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000147","name":"Division of Civil, Mechanical and Manufacturing Innovation","doi-asserted-by":"publisher","award":["2145432"],"award-info":[{"award-number":["2145432"]}],"id":[{"id":"10.13039\/100000147","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["cambridge.org"],"crossmark-restriction":true},"short-container-title":["AIEDAM"],"published-print":{"date-parts":[[2022]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Inspirational stimuli are known to be effective in supporting ideation during early-stage design. However, prior work has predominantly constrained designers to using text-only queries when searching for stimuli, which is not consistent with real-world design behavior where fluidity across modalities (e.g., visual, semantic, etc.) is standard practice. In the current work, we introduce a multi-modal search platform that retrieves inspirational stimuli in the form of 3D-model parts using text, appearance, and function-based search inputs. Computational methods leveraging a deep-learning approach are presented for designing and supporting this platform, which relies on deep-neural networks trained on a large dataset of 3D-model parts. This work further presents the results of a cognitive study (<jats:italic>n<\/jats:italic>= 21) where the aforementioned search platform was used to find parts to inspire solutions to a design challenge. Participants engaged with three different search modalities: by keywords, 3D parts, and user-assembled 3D parts in their workspace. When searching by parts that are selected or in their workspace, participants had additional control over the similarity of appearance and function of results relative to the input. The results of this study demonstrate that the modality used impacts search behavior, such as in search frequency, how retrieved search results are engaged with, and how broadly the search space is covered. Specific results link interactions with the interface to search strategies participants may have used during the task. Findings suggest that when searching for inspirational stimuli, desired results can be achieved both by direct search inputs (e.g., by keyword) as well as by more randomly discovered examples, where a specific goal was not defined. Both search processes are found to be important to enable when designing search platforms for inspirational stimuli retrieval.<\/jats:p>","DOI":"10.1017\/s0890060422000130","type":"journal-article","created":{"date-parts":[[2022,7,27]],"date-time":"2022-07-27T11:40:33Z","timestamp":1658922033000},"update-policy":"https:\/\/doi.org\/10.1017\/policypage","source":"Crossref","is-referenced-by-count":14,"title":["Enabling multi-modal search for inspirational design stimuli using deep learning"],"prefix":"10.1017","volume":"36","author":[{"given":"Elisa","family":"Kwon","sequence":"first","affiliation":[]},{"given":"Forrest","family":"Huang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0850-9197","authenticated-orcid":false,"given":"Kosa","family":"Goucher-Lambert","sequence":"additional","affiliation":[]}],"member":"56","published-online":{"date-parts":[[2022,7,27]]},"reference":[{"key":"S0890060422000130_ref2","first-page":"2645","article-title":"Exploratory search: from finding to understanding","volume":"67","author":"Athukorala","year":"2016","journal-title":"Journal of the Association for Information Science and Technology"},{"key":"S0890060422000130_ref32","doi-asserted-by":"publisher","DOI":"10.1115\/1.4049214"},{"key":"S0890060422000130_ref49","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2019.112995"},{"key":"S0890060422000130_ref28","doi-asserted-by":"publisher","DOI":"10.1080\/09544821003598573"},{"key":"S0890060422000130_ref57","doi-asserted-by":"publisher","DOI":"10.1016\/j.destud.2008.01.003"},{"key":"S0890060422000130_ref38","doi-asserted-by":"crossref","unstructured":"Lee, B , Srivastava, S , Kumar, R , Brafman, R and Klemmer, SR (2010) Designing with interactive example galleries. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI \u201810. New York, NY, USA: Association for Computing Machinery.","DOI":"10.1145\/1753326.1753667"},{"key":"S0890060422000130_ref29","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201287"},{"key":"S0890060422000130_ref44","doi-asserted-by":"publisher","DOI":"10.1504\/JDR.2008.026987"},{"key":"S0890060422000130_ref13","doi-asserted-by":"crossref","first-page":"383","DOI":"10.1016\/j.knosys.2005.07.001","article-title":"Design sketches and sketch design tools","volume":"18","author":"Do","year":"2005","journal-title":"Knowledge-Based Systems"},{"key":"S0890060422000130_ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.cad.2011.03.010"},{"key":"S0890060422000130_ref11","doi-asserted-by":"publisher","DOI":"10.1115\/1.4003249"},{"key":"S0890060422000130_ref15","doi-asserted-by":"publisher","DOI":"10.1115\/1.4023484"},{"key":"S0890060422000130_ref42","doi-asserted-by":"publisher","DOI":"10.1016\/j.destud.2006.02.001"},{"key":"S0890060422000130_ref9","doi-asserted-by":"publisher","DOI":"10.1145\/1882261.1866205"},{"key":"S0890060422000130_ref26","doi-asserted-by":"publisher","DOI":"10.1115\/1.4044399"},{"key":"S0890060422000130_ref14","doi-asserted-by":"publisher","DOI":"10.1504\/JDR.2003.009826"},{"key":"S0890060422000130_ref47","doi-asserted-by":"publisher","DOI":"10.1016\/0950-7051(92)90026-C"},{"key":"S0890060422000130_ref7","doi-asserted-by":"publisher","DOI":"10.1115\/1.4004396"},{"key":"S0890060422000130_ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.destud.2014.08.001"},{"key":"S0890060422000130_ref19","doi-asserted-by":"publisher","DOI":"10.1017\/dsj.2016.10"},{"key":"S0890060422000130_ref45","doi-asserted-by":"publisher","DOI":"10.1115\/1.4028093"},{"key":"S0890060422000130_ref3","unstructured":"Borgianni, Y , Rotini, F and Tomassini, M (2017) Fostering ideation in the very early design phases: how textual, pictorial and combined stimuli affect creativity. Proceedings of the 21st International Conference on Engineering Design, ICED17. The Design Society."},{"key":"S0890060422000130_ref20","doi-asserted-by":"publisher","DOI":"10.1016\/j.destud.2019.01.001"},{"key":"S0890060422000130_ref37","doi-asserted-by":"publisher","DOI":"10.1115\/DETC2021-71825"},{"key":"S0890060422000130_ref24","doi-asserted-by":"publisher","DOI":"10.1017\/dsj.2018.7"},{"key":"S0890060422000130_ref17","doi-asserted-by":"publisher","DOI":"10.1017\/S0890060409000080"},{"key":"S0890060422000130_ref60","doi-asserted-by":"publisher","DOI":"10.1115\/DETC2020-22394"},{"key":"S0890060422000130_ref30","doi-asserted-by":"crossref","unstructured":"Hua, M , Han, J , Ma, X and Childs, P (2019) Exploring the effect of combinational pictorial stimuli on creative design performance. Proc. of the 22nd International Conference on Engineering Design, ICED19. Cambridge University Press.","DOI":"10.1017\/dsi.2019.182"},{"key":"S0890060422000130_ref31","doi-asserted-by":"publisher","DOI":"10.1115\/DETC2020-22048"},{"key":"S0890060422000130_ref21","doi-asserted-by":"publisher","DOI":"10.1016\/j.destud.2018.07.001"},{"key":"S0890060422000130_ref51","doi-asserted-by":"publisher","DOI":"10.1115\/1.3482061"},{"key":"S0890060422000130_ref54","doi-asserted-by":"publisher","DOI":"10.1016\/S0142-694X(96)00019-1"},{"key":"S0890060422000130_ref36","doi-asserted-by":"publisher","DOI":"10.1115\/1.4042336"},{"key":"S0890060422000130_ref1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.13081"},{"key":"S0890060422000130_ref55","doi-asserted-by":"publisher","DOI":"10.1016\/S0953-5438(98)00013-7"},{"key":"S0890060422000130_ref10","doi-asserted-by":"publisher","DOI":"10.1115\/1.4052297"},{"key":"S0890060422000130_ref46","doi-asserted-by":"publisher","DOI":"10.1017\/S0890060412000054"},{"key":"S0890060422000130_ref25","first-page":"020802","article-title":"Semantic networks for engineering design: state of the art and future directions","volume":"144","author":"Han","year":"2022","journal-title":"Journal of Mechanical Design"},{"key":"S0890060422000130_ref33","doi-asserted-by":"publisher","DOI":"10.1115\/1.4051681"},{"key":"S0890060422000130_ref41","doi-asserted-by":"publisher","DOI":"10.1145\/1121949.1121979"},{"key":"S0890060422000130_ref39","doi-asserted-by":"publisher","DOI":"10.1017\/S0890060408000061"},{"key":"S0890060422000130_ref52","doi-asserted-by":"crossref","unstructured":"Siangliulue, P , Chan, J , Gajos, KZ and Dow, SP (2015) Providing timely examples improves the quantity and quality of generated ideas. Proceedings of the 2015 ACM SIGCHI Conference on Creativity and Cognition, C&C \u201815. New York: Association for Computing Machinery.","DOI":"10.1145\/2757226.2757230"},{"key":"S0890060422000130_ref56","doi-asserted-by":"publisher","DOI":"10.1115\/1.4027639"},{"key":"S0890060422000130_ref12","unstructured":"Devlin, J , Chang, M-W , Lee, K and Toutanova, K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: human Language Technologies (NAACL-HLT). Minneapolis, MN, June 2\u20137, pp. 4171\u20134186."},{"key":"S0890060422000130_ref35","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1807185116"},{"key":"S0890060422000130_ref27","doi-asserted-by":"crossref","unstructured":"Herring, SR , Chang, CC , Krantzler, J and Bailey, BP (2009) Getting inspired! Understanding how and why examples are used in creative design practice. Proc. of SIGCHI Conference on Human Factors in Computing Systems, CHI\u201909. New York, NY, USA: Association for Computing Machinery.","DOI":"10.1145\/1518701.1518717"},{"key":"S0890060422000130_ref58","doi-asserted-by":"publisher","DOI":"10.1115\/1.4035540"},{"key":"S0890060422000130_ref16","doi-asserted-by":"publisher","DOI":"10.1115\/1.4023158"},{"key":"S0890060422000130_ref61","doi-asserted-by":"publisher","DOI":"10.1115\/DETC2021-70961"},{"key":"S0890060422000130_ref53","doi-asserted-by":"publisher","DOI":"10.1115\/1.4053683"},{"key":"S0890060422000130_ref5","doi-asserted-by":"crossref","unstructured":"Cer, D , Yang, Y , Kong, S-y , Hua, N , Limtiaco, N , St. John, R , Constant, N , GuajardoCespedes, M , Yuan, S , Tar, C , Strope, B and Kurzweil, R (2018) Universal sentence encoder for English. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Brussels, Belgium: Association for Computational Linguistics.","DOI":"10.18653\/v1\/D18-2029"},{"key":"S0890060422000130_ref6","doi-asserted-by":"publisher","DOI":"10.1017\/S0890060405050109"},{"key":"S0890060422000130_ref50","doi-asserted-by":"publisher","DOI":"10.1017\/S0890060421000020"},{"key":"S0890060422000130_ref48","doi-asserted-by":"publisher","DOI":"10.1115\/1.2722329"},{"key":"S0890060422000130_ref22","doi-asserted-by":"publisher","DOI":"10.1115\/1.4046077"},{"key":"S0890060422000130_ref4","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2018.02266"},{"key":"S0890060422000130_ref34","doi-asserted-by":"crossref","unstructured":"Kazi, RH , Grossman, T , Cheong, H , Hashemi, A and Fitzmaurice, G (2017) Dreamsketch: early stage 3D design explorations with sketching and generative design. Proc. of the 30th Annual ACM Symposium on User Interface Software and Technology, UIST \u201817. New York, NY, USA: Association for Computing Machinery.","DOI":"10.1145\/3126594.3126662"},{"key":"S0890060422000130_ref23","first-page":"51","volume-title":"Visual and Spatial Reason in Design II","author":"Gross","year":"2001"},{"key":"S0890060422000130_ref59","doi-asserted-by":"publisher","DOI":"10.1007\/978-0-85729-224-7_16"},{"key":"S0890060422000130_ref43","unstructured":"Mo, K , Zhu, S , Chang, AX , Yi, L , Tripathi, S , Guibas, LJ and Su, H (2018) Partnet: a large-scale benchmark for fine-grained and hierarchical part-level 3D object understanding. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 909\u2013918."},{"key":"S0890060422000130_ref40","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.106873"}],"container-title":["Artificial Intelligence for Engineering Design, Analysis and Manufacturing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.cambridge.org\/core\/services\/aop-cambridge-core\/content\/view\/S0890060422000130","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,12]],"date-time":"2023-02-12T18:18:26Z","timestamp":1676225906000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.cambridge.org\/core\/product\/identifier\/S0890060422000130\/type\/journal_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":61,"alternative-id":["S0890060422000130"],"URL":"https:\/\/doi.org\/10.1017\/s0890060422000130","relation":{},"ISSN":["0890-0604","1469-1760"],"issn-type":[{"value":"0890-0604","type":"print"},{"value":"1469-1760","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"Copyright \u00a9 The Author(s), 2022. Published by Cambridge University Press","name":"copyright","label":"Copyright","group":{"name":"copyright_and_licensing","label":"Copyright and Licensing"}},{"value":"This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http:\/\/creativecommons.org\/licenses\/by\/4.0\/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.","name":"license","label":"License","group":{"name":"copyright_and_licensing","label":"Copyright and Licensing"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}],"article-number":"e22"}}