{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,21]],"date-time":"2025-10-21T15:21:13Z","timestamp":1761060073443,"version":"3.41.0"},"reference-count":82,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2016,1,29]],"date-time":"2016-01-29T00:00:00Z","timestamp":1454025600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100001871","name":"Funda\u00e7\u00e3o para a Ci\u00eancia e a Tecnologia","doi-asserted-by":"publisher","award":["UID\/CEC\/50021\/2013"],"award-info":[{"award-number":["UID\/CEC\/50021\/2013"]}],"id":[{"id":"10.13039\/501100001871","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Funda\u00e7\u00e3o para a Ci\u00eancia e Tecnologia","award":["INCENTIVO\/EEI\/LA0021\/2014"],"award-info":[{"award-number":["INCENTIVO\/EEI\/LA0021\/2014"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Access. Comput."],"published-print":{"date-parts":[[2016,1,29]]},"abstract":"<jats:p>\n            The widespread availability of digital media has changed the way that people consume information and has impacted the consumption of auditory information. Despite this recent popularity among sighted people, the use of auditory feedback to access digital information is not new for visually impaired users. However, its sequential nature undermines both blind and sighted people\u2019s ability to efficiently find relevant information in the midst of several potentially useful items. We propose taking advantage of the\n            <jats:italic>Cocktail Party Effect<\/jats:italic>\n            , which states that people are able to focus on a single speech source among several conversations, but still identify relevant content in the background. Therefore, in contrast to one sequential speech channel, we hypothesize that people can leverage concurrent speech channels to quickly get the gist of digital information. In this article, we present an experiment with 46 (23 blind, 23 sighted) participants, which aims to understand people\u2019s ability to search for relevant content listening to two, three, or four concurrent speech channels. Our results suggest that both blind and sighted people are able to process concurrent speech in scanning scenarios. In particular, the use of two concurrent sources may be used both to identify and understand the content of the relevant sentence. Moreover, three sources may be used for most people depending on the task intelligibility demands and user characteristics. Contrasting with related work, the use of different voices did not affect the perception of concurrent speech but was highly preferred by participants. To complement the analysis, we propose a set of scenarios that may benefit from the use of concurrent speech sources, for both blind and sighted people, toward a\n            <jats:italic>Design for All<\/jats:italic>\n            paradigm.\n          <\/jats:p>","DOI":"10.1145\/2822910","type":"journal-article","created":{"date-parts":[[2016,2,1]],"date-time":"2016-02-01T20:37:54Z","timestamp":1454359074000},"page":"1-28","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":18,"title":["Scanning for Digital Content"],"prefix":"10.1145","volume":"8","author":[{"given":"Jo\u00e3o","family":"Guerreiro","sequence":"first","affiliation":[{"name":"Instituto Superior T\u00e9cnico, Universidade de Lisboa\/INESC-ID, Lisboa, Portugal"}]},{"given":"Daniel","family":"Gon\u00e7alves","sequence":"additional","affiliation":[{"name":"Instituto Superior T\u00e9cnico, Universidade de Lisboa\/INESC-ID, Lisboa, Portugal"}]}],"member":"320","published-online":{"date-parts":[[2016,1,29]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/2207016.2207052"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2380116.2380164"},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.heares.2007.01.011"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/642611.642686"},{"key":"e_1_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/244754.244758"},{"key":"e_1_2_1_6_1","volume-title":"Proceedings of ICAD (ICAD\u201903)","author":"Asakawa C.","year":"2015","unstructured":"C. Asakawa , H. Takagi , S. Ino , and T. Ifukube . 2003. Maximum listening speeds for the blind . Proceedings of ICAD (ICAD\u201903) , 276--279. Retrieved December 22, 2015 from http:\/\/www.icad.org\/Proceedings\/2003\/AsakawaTakagi2003.pdf. C. Asakawa, H. Takagi, S. Ino, and T. Ifukube. 2003. Maximum listening speeds for the blind. Proceedings of ICAD (ICAD\u201903), 276--279. Retrieved December 22, 2015 from http:\/\/www.icad.org\/Proceedings\/2003\/AsakawaTakagi2003.pdf."},{"key":"e_1_2_1_7_1","volume-title":"Retrieved","author":"Backhaus Norman","year":"2015","unstructured":"Norman Backhaus and Rico Tuor . 2015 . OLwA: Use of literature . Retrieved December 22, 2015 from http:\/\/www.geo.uzh.ch\/microsite\/olwa\/olwa\/en\/html\/unit3_kap33.html. (2015). Norman Backhaus and Rico Tuor. 2015. OLwA: Use of literature. Retrieved December 22, 2015 from http:\/\/www.geo.uzh.ch\/microsite\/olwa\/olwa\/en\/html\/unit3_kap33.html. (2015)."},{"key":"e_1_2_1_8_1","volume-title":"Neuroscience: Exploring the Brain","author":"Bear M. F.","year":"2006","unstructured":"M. F. Bear , B. W. Connors , and M. A. Paradiso . 2006 . Neuroscience: Exploring the Brain . Lippincott Williams & Wilkins , Philadelphia, PA . M. F. Bear, B. W. Connors, and M. A. Paradiso. 2006. Neuroscience: Exploring the Brain. Lippincott Williams & Wilkins, Philadelphia, PA."},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/1296843.1296854"},{"key":"e_1_2_1_10_1","volume-title":"Retrieved","author":"Boersma Paul","year":"2014","unstructured":"Paul Boersma and David Weenink . 2014 . Praat: doing phonetics by computer . Retrieved December 22, 2015 from http:\/\/www.fon.hum.uva.nl\/praat\/. (2014). Accessed in: 06- 2015. Paul Boersma and David Weenink. 2014. Praat: doing phonetics by computer. Retrieved December 22, 2015 from http:\/\/www.fon.hum.uva.nl\/praat\/. (2014). Accessed in: 06-2015."},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/1805986.1806005"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/2596695.2596728"},{"volume-title":"Auditory Scene Analysis: The Perceptual Organization of Sound","author":"Bregman A. S.","key":"e_1_2_1_13_1","unstructured":"A. S. Bregman . 1990. Auditory Scene Analysis: The Perceptual Organization of Sound . Bradford Books, MIT Press, Cambridge, MA. A. S. Bregman. 1990. Auditory Scene Analysis: The Perceptual Organization of Sound. Bradford Books, MIT Press, Cambridge, MA."},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/169059.169179"},{"key":"e_1_2_1_15_1","first-page":"117","article-title":"The cocktail party phenomenon: A review of research on speech intelligibility in multiple-talker conditions","volume":"86","author":"Bronkhorst Adelbert W.","year":"2000","unstructured":"Adelbert W. Bronkhorst . 2000 . The cocktail party phenomenon: A review of research on speech intelligibility in multiple-talker conditions . Acta Acustica United with Acustica 86 , 1, 117 -- 128 . Adelbert W. Bronkhorst. 2000. The cocktail party phenomenon: A review of research on speech intelligibility in multiple-talker conditions. Acta Acustica United with Acustica 86, 1, 117--128.","journal-title":"Acta Acustica United with Acustica"},{"key":"e_1_2_1_16_1","volume-title":"Simpson","author":"Brungart Douglas S.","year":"2005","unstructured":"Douglas S. Brungart and Brian D . Simpson . 2005 a. Improving multitalker speech communication with advanced audio displays. Air Force Research Lab Wright-Patterson AFB OH. Douglas S. Brungart and Brian D. Simpson. 2005a. Improving multitalker speech communication with advanced audio displays. Air Force Research Lab Wright-Patterson AFB OH."},{"key":"e_1_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/1101530.1101538"},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.23-10-04005.2003"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1177\/108471389800300202"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.3766\/jaaa.19.5.2"},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.tics.2004.08.008"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.1907229"},{"key":"e_1_2_1_23_1","volume-title":"Proceedings of ICAD.","author":"Crispien Kai","year":"1996","unstructured":"Kai Crispien , Klaus Fellbaum , Anthony Savidis , and Constantine Stephanidis . 1996 . A 3D-auditory environment for hierarchical navigation in non-visual interaction . Proceedings of ICAD. Kai Crispien, Klaus Fellbaum, Anthony Savidis, and Constantine Stephanidis. 1996. A 3D-auditory environment for hierarchical navigation in non-visual interaction. Proceedings of ICAD."},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1016\/S1364-6613(97)01097-8"},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.1616924"},{"key":"e_1_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.3109\/00016486409121363"},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.428503"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.1909356"},{"key":"e_1_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661334.2661363"},{"key":"e_1_2_1_30_1","volume-title":"Retrieved","author":"Gardner Bill","year":"2000","unstructured":"Bill Gardner and Keith Martin . 2000 . HRTF Measurements of a KEMAR Dummy-Head Microphone . Retrieved December 22, 2015 from http:\/\/sound.media.mit.edu\/resources\/KEMAR.html. Bill Gardner and Keith Martin. 2000. HRTF Measurements of a KEMAR Dummy-Head Microphone. Retrieved December 22, 2015 from http:\/\/sound.media.mit.edu\/resources\/KEMAR.html."},{"key":"e_1_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1207\/s15327051hci0202_3"},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/319463.319649"},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661334.2661367"},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2700648.2809840"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702123.2702373"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/2817721.2817743"},{"volume-title":"Deep accessibility: Adapting interfaces to suit our senses","author":"Harper Simon","key":"e_1_2_1_37_1","unstructured":"Simon Harper . 2012. Deep accessibility: Adapting interfaces to suit our senses . Invited Talk-Technical Superior Institute , La SIGE , Lisbon, Portugal. Simon Harper. 2012. Deep accessibility: Adapting interfaces to suit our senses. Invited Talk-Technical Superior Institute, LaSIGE, Lisbon, Portugal."},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/1090785.1090804"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cogbrainres.2003.10.015"},{"key":"e_1_2_1_40_1","volume-title":"Sonification Report: Status of the Field and Research Agenda","author":"Kramer Gregory","year":"2010","unstructured":"Gregory Kramer , Bruce Walker , Terri Bonebright , Perry Cook , John H. Flowers , Nadine Miner , and John Neuhoff . 2010 . Sonification Report: Status of the Field and Research Agenda . Faculty Publications, Department of Psychology , Paper 444. Gregory Kramer, Bruce Walker, Terri Bonebright, Perry Cook, John H. Flowers, Nadine Miner, and John Neuhoff. 2010. Sonification Report: Status of the Field and Research Agenda. Faculty Publications, Department of Psychology, Paper 444."},{"key":"e_1_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1469-8986.1997.tb02134.x"},{"volume-title":"Retrieved","year":"2015","key":"e_1_2_1_42_1","unstructured":"L2F 2015 . INESC-ID\u2019s Spoken Language Systems Laboratory . Retrieved December 22, 2015 from http:\/\/www.l2f.inesc-id.pt\/. L2F 2015. INESC-ID\u2019s Spoken Language Systems Laboratory. Retrieved December 22, 2015 from http:\/\/www.l2f.inesc-id.pt\/."},{"key":"e_1_2_1_43_1","volume-title":"Retrieved","author":"Lamb Paul","year":"2015","unstructured":"Paul Lamb . 2015 . Paul Lamb\u2019s 3D Sound System . Retrieved December 22, 2015 from http:\/\/www.paulscode.com\/forum\/index.php?topic&equals;4.0. (2015). Accessed in: 06- 2015. Paul Lamb. 2015. Paul Lamb\u2019s 3D Sound System. Retrieved December 22, 2015 from http:\/\/www.paulscode.com\/forum\/index.php?topic&equals;4.0. (2015). Accessed in: 06-2015."},{"key":"e_1_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/1357054.1357273"},{"key":"e_1_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1097\/AUD.0b013e3181a165be"},{"key":"e_1_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/1952388.1952390"},{"key":"e_1_2_1_47_1","volume-title":"Retrieved","author":"LWJGL","year":"2015","unstructured":"LWJGL 2.0. 2015 . LightWeight Java Game Library . Retrieved December 22, 2015 from http:\/\/legacy.lwjgl.org\/. LWJGL 2.0. 2015. LightWeight Java Game Library. Retrieved December 22, 2015 from http:\/\/legacy.lwjgl.org\/."},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/1216295.1216362"},{"volume-title":"Handbook of perception and cognition","author":"Moore B. C. J.","key":"e_1_2_1_49_1","unstructured":"B. C. J. Moore . 1995. Hearing. Handbook of perception and cognition . Academic Press . B. C. J. Moore. 1995. Hearing. Handbook of perception and cognition. Academic Press."},{"key":"e_1_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.heares.2011.01.003"},{"key":"e_1_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.3109\/00206098109072719"},{"key":"e_1_2_1_52_1","volume-title":"Retrieved","author":"Soft AL","year":"2014","unstructured":"Open AL Soft 2014 . OpenAL Soft 1.15.1 . Retrieved December 22, 2015 from http:\/\/kcat.strangesoft.net\/openal.html. OpenAL Soft 2014. OpenAL Soft 1.15.1. Retrieved December 22, 2015 from http:\/\/kcat.strangesoft.net\/openal.html."},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.3109\/07434618.2010.522200"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/1127564.1127571"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-85980-2_10"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1080\/10691316.2010.550529"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/1978942.1979353"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/223355.223533"},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4718637"},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1080\/14640746708400069"},{"key":"e_1_2_1_61_1","unstructured":"B. Shinn-Cunningham and A. Ihlefeld. 2004. Selective and divided attention: Extracting information from simultaneous sound sources. (2004).  B. Shinn-Cunningham and A. Ihlefeld. 2004. Selective and divided attention: Extracting information from simultaneous sound sources. (2004)."},{"key":"e_1_2_1_62_1","volume-title":"Proceedings of the World Congress on Engineering and Computer Science","author":"Sodnik Jaka","year":"2009","unstructured":"Jaka Sodnik and Sa\u0161o Toma\u017ei\u010d . 2009 . Spatial speaker : 3D Java text-to-speech converter . In Proceedings of the World Congress on Engineering and Computer Science , Vol. II . Jaka Sodnik and Sa\u0161o Toma\u017ei\u010d. 2009. Spatial speaker : 3D Java text-to-speech converter. In Proceedings of the World Congress on Engineering and Computer Science, Vol. II."},{"key":"e_1_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/2049536.2049574"},{"key":"e_1_2_1_64_1","first-page":"3","article-title":"User interfaces for all: New perspectives into human-- computer interaction","volume":"1","author":"Stephanidis Constantine","year":"2001","unstructured":"Constantine Stephanidis . 2001 . User interfaces for all: New perspectives into human-- computer interaction . User Interfaces for All-Concepts, Methods, and Tools 1 , 3 -- 17 . Constantine Stephanidis. 2001. User interfaces for all: New perspectives into human-- computer interaction. User Interfaces for All-Concepts, Methods, and Tools 1, 3--17.","journal-title":"User Interfaces for All-Concepts, Methods, and Tools"},{"key":"e_1_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/1279700.1279703"},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1152\/jn.01125.2002"},{"key":"e_1_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jphysparis.2004.03.009"},{"key":"e_1_2_1_68_1","first-page":"5","article-title":"On the comprehension of extremely fast synthetic speech","volume":"1","author":"Trouvain J.","year":"2007","unstructured":"J. Trouvain . 2007 . On the comprehension of extremely fast synthetic speech . Saarland Working Papers in Linguistics 1 , 5 -- 13 . http:\/\/scidok.sulb.uni-saarland.de\/volltexte\/2007\/1176\/ J. Trouvain. 2007. On the comprehension of extremely fast synthetic speech. Saarland Working Papers in Linguistics 1, 5--13. http:\/\/scidok.sulb.uni-saarland.de\/volltexte\/2007\/1176\/","journal-title":"Saarland Working Papers in Linguistics"},{"key":"e_1_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1037\/0096-1523.31.5.939"},{"key":"e_1_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/1978942.1979258"},{"key":"e_1_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.3050321"},{"key":"e_1_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2013.08.002"},{"key":"e_1_2_1_73_1","volume-title":"Retrieved","author":"Interaction Voice","year":"2014","unstructured":"Voice Interaction 2014 . Voice Interaction . Retrieved December 22, 2015 from http:\/\/www.voiceinteraction.eu\/. Voice Interaction 2014. Voice Interaction. Retrieved December 22, 2015 from http:\/\/www.voiceinteraction.eu\/."},{"key":"e_1_2_1_74_1","volume-title":"Proceedings of ICAD. 63--68","author":"Walker B. N.","year":"2015","unstructured":"B. N. Walker , A. Nance , and J. Lindsay . 2006. Spearcons: Speech-based earcons improve navigation performance in auditory menus . Proceedings of ICAD. 63--68 . Retrieved December 22, 2015 from http:\/\/sonify.psych.gatech.edu\/&sim;walkerb\/publications\/pdfs\/2006ICAD-WalkerNanceLindsay.pdf. B. N. Walker, A. Nance, and J. Lindsay. 2006. Spearcons: Speech-based earcons improve navigation performance in auditory menus. Proceedings of ICAD. 63--68. Retrieved December 22, 2015 from http:\/\/sonify.psych.gatech.edu\/&sim;walkerb\/publications\/pdfs\/2006ICAD-WalkerNanceLindsay.pdf."},{"key":"e_1_2_1_75_1","doi-asserted-by":"crossref","unstructured":"Takayuki Watanabe. 2007. Experimental evaluation of usability and accessibility of heading elements components of web accessibility. Disability & Rehabilitation: Assistive Technology 1--8.  Takayuki Watanabe. 2007. Experimental evaluation of usability and accessibility of heading elements components of web accessibility. Disability & Rehabilitation: Assistive Technology 1--8.","DOI":"10.1145\/1243441.1243473"},{"volume-title":"WAIS-R Manual: Wechsler Adult Intelligence Scale-Revised","author":"Wechsler David","key":"e_1_2_1_76_1","unstructured":"David Wechsler . 1981. WAIS-R Manual: Wechsler Adult Intelligence Scale-Revised . Harcourt Brace Jovanovich, New York, NY. David Wechsler. 1981. WAIS-R Manual: Wechsler Adult Intelligence Scale-Revised. Harcourt Brace Jovanovich, New York, NY."},{"key":"e_1_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.407089"},{"key":"e_1_2_1_78_1","unstructured":"Elizabeth Grace Winkler. 2008. Understanding language. Continuum International.  Elizabeth Grace Winkler. 2008. Understanding language. Continuum International."},{"key":"e_1_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.211220498"},{"key":"e_1_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1145\/1279700.1279704"},{"key":"e_1_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1097\/00003446-199804000-00009"},{"key":"e_1_2_1_82_1","first-page":"255","article-title":"Binaural advantages and directional effects in speech intelligibility","volume":"2","author":"Zurek P. M.","year":"1993","unstructured":"P. M. Zurek . 1993 . Binaural advantages and directional effects in speech intelligibility . Acoustical Factors Affecting Hearing Aid Performance 2 , 255 -- 275 . P. M. Zurek. 1993. Binaural advantages and directional effects in speech intelligibility. Acoustical Factors Affecting Hearing Aid Performance 2, 255--275.","journal-title":"Acoustical Factors Affecting Hearing Aid Performance"}],"container-title":["ACM Transactions on Accessible Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/2822910","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/2822910","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T05:48:31Z","timestamp":1750225711000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/2822910"}},"subtitle":["How Blind and Sighted People Perceive Concurrent Speech"],"short-title":[],"issued":{"date-parts":[[2016,1,29]]},"references-count":82,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2016,1,29]]}},"alternative-id":["10.1145\/2822910"],"URL":"https:\/\/doi.org\/10.1145\/2822910","relation":{},"ISSN":["1936-7228","1936-7236"],"issn-type":[{"type":"print","value":"1936-7228"},{"type":"electronic","value":"1936-7236"}],"subject":[],"published":{"date-parts":[[2016,1,29]]},"assertion":[{"value":"2015-03-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2015-09-01","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2016-01-29","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}