{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T15:26:00Z","timestamp":1775229960889,"version":"3.50.1"},"publisher-location":"ISCA","reference-count":0,"publisher":"ISCA","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"DOI":"10.21437\/interspeech.2025-839","type":"proceedings-article","created":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T11:47:41Z","timestamp":1761133661000},"page":"1788-1792","source":"Crossref","is-referenced-by-count":3,"title":["SAKURA: On the Multi-hop Reasoning of Large Audio-Language Models Based on Speech and Audio Information"],"prefix":"10.21437","author":[{"given":"Chih-Kai","family":"Yang","sequence":"first","affiliation":[]},{"given":"Neo","family":"Ho","sequence":"additional","affiliation":[]},{"given":"Yen-Ting","family":"Piao","sequence":"additional","affiliation":[]},{"given":"Hung-yi","family":"Lee","sequence":"additional","affiliation":[]}],"member":"8866","published-online":{"date-parts":[[2025,8,17]]},"event":{"name":"Interspeech 2025","acronym":"interspeech_2025"},"container-title":["Interspeech 2025"],"original-title":[],"deposited":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T11:48:39Z","timestamp":1761133719000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.isca-archive.org\/interspeech_2025\/yang25g_interspeech.html"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,17]]},"references-count":0,"URL":"https:\/\/doi.org\/10.21437\/interspeech.2025-839","relation":{},"subject":[],"published":{"date-parts":[[2025,8,17]]}}}