{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T19:37:12Z","timestamp":1730230632059,"version":"3.28.0"},"reference-count":27,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,6,6]]},"DOI":"10.1109\/icassp39728.2021.9414884","type":"proceedings-article","created":{"date-parts":[[2021,5,13]],"date-time":"2021-05-13T15:53:45Z","timestamp":1620921225000},"page":"191-195","source":"Crossref","is-referenced-by-count":3,"title":["SepNet: A Deep Separation Matrix Prediction Network for Multichannel Audio Source Separation"],"prefix":"10.1109","author":[{"given":"Shota","family":"Inoue","sequence":"first","affiliation":[]},{"given":"Hirokazu","family":"Kameoka","sequence":"additional","affiliation":[]},{"given":"Li","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shoji","family":"Makino","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"article-title":"Semi-blind source separation with multichannel variational autoencoder","year":"2018","author":"kameoka","key":"ref10"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.23919\/APSIPA.2018.8659591"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683704"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1162\/neco_a_01217"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682623"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683497"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2009.2031510"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2013.2239990"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP40776.2020.9053556"},{"article-title":"Deep unfolding: Model-based inspiration of novel deep architectures","year":"2014","author":"hershey","key":"ref19"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ASPAA.2011.6082320"},{"key":"ref27","first-page":"1462","article-title":"Performance measurement in blind audio source separation","volume":"14","author":"vincent","year":"2006","journal-title":"IEEE Trans ASLP"},{"key":"ref3","first-page":"601","article-title":"Solution of permutation problem in frequency domain ICA, using multivariate probability density functions","author":"hiroe","year":"2006","journal-title":"Proc ICA"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2016.2577880"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-15995-4_31"},{"key":"ref8","first-page":"1652","article-title":"Multichannel audio source separation with deep neural networks","volume":"24","author":"nugraha","year":"2016","journal-title":"IEEE\/ACM Trans ASLP"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-73031-8_6"},{"key":"ref2","first-page":"165","article-title":"Independent vector analysis: An extension of ICA to multivariate components","author":"kim","year":"2006","journal-title":"Proc ICA"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO.2018.8553246"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/S0893-6080(00)00026-5"},{"key":"ref20","first-page":"241","article-title":"Permutation invariant training of deep models for speaker-independent multi-talker speech separation","author":"tan","year":"2017","journal-title":"Proc ICASSP"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-00599-2_97"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ASPAA.2005.1540173"},{"key":"ref24","doi-asserted-by":"crossref","DOI":"10.21437\/Odyssey.2018-28","article-title":"The voice conversion challenge 2018: Promoting development of parallel and nonparallel methods","author":"lorenzo-trueba","year":"2018"},{"key":"ref23","first-page":"223","article-title":"The CMU Arctic speech databases","author":"kominek","year":"2004","journal-title":"Proc SSW"},{"article-title":"Language modeling with gated convolutional networks","year":"2016","author":"dauphin","key":"ref26"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1121\/1.382599"}],"event":{"name":"ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","start":{"date-parts":[[2021,6,6]]},"location":"Toronto, ON, Canada","end":{"date-parts":[[2021,6,11]]}},"container-title":["ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9413349\/9413350\/09414884.pdf?arnumber=9414884","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T11:40:51Z","timestamp":1652182851000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9414884\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,6]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/icassp39728.2021.9414884","relation":{},"subject":[],"published":{"date-parts":[[2021,6,6]]}}}