{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T21:57:01Z","timestamp":1769551021712,"version":"3.49.0"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,5,1]],"date-time":"2020-05-01T00:00:00Z","timestamp":1588291200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,5]]},"DOI":"10.1109\/icassp40776.2020.9053975","type":"proceedings-article","created":{"date-parts":[[2020,4,9]],"date-time":"2020-04-09T16:21:13Z","timestamp":1586449273000},"page":"8034-8038","source":"Crossref","is-referenced-by-count":17,"title":["A Simple But Effective Bert Model for Dialog State Tracking on Resource-Limited Systems"],"prefix":"10.1109","author":[{"given":"Tuan Manh","family":"Lai","sequence":"first","affiliation":[]},{"given":"Quan","family":"Hung Tran","sequence":"additional","affiliation":[]},{"given":"Trung","family":"Bui","sequence":"additional","affiliation":[]},{"given":"Daisuke","family":"Kihara","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/W14-4337"},{"key":"ref11","article-title":"A network-based end-to-end trainable task-oriented dialogue system","author":"wen","year":"2016","journal-title":"EACL"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"key":"ref13","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"devlin","year":"2019","journal-title":"NAACL-HLT"},{"key":"ref14","article-title":"Distilling the knowledge in a neural network","volume":"abs 1503 2531","author":"hinton","year":"2015","journal-title":"ArXiv"},{"key":"ref15","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"NIPS"},{"key":"ref16","article-title":"Do deep nets really need to be deep&#x0192;","author":"ba","year":"2013","journal-title":"NIPS"},{"key":"ref17","article-title":"Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding","volume":"abs 1510 149","author":"han","year":"2015","journal-title":"CoRR"},{"key":"ref18","article-title":"Learning both weights and connections for efficient neural networks","author":"han","year":"2015","journal-title":"NIPS"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.11"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-1135"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2017-1326"},{"key":"ref6","article-title":"Toward scalable neural dialogue state tracking model","author":"nouri","year":"2018","journal-title":"arXiv preprint arXiv 1812 02588"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1299"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683412"},{"key":"ref7","article-title":"Fully statistical neural belief tracking","author":"mrk\u0161i?","year":"2018","journal-title":"ACL"},{"key":"ref2","article-title":"Neural belief tracker: Data-driven dialogue state tracking","author":"mrksic","year":"2016","journal-title":"ACL"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2019-1355"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2012.2225812"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2012.6289079"},{"key":"ref22","article-title":"Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter","volume":"abs 1910 1108","author":"sanh","year":"2019","journal-title":"ArXiv"},{"key":"ref21","article-title":"Distilling task-specific knowledge from bert into simple neural networks","volume":"abs 1903 12136","author":"tang","year":"2019","journal-title":"ArXiv"},{"key":"ref24","article-title":"Mixed precision training","volume":"abs 1710 3740","author":"micikevicius","year":"2017","journal-title":"ArXiv"},{"key":"ref23","year":"0","journal-title":"Core ml models"},{"key":"ref25","article-title":"Multiwoz - a large-scale multidomain wizard-of-oz dataset for task-oriented dialogue modelling","author":"budzianowski","year":"2018","journal-title":"EMNLP"}],"event":{"name":"ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","location":"Barcelona, Spain","start":{"date-parts":[[2020,5,4]]},"end":{"date-parts":[[2020,5,8]]}},"container-title":["ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9040208\/9052899\/09053975.pdf?arnumber=9053975","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,27]],"date-time":"2022-06-27T20:11:45Z","timestamp":1656360705000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9053975\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,5]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/icassp40776.2020.9053975","relation":{},"subject":[],"published":{"date-parts":[[2020,5]]}}}