{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T07:28:58Z","timestamp":1743060538267,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031044465"},{"type":"electronic","value":"9783031044472"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-04447-2_19","type":"book-chapter","created":{"date-parts":[[2022,4,21]],"date-time":"2022-04-21T16:05:24Z","timestamp":1650557124000},"page":"281-295","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["AmLDA: A Non-VAE Neural Topic Model"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8358-3699","authenticated-orcid":false,"given":"Tomonari","family":"Masada","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,20]]},"reference":[{"key":"19_CR1","doi-asserted-by":"crossref","unstructured":"Bianchi, F., Terragni, S., Hovy, D.: Pre-training is a hot topic: contextualized document embeddings improve topic coherence. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing - Volume 2, pp. 759\u2013766 (2021)","DOI":"10.18653\/v1\/2021.acl-short.96"},{"key":"19_CR2","first-page":"993","volume":"3","author":"DM Blei","year":"2003","unstructured":"Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, 993\u20131022 (2003)","journal-title":"J. Mach. Learn. Res."},{"issue":"1","key":"19_CR3","first-page":"1","volume":"2","author":"MR Brett","year":"2012","unstructured":"Brett, M.R.: Topic modeling: a basic introduction. J. Digital Humanities 2(1), 1\u20132 (2012)","journal-title":"J. Digital Humanities"},{"key":"19_CR4","doi-asserted-by":"publisher","first-page":"439","DOI":"10.1162\/tacl_a_00325","volume":"8","author":"AB Dieng","year":"2020","unstructured":"Dieng, A.B., Ruiz, F.J.R., Blei, D.M.: Topic modeling in embedding spaces. Trans. Assoc. Comput. Linguist. 8, 439\u2013453 (2020)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"19_CR5","unstructured":"Dieng, A.B., Wang, C., Gao, J., Paisley, J.W.: TopicRNN: a recurrent neural network with long-range semantic dependency. CoRR abs\/1611.01702 (2016)"},{"key":"19_CR6","unstructured":"Hoffman, M.D., Blei, D.M., Bach, F.: Online learning for Latent Dirichlet allocation. In: Proceedings of the 23rd International Conference on Neural Information Processing Systems - Volume 1, pp. 856\u2013864 (2010)"},{"key":"19_CR7","unstructured":"Hoffman, M.D., Blei, D.M., Wang, C., Paisley, J.: Stochastic variational inference. J. Mach. Learn. Res. 14(5) (2013)"},{"key":"19_CR8","doi-asserted-by":"crossref","unstructured":"Kim, Y., Dyer, C., Rush, A.: Compound probabilistic context-free grammars for grammar induction. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 2369\u20132385 (2019)","DOI":"10.18653\/v1\/P19-1228"},{"key":"19_CR9","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR abs\/1412.6980 (2014)"},{"key":"19_CR10","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. CoRR abs\/1312.6114 (2013)"},{"key":"19_CR11","doi-asserted-by":"crossref","unstructured":"Li, B., He, J., Neubig, G., Berg-Kirkpatrick, T., Yang, Y.: A surprisingly effective fix for deep latent variable modeling of text. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3603\u20133614 (2019)","DOI":"10.18653\/v1\/D19-1370"},{"key":"19_CR12","doi-asserted-by":"crossref","unstructured":"Lindstedt, N.C.: Structural topic modeling for social scientists: a brief case study with social movement studies literature, 2005\u20132017. Soc. Currents 6(4), 307\u2013318 (2019)","DOI":"10.1177\/2329496519846505"},{"key":"19_CR13","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: 7th International Conference on Learning Representations (2019)"},{"key":"19_CR14","doi-asserted-by":"crossref","unstructured":"McCarthy, A.D., Li, X., Gu, J., Dong, N.: Addressing posterior collapse with mutual information for improved variational neural machine translation. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8512\u20138525 (2020)","DOI":"10.18653\/v1\/2020.acl-main.753"},{"key":"19_CR15","unstructured":"Miao, Y., Grefenstette, E., Blunsom, P.: Discovering discrete latent topics with neural variational inference. In: Proceedings of the 34th International Conference on Machine Learning, pp. 2410\u20132419 (2017)"},{"key":"19_CR16","unstructured":"Miao, Y., Yu, L., Blunsom, P.: Neural variational inference for text processing. In: Proceedings of the 33rd International Conference on Machine Learning, pp. 1727\u20131736 (2016)"},{"key":"19_CR17","doi-asserted-by":"crossref","unstructured":"Nallapati, R., Cohen, W., Lafferty, J.: Parallelized variational EM for latent Dirichlet allocation: an experimental evaluation of speed and scalability. In: 7th IEEE International Conference on Data Mining Workshops, pp. 349\u2013354 (2007)","DOI":"10.1109\/ICDMW.2007.33"},{"key":"19_CR18","doi-asserted-by":"crossref","unstructured":"Nan, F., Ding, R., Nallapati, R., Xiang, B.: Topic modeling with Wasserstein autoencoders. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. pp. 6345\u20136381 (2019)","DOI":"10.18653\/v1\/P19-1640"},{"key":"19_CR19","doi-asserted-by":"crossref","unstructured":"Ponte, J.M., Croft, W.B.: A language modeling approach to information retrieval. In: Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 275\u2013281 (1998)","DOI":"10.1145\/290941.291008"},{"key":"19_CR20","unstructured":"Razavi, A., van den Oord, A., Poole, B., Vinyals, O.: Preventing posterior collapse with delta-VAEs. In: 7th International Conference on Learning Representations (2019)"},{"key":"19_CR21","doi-asserted-by":"crossref","unstructured":"Rezaee, M., Ferraro, F.: A discrete variational recurrent topic model without the reparametrization trick. In: Advances in Neural Information Processing Systems, vol. 33, pp. 13831\u201313843 (2020)","DOI":"10.1007\/s00521-021-06020-8"},{"issue":"515","key":"19_CR22","doi-asserted-by":"publisher","first-page":"988","DOI":"10.1080\/01621459.2016.1141684","volume":"111","author":"ME Roberts","year":"2016","unstructured":"Roberts, M.E., Stewart, B.M., Airoldi, E.M.: A model of text for experimentation in the social sciences. J. Am. Stat. Assoc. 111(515), 988\u20131003 (2016)","journal-title":"J. Am. Stat. Assoc."},{"key":"19_CR23","unstructured":"Srivastava, A., Sutton, C.: Autoencoding variational inference for topic models. In: 5th International Conference on Learning Representations (2017)"},{"key":"19_CR24","unstructured":"Stone, H., Sports, T.: The Push and Pull of Digital Humanities: Topic Modeling the \"What is digital humanities?\" Genre. DHQ 14(1) (2020)"},{"key":"19_CR25","unstructured":"Wang, W., et al.: Topic compositional neural language model. In: Proceedings of the 21st International Conference on Artificial Intelligence and Statistics, pp. 356\u2013365 (2018)"},{"key":"19_CR26","doi-asserted-by":"crossref","unstructured":"Wang, W., Gan, Z., Xu, H., Zhang, R., Wang, G., Shen, D., Chen, C., Carin, L.: Topic-guided variational auto-encoder for text generation. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics, pp. 166\u2013177 (2019)","DOI":"10.18653\/v1\/N19-1015"},{"key":"19_CR27","unstructured":"Xu, H., Wang, W., Liu, W., Carin, L.: Distilled wasserstein learning for word embedding and topic modeling. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 1723\u20131732 (2018)"}],"container-title":["Communications in Computer and Information Science","Information Management and Big Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-04447-2_19","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,21]],"date-time":"2022-04-21T16:16:58Z","timestamp":1650557818000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-04447-2_19"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031044465","9783031044472"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-04447-2_19","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"20 April 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"SIMBig","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Annual International Conference on Information Management and Big Data","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"simbig2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/simbig.org\/SIMBig2021\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"67","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"25","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}