{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T10:35:48Z","timestamp":1763202948313,"version":"3.40.3"},"publisher-location":"Cham","reference-count":40,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031703409"},{"type":"electronic","value":"9783031703416"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-70341-6_25","type":"book-chapter","created":{"date-parts":[[2024,8,30]],"date-time":"2024-08-30T20:26:39Z","timestamp":1725049599000},"page":"422-439","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["The Future is Different: Predicting Reddits Popularity with\u00a0Variational Dynamic Language Models"],"prefix":"10.1007","author":[{"given":"Kostadin","family":"Cvejoski","sequence":"first","affiliation":[]},{"given":"Rams\u00e9s J.","family":"S\u00e1nchez","sequence":"additional","affiliation":[]},{"given":"C\u00e9sar","family":"Ojeda","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,22]]},"reference":[{"key":"25_CR1","doi-asserted-by":"publisher","first-page":"904","DOI":"10.1162\/tacl_a_00497","volume":"10","author":"O Agarwal","year":"2022","unstructured":"Agarwal, O., Nenkova, A.: Temporal effects on pre-trained models for language processing tasks. Trans. Assoc. Comput. Linguist. 10, 904\u2013921 (2022)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"25_CR2","doi-asserted-by":"publisher","unstructured":"Amba, S., Chen, H.T., Zhang, M., Bendersky, M., Najork, M., Ben, M.: Dynamic language models for continuously evolving content; dynamic language models for continuously evolving content, vol. 11 (2021). https:\/\/doi.org\/10.1145\/3447548.3467162","DOI":"10.1145\/3447548.3467162"},{"key":"25_CR3","doi-asserted-by":"crossref","unstructured":"Amba\u00a0Hombaiah, S., Chen, T., Zhang, M., Bendersky, M., Najork, M.: Dynamic language models for continuously evolving content. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 2514\u20132524 (2021)","DOI":"10.1145\/3447548.3467162"},{"key":"25_CR4","unstructured":"Bishop, C.M.: Pattern recognition and machine learning. Springer (2006)"},{"key":"25_CR5","doi-asserted-by":"crossref","unstructured":"Blei, D.M., Lafferty, J.D.: Dynamic topic models. In: Proceedings of the 23rd International Conference on Machine Learning, pp. 113\u2013120 (2006)","DOI":"10.1145\/1143844.1143859"},{"key":"25_CR6","unstructured":"Brown, T., et al.: Language models are few-shot learners. In: Advances in Neural Information Processing Systems, vol.\u00a033 (2020)"},{"key":"25_CR7","unstructured":"Chawla, S., Singh, N., Drori, I.: Quantifying and alleviating distribution shifts in foundation models on review classification. In: NeurIPS 2021 Workshop on Distribution Shifts: Connecting Methods and Applications (2021). https:\/\/openreview.net\/forum?id=OG78-TuPcvL"},{"key":"25_CR8","doi-asserted-by":"crossref","unstructured":"Cho, K., Van\u00a0Merri\u00ebnboer, B., Bahdanau, D., Bengio, Y.: On the properties of neural machine translation: Encoder-decoder approaches (2014). arXiv preprint arXiv:1409.1259","DOI":"10.3115\/v1\/W14-4012"},{"key":"25_CR9","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1007\/978-3-658-36295-9_10","volume-title":"Data Science \u2013 Analytics and Applications","author":"K Cvejoski","year":"2022","unstructured":"Cvejoski, K., S\u00e1nchez, R.J., Bauckhage, C., Ojeda, C.: Dynamic review-based recommenders. In: Data Science \u2013 Analytics and Applications, pp. 66\u201371. Springer, Wiesbaden (2022). https:\/\/doi.org\/10.1007\/978-3-658-36295-9_10"},{"key":"25_CR10","doi-asserted-by":"publisher","unstructured":"Cvejoski, K., S\u00e1nchez, R.J., Georgiev, B., Bauckhage, C., Ojeda, C.: Recurrent point review models. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp.\u00a01\u20138 (2020). https:\/\/doi.org\/10.1109\/IJCNN48605.2020.9206768","DOI":"10.1109\/IJCNN48605.2020.9206768"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"Danescu-Niculescu-Mizil, C., West, R., Jurafsky, D., Leskovec, J., Potts, C.: No country for old members: user lifecycle and linguistic change in online communities. In: Proceedings of the 22nd International Conference on World Wide Web, pp. 307\u2013318. Association for Computing Machinery, New York, NY, USA (2013)","DOI":"10.1145\/2488388.2488416"},{"key":"25_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"282","DOI":"10.1007\/978-3-030-36718-3_24","volume-title":"Neural Information Processing","author":"E Delasalles","year":"2019","unstructured":"Delasalles, E., Lamprier, S., Denoyer, L.: Dynamic neural language models. In: Gedeon, T., Wong, K.W., Lee, M. (eds.) ICONIP 2019. LNCS, vol. 11955, pp. 282\u2013294. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-36718-3_24"},{"key":"25_CR13","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding (2018). arXiv preprint arXiv:1810.04805"},{"key":"25_CR14","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1162\/tacl_a_00459","volume":"10","author":"B Dhingra","year":"2022","unstructured":"Dhingra, B., Cole, J.R., Eisenschlos, J.M., Gillick, D., Eisenstein, J., Cohen, W.W.: Time-aware language models as temporal knowledge bases. Trans. Assoc. Comput. Linguist. 10, 257\u2013273 (2022)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"25_CR15","doi-asserted-by":"crossref","unstructured":"Guo, H., Zhu, H., Guo, Z., Zhang, X., Wu, X., Su, Z.: Domain adaptation with latent semantic association for named entity recognition. In: Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp. 281\u2013289. Association for Computational Linguistics, Boulder, Colorado (2009). https:\/\/aclanthology.org\/N09-1032","DOI":"10.3115\/1620754.1620795"},{"key":"25_CR16","doi-asserted-by":"publisher","unstructured":"Hendrycks, D., Liu, X., Wallace, E., Dziedzic, A., Krishnan, R., Song, D.: Pretrained transformers improve out-of-distribution robustness. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 2744\u20132751. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.244, https:\/\/aclanthology.org\/2020.acl-main.244","DOI":"10.18653\/v1\/2020.acl-main.244"},{"key":"25_CR17","doi-asserted-by":"publisher","unstructured":"Hofmann, V., Pierrehumbert, J., Sch\u00fctze, H.: Dynamic contextualized word embeddings. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6970\u20136984. Association for Computational Linguistics, Online (2021). https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.542, https:\/\/aclanthology.org\/2021.acl-long.542","DOI":"10.18653\/v1\/2021.acl-long.542"},{"key":"25_CR18","doi-asserted-by":"publisher","unstructured":"Hu, Y., Zhai, K., Eidelman, V., Boyd-Graber, J.: Polylingual tree-based topic models for translation domain adaptation. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1166\u20131176. Association for Computational Linguistics, Baltimore, Maryland (2014). https:\/\/doi.org\/10.3115\/v1\/P14-1110, https:\/\/aclanthology.org\/P14-1110","DOI":"10.3115\/v1\/P14-1110"},{"key":"25_CR19","unstructured":"Koh, P.W., et al.: WILDS: a benchmark of in-the-wild distribution shifts. In: Meila, M., Zhang, T. (eds.) Proceedings of the 38th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol.\u00a0139, pp. 5637\u20135664. PMLR (2021). https:\/\/proceedings.mlr.press\/v139\/koh21a.html"},{"key":"25_CR20","unstructured":"Krishnan, R.G., Shalit, U., Sontag, D.: Deep Kalman filters (2015)"},{"key":"25_CR21","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: ALBERT: A lite BERT for self-supervised learning of language representations (2019). arXiv preprint arXiv:1909.11942"},{"key":"25_CR22","first-page":"29348","volume":"34","author":"A Lazaridou","year":"2021","unstructured":"Lazaridou, A., et al.: Mind the Gap: assessing temporal generalization in neural language models. Adv. Neural. Inf. Process. Syst. 34, 29348\u201329363 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"25_CR23","unstructured":"Liska, A., et\u00a0al.: StreamingQA: a benchmark for adaptation to new knowledge over time in question answering models. In: International Conference on Machine Learning, pp. 13604\u201313622. PMLR (2022)"},{"key":"25_CR24","unstructured":"Liu, Y., et al.: RoBERTa: A robustly optimized BERT pretraining approach (2019). arXiv preprint arXiv:1907.11692"},{"key":"25_CR25","doi-asserted-by":"publisher","unstructured":"Loureiro, D., Barbieri, F., Neves, L., Anke, L.E., Camacho-Collados, J.: TimeLMS: Diachronic language models from twitter, pp. 251\u2013260 (2022). https:\/\/doi.org\/10.48550\/arxiv.2202.03829, https:\/\/arxiv.org\/abs\/2202.03829v2","DOI":"10.48550\/arxiv.2202.03829"},{"key":"25_CR26","doi-asserted-by":"crossref","unstructured":"Luu, K., Khashabi, D., Gururangan, S., Mandyam, K., Smith, N.A.: Time waits for no one! analysis and challenges of temporal misalignment (2021). arXiv preprint arXiv:2111.07408","DOI":"10.18653\/v1\/2022.naacl-main.435"},{"key":"25_CR27","doi-asserted-by":"publisher","unstructured":"Ma, X., Xu, P., Wang, Z., Nallapati, R., Xiang, B.: Domain adaptation with BERT-based domain classification and data selection. In: Proceedings of the 2nd Workshop on Deep Learning Approaches for Low-Resource NLP (DeepLo 2019), pp. 76\u201383. Association for Computational Linguistics, Hong Kong, China (2019). https:\/\/doi.org\/10.18653\/v1\/D19-6109, https:\/\/aclanthology.org\/D19-6109","DOI":"10.18653\/v1\/D19-6109"},{"key":"25_CR28","doi-asserted-by":"publisher","unstructured":"Oren, Y., Sagawa, S., Hashimoto, T.B., Liang, P.: Distributionally robust language modeling. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 4227\u20134237. Association for Computational Linguistics, Hong Kong, China (2019). https:\/\/doi.org\/10.18653\/v1\/D19-1432, https:\/\/aclanthology.org\/D19-1432","DOI":"10.18653\/v1\/D19-1432"},{"key":"25_CR29","doi-asserted-by":"publisher","unstructured":"Peters, M.E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., Zettlemoyer, L.: Deep contextualized word representations. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp. 2227\u20132237. Association for Computational Linguistics, New Orleans, Louisiana (2018). https:\/\/doi.org\/10.18653\/v1\/N18-1202, https:\/\/aclanthology.org\/N18-1202","DOI":"10.18653\/v1\/N18-1202"},{"issue":"10","key":"25_CR30","doi-asserted-by":"publisher","first-page":"1872","DOI":"10.1007\/s11431-020-1647-3","volume":"63","author":"X Qiu","year":"2020","unstructured":"Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., Huang, X.: Pre-trained models for natural language processing: a survey. SCIENCE CHINA Technol. Sci. 63(10), 1872\u20131897 (2020)","journal-title":"SCIENCE CHINA Technol. Sci."},{"key":"25_CR31","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I., et\u00a0al.: Improving language understanding by generative pre-training (2018)"},{"issue":"8","key":"25_CR32","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)","journal-title":"OpenAI Blog"},{"key":"25_CR33","doi-asserted-by":"publisher","unstructured":"Rosin, G.D., Guy, I., Radinsky, K.: Time masking for temporal language models. In: WSDM 2022 - Proceedings of the 15th ACM International Conference on Web Search and Data Mining, pp. 833\u2013841 (10 2021). https:\/\/doi.org\/10.48550\/arxiv.2110.06366, https:\/\/arxiv.org\/abs\/2110.06366v4","DOI":"10.48550\/arxiv.2110.06366"},{"key":"25_CR34","doi-asserted-by":"publisher","unstructured":"Rosin, G.D., Radinsky, K.: Temporal attention for language models. In: Findings of the Association for Computational Linguistics: NAACL 2022. pp. 1498\u20131508. Association for Computational Linguistics, Seattle, United States (2022).https:\/\/doi.org\/10.18653\/v1\/2022.findings-naacl.112, https:\/\/aclanthology.org\/2022.findings-naacl.112","DOI":"10.18653\/v1\/2022.findings-naacl.112"},{"key":"25_CR35","doi-asserted-by":"publisher","unstructured":"R\u00f6ttger, P., Pierrehumbert, J.: Temporal adaptation of BERT and performance on downstream document classification: Insights from social media. In: Findings of the Association for Computational Linguistics: EMNLP 2021, pp. 2400\u20132412. Association for Computational Linguistics, Punta Cana, Dominican Republic (2021). https:\/\/doi.org\/10.18653\/v1\/2021.findings-emnlp.206, https:\/\/aclanthology.org\/2021.findings-emnlp.206","DOI":"10.18653\/v1\/2021.findings-emnlp.206"},{"key":"25_CR36","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"25_CR37","unstructured":"Wang, X., YANG, Y.: Neural topic model with attention for supervised learning. In: Chiappa, S., Calandra, R. (eds.) Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol.\u00a0108, pp. 1147\u20131156. PMLR (2020). https:\/\/proceedings.mlr.press\/v108\/wang20c.html"},{"key":"25_CR38","doi-asserted-by":"crossref","unstructured":"Wu, C.Y., Ahmed, A., Beutel, A., Smola, A.J.: Joint training of ratings and reviews with recurrent recommender networks (2016)","DOI":"10.1145\/3018661.3018689"},{"key":"25_CR39","doi-asserted-by":"publisher","unstructured":"Yogatama, D., Wang, C., Routledge, B.R., Smith, N.A., Xing, E.P.: Dynamic Lang. Models Streaming Text. Trans. Assoc. Comput. Linguist. 2, 181\u2013192 (2014). https:\/\/doi.org\/10.1162\/tacl_a_00175","DOI":"10.1162\/tacl_a_00175"},{"key":"25_CR40","doi-asserted-by":"publisher","unstructured":"Zhou, W., Liu, F., Chen, M.: Contrastive out-of-distribution detection for pretrained transformers. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 1100\u20131111. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (2021). https:\/\/doi.org\/10.18653\/v1\/2021.emnlp-main.84, https:\/\/aclanthology.org\/2021.emnlp-main.84","DOI":"10.18653\/v1\/2021.emnlp-main.84"}],"container-title":["Lecture Notes in Computer Science","Machine Learning and Knowledge Discovery in Databases. Research Track"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-70341-6_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,30]],"date-time":"2024-08-30T20:31:30Z","timestamp":1725049890000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-70341-6_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031703409","9783031703416"],"references-count":40,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-70341-6_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"22 August 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECML PKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vilnius","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Lithuania","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 September 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecml2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/2024.ecmlpkdd.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}