{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T14:30:23Z","timestamp":1726065023615},"publisher-location":"Cham","reference-count":16,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030389185"},{"type":"electronic","value":"9783030389192"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-38919-2_53","type":"book-chapter","created":{"date-parts":[[2020,1,16]],"date-time":"2020-01-16T12:03:18Z","timestamp":1579176198000},"page":"639-647","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Automatic Text Generation in\u00a0Slovak\u00a0Language"],"prefix":"10.1007","author":[{"given":"Dominik","family":"Vasko","sequence":"first","affiliation":[]},{"given":"Samuel","family":"Pecar","sequence":"additional","affiliation":[]},{"given":"Marian","family":"Simko","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,1,17]]},"reference":[{"issue":"Feb","key":"53_CR1","first-page":"1137","volume":"3","author":"Y Bengio","year":"2003","unstructured":"Bengio, Y., Ducharme, R., Vincent, P., Jauvin, C.: A neural probabilistic language model. J. Mach. Learn. Res. 3(Feb), 1137\u20131155 (2003)","journal-title":"J. Mach. Learn. Res."},{"key":"53_CR2","unstructured":"Brants, T., Popat, A.C., Xu, P., Och, F.J., Dean, J.: Large language models in machine translation. In: Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL) (2007). http:\/\/aclweb.org\/anthology\/D07-1090"},{"key":"53_CR3","unstructured":"Chung, J., Gulcehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. In: NIPS 2014 Workshop on Deep Learning, December 2014"},{"key":"53_CR4","doi-asserted-by":"publisher","unstructured":"Cotterell, R., Mielke, S.J., Eisner, J., Roark, B.: Are all languages equally hard to language-model? In: Proceedings of the 2018 Conference of the NAACL: Human Language Technologies, Volume 2 (Short Papers), pp. 536\u2013541. ACL, New Orleans, June 2018. https:\/\/doi.org\/10.18653\/v1\/N18-2085","DOI":"10.18653\/v1\/N18-2085"},{"key":"53_CR5","unstructured":"Galu\u0161\u010d\u00e1kov\u00e1, P., Garab\u00edk, R., Bojar, O.: English-Slovak parallel corpus (2012). http:\/\/hdl.handle.net\/11858\/00-097C-0000-0006-AAE0-A"},{"issue":"8","key":"53_CR6","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997). https:\/\/doi.org\/10.1162\/neco.1997.9.8.1735","journal-title":"Neural Comput."},{"key":"53_CR7","unstructured":"J\u00f3zefowicz, R., Vinyals, O., Schuster, M., Shazeer, N., Wu, Y.: Exploring the limits of language modeling (2016). http:\/\/arxiv.org\/abs\/1602.02410"},{"key":"53_CR8","doi-asserted-by":"crossref","unstructured":"Mikolov, T., Karafi\u00e1t, M., Burget, L., \u010cernock\u1ef3, J., Khudanpur, S.: Recurrent neural network based language model. In: Eleventh Annual Conference of the International Speech Communication Association (2010)","DOI":"10.1109\/ICASSP.2011.5947611"},{"key":"53_CR9","doi-asserted-by":"publisher","unstructured":"Pecar, S.: Towards opinion summarization of customer reviews. In: Proceedings of ACL 2018, Student Research Workshop, pp. 1\u20138. ACL, Melbourne, July 2018. https:\/\/doi.org\/10.18653\/v1\/P18-3001","DOI":"10.18653\/v1\/P18-3001"},{"key":"53_CR10","doi-asserted-by":"publisher","unstructured":"Peters, M., et al.: Deep contextualized word representations. In: Proceedings of the 2018 Conference of the NAACL: Human Language Technologies, Volume 1 (Long Papers), pp. 2227\u20132237. ACL, New Orleans, June 2018. https:\/\/doi.org\/10.18653\/v1\/N18-1202","DOI":"10.18653\/v1\/N18-1202"},{"key":"53_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"435","DOI":"10.1007\/978-3-030-10801-4_34","volume-title":"SOFSEM 2019: Theory and Practice of Computer Science","author":"M Pikuliak","year":"2019","unstructured":"Pikuliak, M., Simko, M., Bielikova, M.: Towards combining multitask and multilingual learning. In: Catania, B., Kr\u00e1lovi\u010d, R., Nawrocki, J., Pighizzini, G. (eds.) SOFSEM 2019. LNCS, vol. 11376, pp. 435\u2013446. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-10801-4_34"},{"key":"53_CR12","doi-asserted-by":"publisher","unstructured":"Rush, A.M., Chopra, S., Weston, J.: A neural attention model for abstractive sentence summarization. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 379\u2013389. ACL (2015). https:\/\/doi.org\/10.18653\/v1\/D15-1044 , http:\/\/aclweb.org\/anthology\/D15-1044","DOI":"10.18653\/v1\/D15-1044"},{"key":"53_CR13","unstructured":"Salton, G., Ross, R., Kelleher, J.: Attentive language models. In: Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 441\u2013450. Asian Federation of Natural Language Processing, Taipei, November 2017. https:\/\/www.aclweb.org\/anthology\/I17-1045"},{"key":"53_CR14","doi-asserted-by":"crossref","unstructured":"Simko, J., Hanakova, M., Racsko, P., Tomlein, M., Moro, R., Bielikova, M.: Fake news reading on social media: an eye-tracking study. In: Proceedings of the 30th ACM Conference on Hypertext and Social Media, pp. 221\u2013230. ACM (2019)","DOI":"10.1145\/3342220.3343642"},{"key":"53_CR15","doi-asserted-by":"crossref","unstructured":"Sundermeyer, M., Schl\u00fcter, R., Ney, H.: LSTM neural networks for language modeling. In: 13th Annual Conference of the International Speech Communication Association (2012)","DOI":"10.21437\/Interspeech.2012-65"},{"key":"53_CR16","unstructured":"Yang, Z., Dai, Z., Salakhutdinov, R., Cohen, W.W.: Breaking the softmax bottleneck: a high-rank RNN language model. CoRR abs\/1711.03953 (2017). http:\/\/arxiv.org\/abs\/1711.03953"}],"container-title":["Lecture Notes in Computer Science","SOFSEM 2020: Theory and Practice of Computer Science"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-38919-2_53","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,11]],"date-time":"2022-10-11T14:20:18Z","timestamp":1665498018000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-38919-2_53"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030389185","9783030389192"],"references-count":16,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-38919-2_53","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"17 January 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"SOFSEM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Current Trends in Theory and Practice of Informatics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Limassol","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Cyprus","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 January 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 January 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"46","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"sofsem2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/cyprusconferences.org\/sofsem2020\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"125","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"17","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"32% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.9","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.8","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}