{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,30]],"date-time":"2025-07-30T13:39:15Z","timestamp":1753882755090,"version":"3.41.2"},"reference-count":0,"publisher":"World Scientific Pub Co Pte Ltd","issue":"01","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Int. J. Artif. Intell. Tools"],"published-print":{"date-parts":[[2023,2]]},"abstract":"<jats:p> Many problems in NLP such as language translation and sentiment analysis have shown a lot of improvement in recent years. As simpler language problems are solved or better understood, the focus shifts to more complex problems such as semantic analysis and understanding. Unfortunately, a lot of studies in the literature suffer from a too much specificity problem. The algorithms and datasets are too domain specific. In this study, we analyze and elaborate on this notion of generality. Instead of selecting a highly specialized data set for semantic analysis, we take a generic and possibly dry data set, and we study how a plain vanilla Transformer performs in learning higher level semantic patterns beyond what was obvious or expected. We tune our Transformer model on a classic language task to ensure correct performance. Once tuned, the goal is to select sentences with specific key words and study whether higher level semantic patterns may have been learned by our model. We believe that we obtained promising results. The average BLEU score for sentences less than 25 words is equal to 39.79. Our initial qualitative analysis of possible semantic content of interest shows a 17 percent rate in finding interesting semantic patterns. We provide discussion of data driven results of unexpectedness as a measure of semantic learning. <\/jats:p>","DOI":"10.1142\/s0218213023500070","type":"journal-article","created":{"date-parts":[[2022,10,5]],"date-time":"2022-10-05T04:30:36Z","timestamp":1664944236000},"source":"Crossref","is-referenced-by-count":0,"title":["Unexpectedness as a Measure of Semantic Learning When Training Transformer Models"],"prefix":"10.1142","volume":"32","author":[{"given":"Ricardo A.","family":"Calix","sequence":"first","affiliation":[{"name":"Purdue University Northwest, Hammond, IN, USA"}]},{"given":"Leili","family":"Javadpour","sequence":"additional","affiliation":[{"name":"University of the Pacific, Stockton, CA, USA"}]}],"member":"219","published-online":{"date-parts":[[2023,2,28]]},"container-title":["International Journal on Artificial Intelligence Tools"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.worldscientific.com\/doi\/pdf\/10.1142\/S0218213023500070","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,1]],"date-time":"2023-03-01T10:10:52Z","timestamp":1677665452000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.worldscientific.com\/doi\/10.1142\/S0218213023500070"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2]]},"references-count":0,"journal-issue":{"issue":"01","published-print":{"date-parts":[[2023,2]]}},"alternative-id":["10.1142\/S0218213023500070"],"URL":"https:\/\/doi.org\/10.1142\/s0218213023500070","relation":{},"ISSN":["0218-2130","1793-6349"],"issn-type":[{"type":"print","value":"0218-2130"},{"type":"electronic","value":"1793-6349"}],"subject":[],"published":{"date-parts":[[2023,2]]},"article-number":"2350007"}}