{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T21:19:57Z","timestamp":1773350397468,"version":"3.50.1"},"reference-count":48,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,12,15]],"date-time":"2024-12-15T00:00:00Z","timestamp":1734220800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,15]],"date-time":"2024-12-15T00:00:00Z","timestamp":1734220800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,12,15]]},"DOI":"10.1109\/bigdata62323.2024.10825246","type":"proceedings-article","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T18:31:23Z","timestamp":1737052283000},"page":"5819-5828","source":"Crossref","is-referenced-by-count":1,"title":["Pre-trained Transformer Uncovers Meaningful Patterns in Human Mobility Data"],"prefix":"10.1109","author":[{"given":"Alameen","family":"Najjar","sequence":"first","affiliation":[{"name":"Rakuten Institute of Technology,Tokyo,Japan"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1139"},{"key":"ref3","article-title":"Roberta: A robustly optimized bert pretraining approach","author":"Liu","year":"2019"},{"key":"ref4","article-title":"Albert: A lite bert for self-supervised learning of language representations","author":"Lan","year":"2019"},{"key":"ref5","article-title":"Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter","author":"Sanh","year":"2019"},{"key":"ref6","article-title":"Electra: Pre-training text encoders as discriminators rather than generators","author":"Clark","year":"2020"},{"key":"ref7","article-title":"Deberta: Decoding-enhanced bert with disentangled attention","author":"He","year":"2020"},{"key":"ref8","article-title":"Improving language understanding by generative pre-training","author":"Radford","year":"2018"},{"issue":"8","key":"ref9","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI blog"},{"key":"ref10","article-title":"Xlnet: Generalized autoregressive pretraining for language understanding","volume":"32","author":"Yang","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref11","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref12","article-title":"Bloom: A 176b-parameter open-access multilingual language model","author":"Workshop","year":"2022"},{"key":"ref13","article-title":"Opt: Open pre-trained transformer language models","author":"Zhang","year":"2022"},{"key":"ref14","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3557915.3560972"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3557915.3565529"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1145\/3555041.3589733"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICDMW60847.2023.00112"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/3557915.3561026"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2023.3342137"},{"key":"ref21","article-title":"Large language models for spatial trajectory patterns mining","author":"Zhang","year":"2023"},{"key":"ref22","article-title":"Where would i go next? large language models as human mobility predictors","author":"Wang","year":"2023"},{"key":"ref23","article-title":"Exploring large language models for human mobility prediction under public events","author":"Liang","year":"2023"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3557915.3561043"},{"key":"ref25","article-title":"On the opportunities and challenges of foundation models for geospatial artificial intelligence","author":"Mai","year":"2023"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/3589132.3625652"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/3589132.3625611"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1201\/9781003308423-6"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/BigData55660.2022.10020360"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/2020408.2020579"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1145\/2814575"},{"key":"ref32","article-title":"Weeplaces check-in dataset","author":"Liu","year":"2022"},{"key":"ref33","article-title":"Acm sigspatial cup","year":"2017"},{"key":"ref34","article-title":"Loc2vec: Learning location embeddings with tripletloss networks","author":"Spruyt","year":"2018"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2018.8545376"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3358001"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5450"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/3404555.3404613"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3486635.3491076"},{"key":"ref40","article-title":"Global High Resolution Population Denominators Project","year":"2018"},{"key":"ref41","article-title":"Administrative divisions of japan","year":"2024"},{"key":"ref42","article-title":"Gadm, the database of global administrative areas","year":"2024"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/IGARSS47720.2021.9553499"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/ICDMW60847.2023.00113"},{"key":"ref45","article-title":"H3: Uber\u2019s hexagonal hierarchical spatial index","author":"Brodsky","year":"2018"},{"key":"ref46","article-title":"Huggingface\u2019s transformers: State-of-the-art natural language processing","author":"Wolf","year":"2019"},{"key":"ref47","article-title":"Google\u2019s neural machine translation system: Bridging the gap between human and machine translation","author":"Wu","year":"2016"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/5.18626"}],"event":{"name":"2024 IEEE International Conference on Big Data (BigData)","location":"Washington, DC, USA","start":{"date-parts":[[2024,12,15]]},"end":{"date-parts":[[2024,12,18]]}},"container-title":["2024 IEEE International Conference on Big Data (BigData)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10824975\/10824942\/10825246.pdf?arnumber=10825246","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T07:54:37Z","timestamp":1737100477000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10825246\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,15]]},"references-count":48,"URL":"https:\/\/doi.org\/10.1109\/bigdata62323.2024.10825246","relation":{},"subject":[],"published":{"date-parts":[[2024,12,15]]}}}