{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T14:10:11Z","timestamp":1755871811089,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":13,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,1,4]],"date-time":"2024-01-04T00:00:00Z","timestamp":1704326400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,1,4]]},"DOI":"10.1145\/3632410.3632456","type":"proceedings-article","created":{"date-parts":[[2024,1,3]],"date-time":"2024-01-03T18:15:16Z","timestamp":1704305716000},"page":"449-453","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Comparative Analysis of Transformers for Modeling Tabular Data: A Casestudy using Industry Scale Dataset"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5047-6590","authenticated-orcid":false,"given":"Usneek","family":"Singh","sequence":"first","affiliation":[{"name":"BITS Pilani, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4261-2860","authenticated-orcid":false,"given":"Piyush","family":"Arora","sequence":"additional","affiliation":[{"name":"American Express AI Labs, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9501-8323","authenticated-orcid":false,"given":"Shamika","family":"Ganesan","sequence":"additional","affiliation":[{"name":"American Express AI Labs, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-5254-4936","authenticated-orcid":false,"given":"Mohit","family":"Kumar","sequence":"additional","affiliation":[{"name":"American Express AI Labs, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9100-4565","authenticated-orcid":false,"given":"Siddhant","family":"Kulkarni","sequence":"additional","affiliation":[{"name":"American Express AI Labs, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-3476-4434","authenticated-orcid":false,"given":"Salil Rajeev","family":"Joshi","sequence":"additional","affiliation":[{"name":"American Express AI Labs, India"}]}],"member":"320","published-online":{"date-parts":[[2024,1,4]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","unstructured":"Tianqi Chen and Carlos Guestrin. 2016. XGBoost: A Scalable Tree Boosting System. ACM. 785\u2013794 pages. https:\/\/doi.org\/10.1145\/2939672.2939785","DOI":"10.1145\/2939672.2939785"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Ranak\u00a0Roy Chowdhury Xiyuan Zhang Jingbo Shang Rajesh\u00a0K Gupta and Dezhi Hong. 2022. TARNet: Task-Aware Reconstruction for Time-Series Transformer. (2022) 14\u201318.","DOI":"10.1145\/3534678.3539329"},{"key":"e_1_3_2_1_3_1","volume-title":"A survey of quantization methods for efficient neural network inference. ArXiv preprint abs\/2103.13630","author":"Gholami Amir","year":"2021","unstructured":"Amir Gholami, Sehoon Kim, Zhen Dong, Zhewei Yao, Michael\u00a0W Mahoney, and Kurt Keutzer. 2021. A survey of quantization methods for efficient neural network inference. ArXiv preprint abs\/2103.13630 (2021). https:\/\/arxiv.org\/abs\/2103.13630"},{"key":"e_1_3_2_1_4_1","volume-title":"LUNA: Language Understanding with Number Augmentations on Transformers via Number Plugins and Pre-training. ArXiv preprint abs\/2212.02691","author":"Han Hongwei","year":"2022","unstructured":"Hongwei Han, Jialiang Xu, Mengyu Zhou, Yijia Shao, Shi Han, and Dongmei Zhang. 2022. LUNA: Language Understanding with Number Augmentations on Transformers via Number Plugins and Pre-training. ArXiv preprint abs\/2212.02691 (2022). https:\/\/arxiv.org\/abs\/2212.02691"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3358132"},{"key":"e_1_3_2_1_6_1","volume-title":"Tabtransformer: Tabular data modeling using contextual embeddings. ArXiv preprint abs\/2012.06678","author":"Huang Xin","year":"2020","unstructured":"Xin Huang, Ashish Khetan, Milan Cvitkovic, and Zohar Karnin. 2020. Tabtransformer: Tabular data modeling using contextual embeddings. ArXiv preprint abs\/2012.06678 (2020). https:\/\/arxiv.org\/abs\/2012.06678"},{"key":"e_1_3_2_1_7_1","volume-title":"Tune: A research platform for distributed model selection and training. ArXiv preprint abs\/1807.05118","author":"Liaw Richard","year":"2018","unstructured":"Richard Liaw, Eric Liang, Robert Nishihara, Philipp Moritz, Joseph\u00a0E Gonzalez, and Ion Stoica. 2018. Tune: A research platform for distributed model selection and training. ArXiv preprint abs\/1807.05118 (2018). https:\/\/arxiv.org\/abs\/1807.05118"},{"key":"e_1_3_2_1_8_1","volume-title":"Gated transformer networks for multivariate time series classification. ArXiv preprint abs\/2103.14438","author":"Liu Minghao","year":"2021","unstructured":"Minghao Liu, Shengqi Ren, Siyuan Ma, Jiahui Jiao, Yizhou Chen, Zhiguang Wang, and Wei Song. 2021. Gated transformer networks for multivariate time series classification. ArXiv preprint abs\/2103.14438 (2021). https:\/\/arxiv.org\/abs\/2103.14438"},{"key":"e_1_3_2_1_9_1","volume-title":"One Transformer for All Time Series: Representing and Training with Time-Dependent Heterogeneous Tabular Data. ArXiv preprint abs\/2302.06375","author":"Luetto Simone","year":"2023","unstructured":"Simone Luetto, Fabrizio Garuti, Enver Sangineto, Lorenzo Forni, and Rita Cucchiara. 2023. One Transformer for All Time Series: Representing and Training with Time-Dependent Heterogeneous Tabular Data. ArXiv preprint abs\/2302.06375 (2023). https:\/\/arxiv.org\/abs\/2302.06375"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP39728.2021.9414142"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCIDS.2019.8862112"},{"key":"e_1_3_2_1_12_1","volume-title":"Saint: Improved neural networks for tabular data via row attention and contrastive pre-training. ArXiv preprint abs\/2106.01342","author":"Somepalli Gowthami","year":"2021","unstructured":"Gowthami Somepalli, Micah Goldblum, Avi Schwarzschild, C\u00a0Bayan Bruss, and Tom Goldstein. 2021. Saint: Improved neural networks for tabular data via row attention and contrastive pre-training. ArXiv preprint abs\/2106.01342 (2021). https:\/\/arxiv.org\/abs\/2106.01342"},{"key":"e_1_3_2_1_13_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Zhang Yunhao","year":"2023","unstructured":"Yunhao Zhang and Junchi Yan. 2023. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The Eleventh International Conference on Learning Representations."}],"event":{"name":"CODS-COMAD 2024: 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD)","acronym":"CODS-COMAD 2024","location":"Bangalore India"},"container-title":["Proceedings of the 7th Joint International Conference on Data Science &amp; Management of Data (11th ACM IKDD CODS and 29th COMAD)"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3632456","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3632410.3632456","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T13:36:33Z","timestamp":1755869793000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3632456"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,4]]},"references-count":13,"alternative-id":["10.1145\/3632410.3632456","10.1145\/3632410"],"URL":"https:\/\/doi.org\/10.1145\/3632410.3632456","relation":{},"subject":[],"published":{"date-parts":[[2024,1,4]]},"assertion":[{"value":"2024-01-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}