{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T00:08:55Z","timestamp":1755907735316,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":21,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,17]],"date-time":"2024-05-17T00:00:00Z","timestamp":1715904000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"JSPS Core-to-Core Program","award":["JPJSCCA20200006"],"award-info":[{"award-number":["JPJSCCA20200006"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,17]]},"DOI":"10.1145\/3673971.3673999","type":"proceedings-article","created":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T12:22:16Z","timestamp":1725884536000},"page":"256-261","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["BARToM: a Transformer-based approach for multimodal electronic health records transformation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-0075-0392","authenticated-orcid":false,"given":"Ruiming","family":"Li","sequence":"first","affiliation":[{"name":"Medical Data Mathematical Reasoning Team, RIKEN Information R&amp;D and Strategy Headquarters, RIKEN, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9955-4342","authenticated-orcid":false,"given":"Eiryo","family":"Kawakami","sequence":"additional","affiliation":[{"name":"Medical Data Mathematical Reasoning Team, RIKEN Information R&amp;D and Strategy Headquarters, RIKEN, Japan"}]}],"member":"320","published-online":{"date-parts":[[2024,9,9]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41398-023-02720-y"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-023-00947-z"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-020-0301-z"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-018-0029-1"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-023-38257-9"},{"key":"e_1_3_2_1_6_1","unstructured":"A. Vaswani \"Attention is all you need \" Advances in neural information processing systems vol. 30 2017."},{"key":"e_1_3_2_1_7_1","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume":"35","author":"Ouyang L.","year":"2022","unstructured":"L. Ouyang , \"Training language models to follow instructions with human feedback,\" Advances in Neural Information Processing Systems, vol. 35, pp. 27730-27744, 2022.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_8_1","volume-title":"Transformers for image recognition at scale,\" arXiv preprint arXiv:2010.11929","author":"Dosovitskiy A.","year":"2020","unstructured":"A. Dosovitskiy , \"An image is worth 16x16 words: Transformers for image recognition at scale,\" arXiv preprint arXiv:2010.11929, 2020."},{"key":"e_1_3_2_1_9_1","first-page":"10347","volume-title":"Training data-efficient image transformers & distillation through attention,\" in International conference on machine learning","author":"Touvron H.","year":"2021","unstructured":"H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, and H. J\u00e9gou, \"Training data-efficient image transformers & distillation through attention,\" in International conference on machine learning, 2021: PMLR, pp. 10347-10357."},{"key":"e_1_3_2_1_10_1","first-page":"213","volume-title":"End-to-end object detection with transformers,\" in European conference on computer vision","author":"Carion N.","year":"2020","unstructured":"N. Carion, F. Massa, G. Synnaeve, N. Usunier, A. Kirillov, and S. Zagoruyko, \"End-to-end object detection with transformers,\" in European conference on computer vision, 2020: Springer, pp. 213-229."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3505244"},{"key":"e_1_3_2_1_12_1","volume-title":"Audio transformers: Transformer architectures for large scale audio understanding. adieu convolutions,\" arXiv preprint arXiv:2105.00335","author":"Verma P.","year":"2021","unstructured":"P. Verma and J. Berger, \"Audio transformers: Transformer architectures for large scale audio understanding. adieu convolutions,\" arXiv preprint arXiv:2105.00335, 2021."},{"key":"e_1_3_2_1_13_1","volume-title":"Audio captioning transformer,\" arXiv preprint arXiv:2107.09817","author":"Mei X.","year":"2021","unstructured":"X. Mei, X. Liu, Q. Huang, M. D. Plumbley, and W. Wang, \"Audio captioning transformer,\" arXiv preprint arXiv:2107.09817, 2021."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i10.21315"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-023-43715-z"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-020-62922-y"},{"key":"e_1_3_2_1_17_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding,\" arXiv preprint arXiv:1810.04805","author":"Devlin J.","year":"2018","unstructured":"J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, \"Bert: Pre-training of deep bidirectional transformers for language understanding,\" arXiv preprint arXiv:1810.04805, 2018."},{"key":"e_1_3_2_1_18_1","first-page":"1","article-title":"Health system-scale language models are all-purpose prediction engines","author":"Jiang L. Y.","year":"2023","unstructured":"L. Y. Jiang , \"Health system-scale language models are all-purpose prediction engines,\" Nature, pp. 1-6, 2023.","journal-title":"Nature"},{"key":"e_1_3_2_1_19_1","volume-title":"Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension,\" arXiv preprint arXiv:1910.13461","author":"Lewis M.","year":"2019","unstructured":"M. Lewis , \"Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension,\" arXiv preprint arXiv:1910.13461, 2019."},{"key":"e_1_3_2_1_20_1","first-page":"311","volume-title":"a method for automatic evaluation of machine translation,\" in Proceedings of the 40th annual meeting of the Association for Computational Linguistics","author":"Papineni K.","year":"2002","unstructured":"K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, \"Bleu: a method for automatic evaluation of machine translation,\" in Proceedings of the 40th annual meeting of the Association for Computational Linguistics, 2002, pp. 311-318."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1002\/jgf2.422"}],"event":{"name":"ICMHI 2024: 2024 8th International Conference on Medical and Health Informatics","acronym":"ICMHI 2024","location":"Yokohama Japan"},"container-title":["Proceedings of the 2024 8th International Conference on Medical and Health Informatics"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3673971.3673999","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3673971.3673999","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T19:48:17Z","timestamp":1755892097000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3673971.3673999"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,17]]},"references-count":21,"alternative-id":["10.1145\/3673971.3673999","10.1145\/3673971"],"URL":"https:\/\/doi.org\/10.1145\/3673971.3673999","relation":{},"subject":[],"published":{"date-parts":[[2024,5,17]]},"assertion":[{"value":"2024-09-09","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}