{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T00:25:09Z","timestamp":1773447909663,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":42,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,4,2]],"date-time":"2020-04-02T00:00:00Z","timestamp":1585785600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"funder":[{"name":"NIH","award":["NIH-R01-EB017205"],"award-info":[{"award-number":["NIH-R01-EB017205"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2020,4,2]]},"DOI":"10.1145\/3368555.3384455","type":"proceedings-article","created":{"date-parts":[[2020,3,20]],"date-time":"2020-03-20T20:37:37Z","timestamp":1584736657000},"page":"214-221","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":36,"title":["Deidentification of free-text medical records using pre-trained bidirectional transformers"],"prefix":"10.1145","author":[{"given":"Alistair E. W.","family":"Johnson","sequence":"first","affiliation":[{"name":"Massachusetts Institute of Technology"}]},{"given":"Lucas","family":"Bulgarelli","sequence":"additional","affiliation":[{"name":"Massachusetts Institute of Technology"}]},{"given":"Tom J.","family":"Pollard","sequence":"additional","affiliation":[{"name":"Massachusetts Institute of Technology"}]}],"member":"320","published-online":{"date-parts":[[2020,4,2]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijmedinf.2010.09.007"},{"key":"e_1_3_2_1_2_1","volume-title":"Publicly available clinical BERT embeddings. arXiv preprint arXiv:1904.03323","author":"Alsentzer Emily","year":"2019"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Iz Beltagy Kyle Lo and Arman Cohan. 2019a. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP. arXiv:1903.10676  Iz Beltagy Kyle Lo and Arman Cohan. 2019a. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP . arXiv:1903.10676","DOI":"10.18653\/v1\/D19-1371"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"crossref","unstructured":"Iz Beltagy Kyle Lo and Arman Cohan. 2019b. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP. arXiv:1903.10676  Iz Beltagy Kyle Lo and Arman Cohan. 2019b. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP . arXiv:1903.10676","DOI":"10.18653\/v1\/D19-1371"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.5555\/3327144.3327272"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-2017"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocw156"},{"key":"e_1_3_2_1_8_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT.","author":"Devlin Jacob","year":"2019"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.23889\/ijpds.v1i1.249"},{"key":"e_1_3_2_1_10_1","first-page":"8","article-title":"The Unreasonable Effectiveness of Data. Intelligent Systems","volume":"24","author":"Halevy Alon","year":"2009","journal-title":"IEEE"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1186\/s12911-020-1026-2"},{"key":"e_1_3_2_1_12_1","volume-title":"Leo Anthony Celi, and Roger G Mark","author":"Johnson Alistair EW","year":"2016"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.13026\/0757-0y85"},{"key":"e_1_3_2_1_14_1","volume-title":"Proceedings of the Eighteenth International Conference on Machine Learning (ICML '01)","author":"Lafferty John D."},{"key":"e_1_3_2_1_15_1","volume-title":"Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942","author":"Lan Zhenzhong","year":"2019"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2017.06.006"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"e_1_3_2_1_19_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2015.06.009"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2017.05.023"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1186\/1472-6947-8-32"},{"key":"e_1_3_2_1_23_1","volume-title":"Pytorch: Tensors and dynamic neural networks in python with strong gpu acceleration.","author":"Paszke Adam","year":"2017"},{"key":"e_1_3_2_1_24_1","volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam"},{"key":"e_1_3_2_1_25_1","volume-title":"Deep contextualized word representations. arXiv preprint arXiv:1802.05365","author":"Peters Matthew E","year":"2018"},{"key":"e_1_3_2_1_26_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. (2018).  Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2016.07.015"},{"key":"e_1_3_2_1_29_1","volume-title":"Auditable Hypothesis Testing. arXiv","author":"Sacha","year":"2019"},{"key":"e_1_3_2_1_30_1","volume-title":"A systematic review and ethical enquiry into public views on the use of patient data for research in the United Kingdom and the Republic of Ireland. Wellcome open research 3","author":"Stockdale Jessica","year":"2018"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.5555\/3169406.3169752"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2015.06.007"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.5555\/2869975.2870333"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2015.07.020"},{"key":"e_1_3_2_1_35_1","volume-title":"of Labor","author":"Employee Benefits Security Administration U.S. Dept.","year":"2004"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1197\/jamia.M2444"},{"key":"e_1_3_2_1_37_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems. 5998--6008.  Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems . 5998--6008."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"crossref","unstructured":"Eric Wallace Yizhong Wang Sujian Li Sameer Singh and Matt Gardner. 2019. Do NLP Models Know Numbers? Probing Numeracy in Embeddings. In Empirical Methods in Natural Language Processing.  Eric Wallace Yizhong Wang Sujian Li Sameer Singh and Matt Gardner. 2019. Do NLP Models Know Numbers? Probing Numeracy in Embeddings. In Empirical Methods in Natural Language Processing .","DOI":"10.18653\/v1\/D19-1534"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"crossref","unstructured":"Thomas Wolf Lysandre Debut Victor Sanh Julien Chaumond Clement Delangue Anthony Moi Pierric Cistac Tim Rault R\u00e9mi Louf Morgan Funtowicz and Jamie Brew. 2019. HuggingFace's Transformers: State-of-the-art Natural Language Processing. [arxiv]cs.CL\/1910.03771  Thomas Wolf Lysandre Debut Victor Sanh Julien Chaumond Clement Delangue Anthony Moi Pierric Cistac Tim Rault R\u00e9mi Louf Morgan Funtowicz and Jamie Brew. 2019. HuggingFace's Transformers: State-of-the-art Natural Language Processing. [arxiv]cs.CL\/1910.03771","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_40_1","unstructured":"Yonghui Wu Mike Schuster Zhifeng Chen Quoc V Le Mohammad Norouzi Wolfgang Macherey Maxim Krikun Yuan Cao Qin Gao Klaus Macherey etal 2016. Google's neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016).  Yonghui Wu Mike Schuster Zhifeng Chen Quoc V Le Mohammad Norouzi Wolfgang Macherey Maxim Krikun Yuan Cao Qin Gao Klaus Macherey et al. 2016. Google's neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016)."},{"key":"e_1_3_2_1_41_1","volume-title":"Le","author":"Yang Zhilin","year":"2019"},{"key":"e_1_3_2_1_42_1","volume-title":"A survey of automatic de-identification of longitudinal clinical narratives. arXiv preprint arXiv:1810.06765","author":"Yogarajan Vithya","year":"2018"}],"event":{"name":"ACM CHIL '20: ACM Conference on Health, Inference, and Learning","location":"Toronto Ontario Canada","acronym":"ACM CHIL '20","sponsor":["ACM Association for Computing Machinery"]},"container-title":["Proceedings of the ACM Conference on Health, Inference, and Learning"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3368555.3384455","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3368555.3384455","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:01:26Z","timestamp":1750197686000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3368555.3384455"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,4,2]]},"references-count":42,"alternative-id":["10.1145\/3368555.3384455","10.1145\/3368555"],"URL":"https:\/\/doi.org\/10.1145\/3368555.3384455","relation":{},"subject":[],"published":{"date-parts":[[2020,4,2]]},"assertion":[{"value":"2020-04-02","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}