{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,28]],"date-time":"2026-04-28T14:10:47Z","timestamp":1777385447117,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":53,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,7,21]],"date-time":"2021-07-21T00:00:00Z","timestamp":1626825600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,7,21]]},"DOI":"10.1145\/3461702.3462596","type":"proceedings-article","created":{"date-parts":[[2021,7,31]],"date-time":"2021-07-31T01:21:32Z","timestamp":1627694492000},"page":"414-424","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":23,"title":["Ensuring Fairness under Prior Probability Shifts"],"prefix":"10.1145","author":[{"given":"Arpita","family":"Biswas","sequence":"first","affiliation":[{"name":"Harvard University, Cambridge, MA, USA"}]},{"given":"Suvam","family":"Mukherjee","sequence":"additional","affiliation":[{"name":"Microsoft Corporation, Redmond, WA, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,7,30]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Agency for Healthcare Research & Quality. 2016. Medical Expenditure Panel Survey. https:\/\/meps.ahrq.gov\/mepsweb\/.  Agency for Healthcare Research & Quality. 2016. Medical Expenditure Panel Survey. https:\/\/meps.ahrq.gov\/mepsweb\/."},{"key":"e_1_3_2_1_2_1","unstructured":"Julia Angwin Jeff Larson Surya Mattu and Lauren Kirchner. 2016. Machine bias: There's software used across the country to predict future criminals. And it's biased against blacks.. In ProPublica. www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing  Julia Angwin Jeff Larson Surya Mattu and Lauren Kirchner. 2016. Machine bias: There's software used across the country to predict future criminals. And it's biased against blacks.. In ProPublica. www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing"},{"key":"e_1_3_2_1_3_1","unstructured":"Maria-Florina F Balcan Travis Dick Ritesh Noothigattu and Ariel D Procaccia. 2019. Envy-free classification. In Advances in Neural Information Processing Systems. 1238--1248. http:\/\/papers.nips.cc\/paper\/8407-envy-free-classification  Maria-Florina F Balcan Travis Dick Ritesh Noothigattu and Ariel D Procaccia. 2019. Envy-free classification. In Advances in Neural Information Processing Systems. 1238--1248. http:\/\/papers.nips.cc\/paper\/8407-envy-free-classification"},{"key":"e_1_3_2_1_4_1","unstructured":"Solon Barocas Moritz Hardt and Arvind Narayanan. 2017. Fairness in machine learning. In NIPS Tutorial.  Solon Barocas Moritz Hardt and Arvind Narayanan. 2017. Fairness in machine learning. In NIPS Tutorial."},{"key":"e_1_3_2_1_5_1","first-page":"671","article-title":"Big data's disparate impact","volume":"104","author":"Barocas Solon","year":"2016","unstructured":"Solon Barocas and Andrew D Selbst . 2016 . Big data's disparate impact . Cal. L. Rev. , Vol. 104 (2016), 671 . Solon Barocas and Andrew D Selbst. 2016. Big data's disparate impact. Cal. L. Rev., Vol. 104 (2016), 671.","journal-title":"Cal. L. Rev."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2010.75"},{"key":"e_1_3_2_1_7_1","volume-title":"John Richards, Diptikalyan Saha, Prasanna Sattigeri, Moninder Singh, Kush R. Varshney, and Yunfeng Zhang.","author":"Bellamy Rachel K. E.","year":"2018","unstructured":"Rachel K. E. Bellamy , Kuntal Dey , Michael Hind , Samuel C. Hoffman , Stephanie Houde , Kalapriya Kannan , Pranay Lohia , Jacquelyn Martino , Sameep Mehta , Aleksandra Mojsilovic , Seema Nagar , Karthikeyan Natesan Ramamurthy , John Richards, Diptikalyan Saha, Prasanna Sattigeri, Moninder Singh, Kush R. Varshney, and Yunfeng Zhang. 2018 . AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias . https:\/\/arxiv.org\/abs\/1810.01943 Rachel K. E. Bellamy, Kuntal Dey, Michael Hind, Samuel C. Hoffman, Stephanie Houde, Kalapriya Kannan, Pranay Lohia, Jacquelyn Martino, Sameep Mehta, Aleksandra Mojsilovic, Seema Nagar, Karthikeyan Natesan Ramamurthy, John Richards, Diptikalyan Saha, Prasanna Sattigeri, Moninder Singh, Kush R. Varshney, and Yunfeng Zhang. 2018. AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias. https:\/\/arxiv.org\/abs\/1810.01943"},{"key":"e_1_3_2_1_8_1","volume-title":"Sociological Methods & Research","author":"Berk Richard","unstructured":"Richard Berk , Hoda Heidari , Shahin Jabbari , Michael Kearns , and Aaron Roth . 2018. Fairness in criminal justice risk assessments: The state of the art . In Sociological Methods & Research . Sage Publications Sage CA : Los Angeles, CA . Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. 2018. Fairness in criminal justice risk assessments: The state of the art. In Sociological Methods & Research. Sage Publications Sage CA: Los Angeles, CA."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.5555\/3306127.3331934"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDMW.2009.83"},{"key":"e_1_3_2_1_11_1","volume-title":"Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT* '19)","author":"Celis L. Elisa","unstructured":"L. Elisa Celis , Lingxiao Huang , Vijay Keswani , and Nisheeth K. Vishnoi . 2019. Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees . In Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT* '19) . ACM, 319--328. L. Elisa Celis, Lingxiao Huang, Vijay Keswani, and Nisheeth K. Vishnoi. 2019. Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees. In Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT* '19). ACM, 319--328."},{"key":"e_1_3_2_1_12_1","volume-title":"Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data","author":"Chouldechova Alexandra","year":"2017","unstructured":"Alexandra Chouldechova . 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data , Vol. 5 , 2 ( 2017 ), 153--163. Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, Vol. 5, 2 (2017), 153--163."},{"key":"e_1_3_2_1_13_1","unstructured":"Alexandra Chouldechova and Aaron Roth. 2018. The Frontiers of Fairness in Machine Learning. In arXiv preprint arXiv:1810.08810.  Alexandra Chouldechova and Aaron Roth. 2018. The Frontiers of Fairness in Machine Learning. In arXiv preprint arXiv:1810.08810."},{"key":"e_1_3_2_1_14_1","unstructured":"Sam Corbett-Davies and Sharad Goel. 2018. The measure and mismeasure of fairness: A critical review of fair machine learning. In arXiv preprint arXiv:1808.00023.  Sam Corbett-Davies and Sharad Goel. 2018. The measure and mismeasure of fairness: A critical review of fair machine learning. In arXiv preprint arXiv:1808.00023."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097983.3098095"},{"key":"e_1_3_2_1_16_1","volume-title":"Training Well-Generalizing Classifiers for Fairness Metrics and Other Data-Dependent Constraints. In International Conference on Machine Learning. 1397--1405","author":"Cotter Andrew","year":"2019","unstructured":"Andrew Cotter , Maya Gupta , Heinrich Jiang , Nathan Srebro , Karthik Sridharan , Serena Wang , Blake Woodworth , and Seungil You . 2019 . Training Well-Generalizing Classifiers for Fairness Metrics and Other Data-Dependent Constraints. In International Conference on Machine Learning. 1397--1405 . Andrew Cotter, Maya Gupta, Heinrich Jiang, Nathan Srebro, Karthik Sridharan, Serena Wang, Blake Woodworth, and Seungil You. 2019. Training Well-Generalizing Classifiers for Fairness Metrics and Other Data-Dependent Constraints. In International Conference on Machine Learning. 1397--1405."},{"key":"e_1_3_2_1_17_1","volume-title":"The accuracy, fairness, and limits of predicting recidivism. Science advances","author":"Dressel Julia","year":"2018","unstructured":"Julia Dressel and Hany Farid . 2018. The accuracy, fairness, and limits of predicting recidivism. Science advances , Vol. 4 , 1 ( 2018 ), eaao5580. Julia Dressel and Hany Farid. 2018. The accuracy, fairness, and limits of predicting recidivism. Science advances, Vol. 4, 1 (2018), eaao5580."},{"key":"e_1_3_2_1_18_1","volume-title":"Adam Tauman Kalai, and Max Leiserson","author":"Dwork Cynthia","year":"2017","unstructured":"Cynthia Dwork , Nicole Immorlica , Adam Tauman Kalai, and Max Leiserson . 2017 . Decoupled classifiers for fair and efficient machine learning. In arXiv preprint arXiv:1707.06613. Cynthia Dwork, Nicole Immorlica, Adam Tauman Kalai, and Max Leiserson. 2017. Decoupled classifiers for fair and efficient machine learning. In arXiv preprint arXiv:1707.06613."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2783311"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1007\/11564096_55"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/1150402.1150423"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287589"},{"key":"e_1_3_2_1_23_1","unstructured":"Gabriel Goh Andrew Cotter Maya Gupta and Michael P Friedlander. 2016. Satisfying real-world goals with dataset constraints. In Advances in Neural Information Processing Systems. 2415--2423.  Gabriel Goh Andrew Cotter Maya Gupta and Michael P Friedlander. 2016. Satisfying real-world goals with dataset constraints. In Advances in Neural Information Processing Systems. 2415--2423."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13748-016-0103-3"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2012.05.028"},{"key":"e_1_3_2_1_26_1","unstructured":"Moritz Hardt Eric Price Nati Srebro etal 2016. Equality of opportunity in supervised learning. In Advances in neural information processing systems. 3315--3323.  Moritz Hardt Eric Price Nati Srebro et al. 2016. Equality of opportunity in supervised learning. In Advances in neural information processing systems. 3315--3323."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380228"},{"key":"e_1_3_2_1_28_1","first-page":"885","article-title":"Proportional Equality","volume":"89","author":"Hunter Nan D","year":"2000","unstructured":"Nan D Hunter . 2000 . Proportional Equality : Readings of Romer. Ky. LJ , Vol. 89 (2000), 885 . Nan D Hunter. 2000. Proportional Equality: Readings of Romer. Ky. LJ, Vol. 89 (2000), 885.","journal-title":"Readings of Romer. Ky. LJ"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-011-0463-8"},{"key":"e_1_3_2_1_30_1","volume-title":"Decision Theory for Discrimination-Aware Classification","author":"Kamiran Faisal","unstructured":"Faisal Kamiran , Asim Karim , and Xiangliang Zhang . 2012. Decision Theory for Discrimination-Aware Classification . In ICDM. IEEE Computer Society , 924--929. Faisal Kamiran, Asim Karim, and Xiangliang Zhang. 2012. Decision Theory for Discrimination-Aware Classification. In ICDM. IEEE Computer Society, 924--929."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33486-3_3"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1093\/jla\/laz001"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Jon Kleinberg Sendhil Mullainathan and Manish Raghavan. 2017. Inherent trade-offs in the fair determination of risk scores. In Innovations in Theoretical Computer Science. ACM.  Jon Kleinberg Sendhil Mullainathan and Manish Raghavan. 2017. Inherent trade-offs in the fair determination of risk scores. In Innovations in Theoretical Computer Science. ACM.","DOI":"10.1145\/3219617.3219634"},{"key":"e_1_3_2_1_34_1","volume-title":"First International Workshop on Learning over Multiple Contexts (LMCE) at ECML-PKDD.","author":"Kull Meelis","year":"2014","unstructured":"Meelis Kull and Peter Flach . 2014 . Patterns of dataset shift . In First International Workshop on Learning over Multiple Contexts (LMCE) at ECML-PKDD. Meelis Kull and Peter Flach. 2014. Patterns of dataset shift. In First International Workshop on Learning over Multiple Contexts (LMCE) at ECML-PKDD."},{"key":"e_1_3_2_1_35_1","first-page":"2346","article-title":"Learning under Concept Drift","volume":"31","author":"Lu Jie","year":"2019","unstructured":"Jie Lu , Anjin Liu , Fan Dong , Feng Gu , Jo a o Gama , and Guangquan Zhang . 2019 . Learning under Concept Drift : A Review. IEEE Transactions on Knowledge and Data Engineering , Vol. 31 , 12 (2019), 2346 -- 2363 . https:\/\/doi.org\/10.1109\/TKDE.2018.2876857 Jie Lu, Anjin Liu, Fan Dong, Feng Gu, Jo a o Gama, and Guangquan Zhang. 2019. Learning under Concept Drift: A Review. IEEE Transactions on Knowledge and Data Engineering, Vol. 31, 12 (2019), 2346--2363. https:\/\/doi.org\/10.1109\/TKDE.2018.2876857","journal-title":"A Review. IEEE Transactions on Knowledge and Data Engineering"},{"key":"e_1_3_2_1_36_1","volume-title":"Advances in Neural Information Processing Systems","volume":"33","author":"Mandal Debmalya","year":"2020","unstructured":"Debmalya Mandal , Samuel Deng , Suman Jana , Jeannette Wing , and Daniel J Hsu . 2020 . Ensuring Fairness Beyond the Training Data . In Advances in Neural Information Processing Systems , Vol. 33 . Curran Associates, Inc. , 18445--18456. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/d6539d3b57159babf6a72e106beb45bd-Paper.pdf Debmalya Mandal, Samuel Deng, Suman Jana, Jeannette Wing, and Daniel J Hsu. 2020. Ensuring Fairness Beyond the Training Data. In Advances in Neural Information Processing Systems, Vol. 33. Curran Associates, Inc., 18445--18456. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/d6539d3b57159babf6a72e106beb45bd-Paper.pdf"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2011.06.019"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.5555\/3013545.3013549"},{"key":"e_1_3_2_1_39_1","unstructured":"Geoff Pleiss Manish Raghavan Felix Wu Jon Kleinberg and Kilian Q Weinberger. 2017. On fairness and calibration. In Advances in Neural Information Processing Systems. 5684--5693.  Geoff Pleiss Manish Raghavan Felix Wu Jon Kleinberg and Kilian Q Weinberger. 2017. On fairness and calibration. In Advances in Neural Information Processing Systems. 5684--5693."},{"key":"e_1_3_2_1_40_1","unstructured":"ProPublica. 2016. COMPAS Recidivism Risk Score Data & Analysis. github.com\/propublica\/compas-analysis.  ProPublica. 2016. COMPAS Recidivism Risk Score Data & Analysis. github.com\/propublica\/compas-analysis."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1017\/S0269888913000039"},{"key":"e_1_3_2_1_42_1","volume-title":"Adjusting the outputs of a classifier to new a priori probabilities: a simple procedure. Neural computation","author":"Saerens Marco","year":"2002","unstructured":"Marco Saerens , Patrice Latinne , and Christine Decaestecker . 2002. Adjusting the outputs of a classifier to new a priori probabilities: a simple procedure. Neural computation , Vol. 14 , 1 ( 2002 ), 21--41. Marco Saerens, Patrice Latinne, and Christine Decaestecker. 2002. Adjusting the outputs of a classifier to new a priori probabilities: a simple procedure. Neural computation, Vol. 14, 1 (2002), 21--41."},{"key":"e_1_3_2_1_43_1","volume-title":"Aequitas: A Bias and Fairness Audit Toolkit. In arXiv preprint arXiv:1811.05577.","author":"Saleiro Pedro","year":"2018","unstructured":"Pedro Saleiro , Benedict Kuester , Abby Stevens , Ari Anisfeld , Loren Hinkson , Jesse London , and Rayid Ghani . 2018 . Aequitas: A Bias and Fairness Audit Toolkit. In arXiv preprint arXiv:1811.05577. Pedro Saleiro, Benedict Kuester, Abby Stevens, Ari Anisfeld, Loren Hinkson, Jesse London, and Rayid Ghani. 2018. Aequitas: A Bias and Fairness Audit Toolkit. In arXiv preprint arXiv:1811.05577."},{"key":"e_1_3_2_1_44_1","volume-title":"Fairness Without Harm: Decoupled Classifiers with Preference Guarantees. In International Conference on Machine Learning. 6373--6382","author":"Ustun Berk","year":"2019","unstructured":"Berk Ustun , Yang Liu , and David Parkes . 2019 . Fairness Without Harm: Decoupled Classifiers with Preference Guarantees. In International Conference on Machine Learning. 6373--6382 . http:\/\/proceedings.mlr.press\/v97\/ustun19a\/ustun19a.pdf Berk Ustun, Yang Liu, and David Parkes. 2019. Fairness Without Harm: Decoupled Classifiers with Preference Guarantees. In International Conference on Machine Learning. 6373--6382. http:\/\/proceedings.mlr.press\/v97\/ustun19a\/ustun19a.pdf"},{"key":"e_1_3_2_1_45_1","volume-title":"Learning in the presence of concept drift and hidden contexts. Machine learning","author":"Widmer Gerhard","year":"1996","unstructured":"Gerhard Widmer and Miroslav Kubat . 1996. Learning in the presence of concept drift and hidden contexts. Machine learning , Vol. 23 , 1 ( 1996 ), 69--101. Gerhard Widmer and Miroslav Kubat. 1996. Learning in the presence of concept drift and hidden contexts. Machine learning, Vol. 23, 1 (1996), 69--101."},{"key":"e_1_3_2_1_46_1","volume-title":"Learning Non-Discriminatory Predictors. In Conference on Learning Theory. 1920--1953","author":"Woodworth Blake","year":"2017","unstructured":"Blake Woodworth , Suriya Gunasekar , Mesrob I Ohannessian , and Nathan Srebro . 2017 . Learning Non-Discriminatory Predictors. In Conference on Learning Theory. 1920--1953 . Blake Woodworth, Suriya Gunasekar, Mesrob I Ohannessian, and Nathan Srebro. 2017. Learning Non-Discriminatory Predictors. In Conference on Learning Theory. 1920--1953."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3038912.3052660"},{"key":"e_1_3_2_1_48_1","volume-title":"Manuel Gomez Rodriguez, and Krishna P Gummadi","author":"Zafar Muhammad Bilal","year":"2017","unstructured":"Muhammad Bilal Zafar , Isabel Valera , Manuel Gomez Rodriguez, and Krishna P Gummadi . 2017 b. Fairness constraints: Mechanisms for fair classification. In Artificial Intelligence and Statistics . 962--970. Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. 2017b. Fairness constraints: Mechanisms for fair classification. In Artificial Intelligence and Statistics. 962--970."},{"key":"e_1_3_2_1_49_1","unstructured":"Muhammad Bilal Zafar Isabel Valera Manuel Rodriguez Krishna Gummadi and Adrian Weller. 2017c. From Parity to Preference-based Notions of Fairness in Classification. In Advances in Neural Information Processing Systems. 229--239.  Muhammad Bilal Zafar Isabel Valera Manuel Rodriguez Krishna Gummadi and Adrian Weller. 2017c. From Parity to Preference-based Notions of Fairness in Classification. In Advances in Neural Information Processing Systems. 229--239."},{"key":"e_1_3_2_1_50_1","volume-title":"Proceedings of the 30th International Conference on Machine Learning (ICML-13)","author":"Zemel Rich","year":"2013","unstructured":"Rich Zemel , Yu Wu , Kevin Swersky , Toni Pitassi , and Cynthia Dwork . 2013 . Learning fair representations . In Proceedings of the 30th International Conference on Machine Learning (ICML-13) . 325--333. Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In Proceedings of the 30th International Conference on Machine Learning (ICML-13). 325--333."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278779"},{"key":"e_1_3_2_1_52_1","volume-title":"Achieving Non-Discrimination in Prediction. In International Joint Conference on Artificial Intelligence, IJCAI. 3097--3103","author":"Zhang Lu","year":"2018","unstructured":"Lu Zhang , Yongkai Wu , and Xintao Wu . 2018 b. Achieving Non-Discrimination in Prediction. In International Joint Conference on Artificial Intelligence, IJCAI. 3097--3103 . Lu Zhang, Yongkai Wu, and Xintao Wu. 2018b. Achieving Non-Discrimination in Prediction. In International Joint Conference on Artificial Intelligence, IJCAI. 3097--3103."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10506-016-9182-5"}],"event":{"name":"AIES '21: AAAI\/ACM Conference on AI, Ethics, and Society","location":"Virtual Event USA","acronym":"AIES '21","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence","AAAI"]},"container-title":["Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3461702.3462596","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3461702.3462596","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:17:06Z","timestamp":1750191426000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3461702.3462596"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,21]]},"references-count":53,"alternative-id":["10.1145\/3461702.3462596","10.1145\/3461702"],"URL":"https:\/\/doi.org\/10.1145\/3461702.3462596","relation":{},"subject":[],"published":{"date-parts":[[2021,7,21]]},"assertion":[{"value":"2021-07-30","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}