{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T02:00:04Z","timestamp":1772935204434,"version":"3.50.1"},"reference-count":20,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,8]]},"DOI":"10.1109\/bigdata66926.2025.11402061","type":"proceedings-article","created":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T20:57:57Z","timestamp":1772830677000},"page":"7002-7011","source":"Crossref","is-referenced-by-count":0,"title":["Quantifying and Mitigating Occupational Bias in Open-Source Large Language Models"],"prefix":"10.1109","author":[{"given":"Animesh","family":"Raj","sequence":"first","affiliation":[{"name":"SRMIST,Data Science and Business Systems,Chennai,India"}]},{"given":"Zayed","family":"Haque","sequence":"additional","affiliation":[{"name":"SRMIST,Data Science and Business Systems,Chennai,India"}]},{"given":"K Shantha","family":"Kumari","sequence":"additional","affiliation":[{"name":"SRMIST,Data Science and Business Systems,Chennai,India"}]},{"given":"P.","family":"Kanmani","sequence":"additional","affiliation":[{"name":"SRMIST,Data Science and Business Systems,Chennai,India"}]},{"given":"Raghava Rao","family":"Mukkamala","sequence":"additional","affiliation":[{"name":"Copenhagen Business School,Department of Digitalization,Frederiksberg,Denmark"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1339"},{"key":"ref3","article-title":"Holistic evaluation of language models","author":"Liang","year":"2022","journal-title":"arXiv preprint"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.656"},{"issue":"140","key":"ref5","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of Machine Learning Research"},{"key":"ref6","article-title":"Longformer: The longdocument transformer","author":"Beltagy","year":"2020","journal-title":"arXiv preprint"},{"key":"ref7","article-title":"Towards measuring societal biases in language models","author":"Liang","year":"2021","journal-title":"arXiv preprint"},{"key":"ref8","article-title":"Stereoset: Measuring stereotypical bias in pretrained language models","author":"Nadeem","year":"2020","journal-title":"arXiv preprint"},{"key":"ref9","volume-title":"User acceptance of gender stereotypes in automated career recommendations","author":"Wang","year":"2021"},{"key":"ref10","doi-asserted-by":"crossref","DOI":"10.1609\/aaai.v36i11.21468","article-title":"Investigations of performance and bias in human-ai teamwork in hiring","volume-title":"AAAI Conference on Artificial Intelligence","author":"Peng","year":"2022"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3457607"},{"key":"ref12","volume-title":"Prompting techniques for reducing social bias in llms through system 1 and system 2 cognitive processes","author":"Kamruzzaman","year":"2024"},{"key":"ref13","volume-title":"Detect and perturb: Neutral rewriting of biased and sensitive text via gradient-based decoding","author":"He","year":"2021"},{"key":"ref14","volume-title":"Nevermind: Instruction override and moderation in large language models","author":"Kim","year":"2024"},{"key":"ref15","volume-title":"End-to-end bias mitigation: Removing gender bias in deep learning","author":"Feldman","year":"2021"},{"key":"ref16","article-title":"Man is to computer programmer as woman is to homemaker? debiasing word embeddings","author":"Bolukbasi","year":"2016","journal-title":"NeurIPS"},{"key":"ref17","article-title":"Gender shades: Intersectional accuracy disparities in commercial gender classification","volume-title":"Conference on Fairness, Accountability, and Transparency","author":"Buolamwini","year":"2018"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445924"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1016\/S0149-7634(99)00036-6"},{"key":"ref20","doi-asserted-by":"crossref","DOI":"10.1609\/aies.v7i1.31748","volume-title":"Gender, race, and intersectional bias in resume screening via language model retrieval","author":"Wilson","year":"2024"}],"event":{"name":"2025 IEEE International Conference on Big Data (BigData)","location":"Macau, China","start":{"date-parts":[[2025,12,8]]},"end":{"date-parts":[[2025,12,11]]}},"container-title":["2025 IEEE International Conference on Big Data (BigData)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11400704\/11400712\/11402061.pdf?arnumber=11402061","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T07:25:53Z","timestamp":1772868353000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11402061\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,8]]},"references-count":20,"URL":"https:\/\/doi.org\/10.1109\/bigdata66926.2025.11402061","relation":{},"subject":[],"published":{"date-parts":[[2025,12,8]]}}}