{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T02:48:20Z","timestamp":1773802100199,"version":"3.50.1"},"reference-count":0,"publisher":"Association for the Advancement of Artificial Intelligence (AAAI)","issue":"12","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["AAAI"],"abstract":"<jats:p>Multimodal large language models (MLLMs) frequently hallucinate by over-committing to spurious visual cues. Prior remedies\u2013Visual and Instruction Contrastive Decoding (VCD, ICD)\u2013mitigate this issue, yet the mechanism remains opaque. We first empirically show that their improvements systematically coincide with redistributions of cross-modal attention. Building on this insight, we propose Attention-Steerable Contrastive Decoding (ASCD), which directly steers the attention scores during decoding. ASCD combines (i) positive steering, which amplifies automatically mined text-centric heads\u2013stable within a model and robust across domains\u2013with (ii) negative steering, which dampens on-the-fly identified critical visual tokens. The method incurs negligible runtime\/memory overhead and requires no additional training. Across five MLLM backbones and three decoding schemes, ASCD reduces hallucination on POPE, CHAIR, and MMHal-Bench by up to 38.2% while improving accuracy on standard VQA benchmarks, including MMMU, MM-VET, ScienceQA, TextVQA, and GQA. These results position attention steering as a simple, model-agnostic, and principled route to safer, more faithful multimodal generation.<\/jats:p>","DOI":"10.1609\/aaai.v40i12.38000","type":"journal-article","created":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T00:18:47Z","timestamp":1773793127000},"page":"10306-10314","source":"Crossref","is-referenced-by-count":0,"title":["ASCD: Attention-Steerable Contrastive Decoding for Reducing Hallucination in MLLM"],"prefix":"10.1609","volume":"40","author":[{"given":"Yujun","family":"Wang","sequence":"first","affiliation":[]},{"family":"Aniri","sequence":"additional","affiliation":[]},{"given":"Jinhe","family":"Bi","sequence":"additional","affiliation":[]},{"given":"Soren","family":"Pirk","sequence":"additional","affiliation":[]},{"given":"Yunpu","family":"Ma","sequence":"additional","affiliation":[]}],"member":"9382","published-online":{"date-parts":[[2026,3,14]]},"container-title":["Proceedings of the AAAI Conference on Artificial Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/38000\/41962","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/38000\/41962","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T00:18:48Z","timestamp":1773793128000},"score":1,"resource":{"primary":{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/38000"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,14]]},"references-count":0,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2026,3,17]]}},"URL":"https:\/\/doi.org\/10.1609\/aaai.v40i12.38000","relation":{},"ISSN":["2374-3468","2159-5399"],"issn-type":[{"value":"2374-3468","type":"electronic"},{"value":"2159-5399","type":"print"}],"subject":[],"published":{"date-parts":[[2026,3,14]]}}}