{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T02:47:17Z","timestamp":1773802037708,"version":"3.50.1"},"reference-count":0,"publisher":"Association for the Advancement of Artificial Intelligence (AAAI)","issue":"13","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["AAAI"],"abstract":"<jats:p>Visual generative models have achieved remarkable progress in synthesizing photorealistic images and videos, yet aligning their outputs with human preferences across critical dimensions remains a persistent challenge.\nThough reinforcement learning from human feedback offers promise for preference alignment, existing reward models for visual generation face limitations, including black-box scoring without interpretability and potentially resultant unexpected biases.\nWe present VisionReward, a general framework for learning human visual preferences in both image and video generation.\nSpecifically, we employ a hierarchical visual assessment framework to capture fine-grained human preferences, and leverages linear weighting to enable interpretable preference learning.\nFurthermore, we propose a multi-dimensional consistent strategy when using VisionReward as a reward model during preference optimization for visual generation.\nExperiments show that VisionReward can significantly outperform existing image and video reward models on both machine metrics and human evaluation.\nNotably, VisionReward surpasses VideoScore by 17.2% in preference prediction accuracy, and text-to-video models with VisionReward achieve a 31.6% higher pairwise win rate compared to the same models using VideoScore.<\/jats:p>","DOI":"10.1609\/aaai.v40i13.38107","type":"journal-article","created":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T00:07:38Z","timestamp":1773792458000},"page":"11269-11277","source":"Crossref","is-referenced-by-count":0,"title":["VisionReward: Fine-Grained Multi-Dimensional Human Preference Learning for Image and Video Generation"],"prefix":"10.1609","volume":"40","author":[{"given":"Jiazheng","family":"Xu","sequence":"first","affiliation":[]},{"given":"Yu","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Jiale","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Yuanming","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Jiajun","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Yuan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Wenbo","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Shen","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Qunlin","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Shurun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jiayan","family":"Teng","sequence":"additional","affiliation":[]},{"given":"Zhuoyi","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Wendi","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Xiaohan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Shiyu","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Xiaotao","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Minlie","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Yuxiao","family":"Dong","sequence":"additional","affiliation":[]}],"member":"9382","published-online":{"date-parts":[[2026,3,14]]},"container-title":["Proceedings of the AAAI Conference on Artificial Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/38107\/42069","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/38107\/42069","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T00:07:38Z","timestamp":1773792458000},"score":1,"resource":{"primary":{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/38107"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,14]]},"references-count":0,"journal-issue":{"issue":"13","published-online":{"date-parts":[[2026,3,17]]}},"URL":"https:\/\/doi.org\/10.1609\/aaai.v40i13.38107","relation":{},"ISSN":["2374-3468","2159-5399"],"issn-type":[{"value":"2374-3468","type":"electronic"},{"value":"2159-5399","type":"print"}],"subject":[],"published":{"date-parts":[[2026,3,14]]}}}