{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T04:42:28Z","timestamp":1773204148622,"version":"3.50.1"},"reference-count":29,"publisher":"Institute of Electronics, Information and Communications Engineers (IEICE)","issue":"1","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEICE Trans. Inf. &amp; Syst."],"published-print":{"date-parts":[[2024,1,1]]},"DOI":"10.1587\/transinf.2023mup0001","type":"journal-article","created":{"date-parts":[[2023,12,31]],"date-time":"2023-12-31T22:39:17Z","timestamp":1704062357000},"page":"13-28","source":"Crossref","is-referenced-by-count":1,"title":["CASEformer \u2014 A Transformer-Based Projection Photometric Compensation Network"],"prefix":"10.1587","volume":"E107.D","author":[{"given":"Yuqiang","family":"ZHANG","sequence":"first","affiliation":[{"name":"School of Computer Science and Technology, Changchun University of Science and Technology"}]},{"given":"Huamin","family":"YANG","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Changchun University of Science and Technology"}]},{"given":"Cheng","family":"HAN","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Changchun University of Science and Technology"}]},{"given":"Chao","family":"ZHANG","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Changchun University of Science and Technology"}]},{"given":"Chaoran","family":"ZHU","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Jilin University"}]}],"member":"532","reference":[{"key":"1","doi-asserted-by":"crossref","unstructured":"[1] M. Ashdown, T. Okabe, I. Sato, and Y. Sato, \u201cRobust content-dependent photometric projector compensation,\u201d 2006 Conference on Computer Vision and Pattern Recognition Workshop (CVPRW&apos;06), pp.6-6, IEEE, 2006. 10.1109\/cvprw.2006.172","DOI":"10.1109\/CVPRW.2006.172"},{"key":"2","doi-asserted-by":"publisher","unstructured":"[2] A. Madi and D. Ziou, \u201cColor constancy for visual compensation of projector displayed image,\u201d Displays, vol.35, no.1, pp.6-17, 2014. 10.1016\/j.displa.2013.10.003","DOI":"10.1016\/j.displa.2013.10.003"},{"key":"3","doi-asserted-by":"crossref","unstructured":"[3] B. Huang and H. Ling, \u201cEnd-to-end projector photometric compensation,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.6810-6819, 2019.","DOI":"10.1109\/CVPR.2019.00697"},{"key":"4","doi-asserted-by":"publisher","unstructured":"[4] B. Huang, T. Sun, and H. Ling, \u201cEnd-to-end full projector compensation,\u201d IEEE Trans. Pattern Anal. Mach. Intell., vol.44, no.6, pp.2953-2967, 2021. 10.1109\/tpami.2021.3050124","DOI":"10.1109\/TPAMI.2021.3050124"},{"key":"5","unstructured":"[5] S. Nayar, \u201cA projection system with radiometric compensation for screen imperfections,\u201d ICCV Workshop on Projector-Camera Systems (PROCAMS), 2003, 2003."},{"key":"6","doi-asserted-by":"publisher","unstructured":"[6] A. Grundh\u00f6fer and D. Iwai, \u201cRobust, error-tolerant photometric projector compensation,\u201d IEEE Trans. Image Process., vol.24, no.12, pp.5086-5099, 2015. 10.1109\/tip.2015.2478388","DOI":"10.1109\/TIP.2015.2478388"},{"key":"7","doi-asserted-by":"publisher","unstructured":"[7] O. Bimber, A. Emmerling, and T. Klemmer, \u201cEmbedded entertainment with smart projectors,\u201d Computer, vol.38, no.1, pp.48-55, 2005. 10.1109\/mc.2005.17","DOI":"10.1109\/MC.2005.17"},{"key":"8","doi-asserted-by":"crossref","unstructured":"[8] B. Sajadi, M. Lazarov, and A. Majumder, \u201cAdict: Accurate direct and inverse color transformation,\u201d European Conference on Computer Vision, pp.72-86, Springer, 2010. 10.1007\/978-3-642-15561-1_6","DOI":"10.1007\/978-3-642-15561-1_6"},{"key":"9","doi-asserted-by":"publisher","unstructured":"[9] W. Zou, H. Xu, and W. L\u00fc, \u201cRadiometric compensation algorithm for color reproduction of projection display on patterned surface,\u201d Chinese Optics Letters, vol.8, no.4, pp.388-391, 2010. 10.3788\/col20100804.0388","DOI":"10.3788\/COL20100804.0388"},{"key":"10","doi-asserted-by":"publisher","unstructured":"[10] D.G. Aliaga, Y.H. Yeung, A. Law, B. Sajadi, and A. Majumder, \u201cFast high-resolution appearance editing using superimposed projections,\u201d ACM Transactions on Graphics (TOG), vol.31, no.2, pp.1-13, 2012. 10.1145\/2159516.2159518","DOI":"10.1145\/2159516.2159518"},{"key":"11","doi-asserted-by":"crossref","unstructured":"[11] N. Hashimoto and K. Kosaka, \u201cPhotometric compensation for practical and complex textures,\u201d ACM SIGGRAPH 2015 Posters, p.1, 2015. 10.1145\/2787626.2787647","DOI":"10.1145\/2787626.2787647"},{"key":"12","doi-asserted-by":"crossref","unstructured":"[12] N. Hashimoto and K. Kosaka, \u201cContinuous photometric compensation for deformable objects,\u201d ACM SIGGRAPH 2017 Posters, no.72, pp.1-2, 2017. 10.1145\/3102163.3102167","DOI":"10.1145\/3102163.3102167"},{"key":"13","doi-asserted-by":"crossref","unstructured":"[13] P. Isola, J.-Y. Zhu, T. Zhou, and A.A. Efros, \u201cImage-to-image translation with conditional adversarial networks,\u201d Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp.1125-1134, 2017.","DOI":"10.1109\/CVPR.2017.632"},{"key":"14","doi-asserted-by":"crossref","unstructured":"[14] B. Huang and H. Ling, \u201cCompennet++: End-to-end full projector compensation,\u201d Proc. IEEE\/CVF International Conference on Computer Vision, pp.7165-7174, 2019.","DOI":"10.1109\/ICCV.2019.00726"},{"key":"15","unstructured":"[15] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A.N. Gomez, \u0141. Kaiser, and I. Polosukhin, \u201cAttention is all you need,\u201d Advances in Neural Information Processing Systems, vol.30, 2017."},{"key":"16","unstructured":"[16] T. Brown, B. Mann, N. Ryder, M. Subbiah, J.D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, et al., \u201cLanguage models are few-shot learners,\u201d Advances in Neural information Processing Systems, vol.33, pp.1877-1901, 2020."},{"key":"17","doi-asserted-by":"crossref","unstructured":"[17] Y. Deng, F. Tang, X. Pan, W. Dong, C. Ma, and C. Xu, \u201c<i>stytr<\/i><sup>2<\/sup>: Unbiased image style transfer with transformers,\u201d arXiv preprint arXiv:2105.14576, 2021.","DOI":"10.1109\/CVPR52688.2022.01104"},{"key":"18","doi-asserted-by":"crossref","unstructured":"[18] N. Carion, F. Massa, G. Synnaeve, N. Usunier, A. Kirillov, and S. Zagoruyko, \u201cEnd-to-end object detection with transformers,\u201d European conference on computer vision, pp.213-229, Springer, 2020. 10.1007\/978-3-030-58452-8_13","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"19","doi-asserted-by":"crossref","unstructured":"[19] R. Strudel, R. Garcia, I. Laptev, and C. Schmid, \u201cSegmenter: Transformer for semantic segmentation,\u201d Proc. IEEE\/CVF International Conference on Computer Vision, pp.7262-7272, 2021.","DOI":"10.1109\/ICCV48922.2021.00717"},{"key":"20","unstructured":"[20] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al., \u201cAn image is worth 16x16 words: Transformers for image recognition at scale,\u201d arXiv preprint arXiv:2010.11929, 2020."},{"key":"21","doi-asserted-by":"crossref","unstructured":"[21] F. Yang, H. Yang, J. Fu, H. Lu, and B. Guo, \u201cLearning texture transformer network for image super-resolution,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.5791-5800, 2020.","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"22","unstructured":"[22] K. Lee, H. Chang, L. Jiang, H. Zhang, Z. Tu, and C. Liu, \u201cVitgan: Training gans with vision transformers,\u201d arXiv preprint arXiv:2107.04589, 2021."},{"key":"23","unstructured":"[23] S. d&apos;Ascoli, H. Touvron, M.L. Leavitt, A.S. Morcos, G. Biroli, and L. Sagun, \u201cConvit: Improving vision transformers with soft convolutional inductive biases,\u201d International Conference on Machine Learning, pp.2286-2296, PMLR, 2021."},{"key":"24","doi-asserted-by":"crossref","unstructured":"[24] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, \u201cSwin transformer: Hierarchical vision transformer using shifted windows,\u201d Proc. IEEE\/CVF International Conference on Computer Vision, pp.10012-10022, 2021.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"25","doi-asserted-by":"crossref","unstructured":"[25] Q. Hou, D. Zhou, and J. Feng, \u201cCoordinate attention for efficient mobile network design,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.13713-13722, 2021.","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"26","doi-asserted-by":"crossref","unstructured":"[26] Z. Zhong, Z.Q. Lin, R. Bidart, X. Hu, I.B. Daya, Z. Li, W.-S. Zheng, J. Li, and A. Wong, \u201cSqueeze-and-attention networks for semantic segmentation,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.13065-13074, 2020.","DOI":"10.1109\/CVPR42600.2020.01308"},{"key":"27","doi-asserted-by":"crossref","unstructured":"[27] X. Zhang, X. Zhou, M. Lin, and J. Sun, \u201cShufflenet: An extremely efficient convolutional neural network for mobile devices,\u201d Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp.6848-6856, 2018.","DOI":"10.1109\/CVPR.2018.00716"},{"key":"28","doi-asserted-by":"crossref","unstructured":"[28] J. Johnson, A. Alahi, and L. Fei-Fei, \u201cPerceptual losses for real-time style transfer and super-resolution,\u201d European conference on computer vision, pp.694-711, Springer, 2016. 10.1007\/978-3-319-46475-6_43","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"29","doi-asserted-by":"crossref","unstructured":"[29] K. He, X. Zhang, S. Ren, and J. Sun, \u201cDeep residual learning for image recognition,\u201d Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp.770-778, 2016.","DOI":"10.1109\/CVPR.2016.90"}],"container-title":["IEICE Transactions on Information and Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E107.D\/1\/E107.D_2023MUP0001\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T05:14:47Z","timestamp":1730956487000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E107.D\/1\/E107.D_2023MUP0001\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,1]]},"references-count":29,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2024]]}},"URL":"https:\/\/doi.org\/10.1587\/transinf.2023mup0001","relation":{},"ISSN":["0916-8532","1745-1361"],"issn-type":[{"value":"0916-8532","type":"print"},{"value":"1745-1361","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1,1]]},"article-number":"2023MUP0001"}}