{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:16:01Z","timestamp":1750220161329,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":61,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,12,13]],"date-time":"2022-12-13T00:00:00Z","timestamp":1670889600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,12,13]]},"DOI":"10.1145\/3551626.3564946","type":"proceedings-article","created":{"date-parts":[[2022,12,7]],"date-time":"2022-12-07T00:55:45Z","timestamp":1670374545000},"page":"1-7","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Multi-Scale Channel Transformer Network for Single Image Deraining"],"prefix":"10.1145","author":[{"given":"Yuto","family":"Namba","sequence":"first","affiliation":[{"name":"Yamaguchi University, Yamaguchi, Japan"}]},{"given":"Xian-Hua","family":"Han","sequence":"additional","affiliation":[{"name":"Yamaguchi University, Yamaguchi, Japan"}]}],"member":"320","published-online":{"date-parts":[[2022,12,13]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Adversarial machine learning at scale. arXiv:1611.01236","author":"Bengio S.","year":"2017","unstructured":"S. Bengio , I. J. Goodfellow , and A. Kurakin . Adversarial machine learning at scale. arXiv:1611.01236 , 2017 . S. Bengio, I. J. Goodfellow, and A. Kurakin. Adversarial machine learning at scale. arXiv:1611.01236, 2017."},{"key":"e_1_3_2_1_2_1","volume-title":"Language models are few-shot learners. arXiv:2005.14165","author":"Brown T. B.","year":"2020","unstructured":"T. B. Brown , B. Mann , N. Ryder , M. Subbiah , J. Kaplan , P. Dhariwal , A. Neelakantan , P. Shyam , G. Sastry , A. Askell , Language models are few-shot learners. arXiv:2005.14165 , 2020 . T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, et al. Language models are few-shot learners. arXiv:2005.14165, 2020."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2014.2308627"},{"key":"e_1_3_2_1_6_1","volume-title":"ICLR","author":"Dosovitskiy A.","year":"2021","unstructured":"A. Dosovitskiy , L. Beyer , A. Kolesnikov , D. Weissenborn , X. Zhai , T. Unterthiner , M. Dehghani , M. Minderer , G. Heigold , S. Gelly , An image is worth 16x16 words: Transformers for image recognition at scale . In ICLR , 2021 . A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021."},{"key":"e_1_3_2_1_7_1","volume-title":"ACMMM","author":"Fan Z.W.","year":"2018","unstructured":"Z.W. Fan , H.F. Wu , X.Y. Fu , and Y. Huang . Residual-guide feature fusion network for single image deraining . ACMMM , 2018 . Z.W. Fan, H.F. Wu, X.Y. Fu, and Y. Huang. Residual-guide feature fusion network for single image deraining. ACMMM, 2018."},{"key":"e_1_3_2_1_8_1","volume-title":"Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. arXiv:2101.03961","author":"Fedus W.","year":"2021","unstructured":"W. Fedus , B. Zoph , and N. Shazeer . Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. arXiv:2101.03961 , 2021 . W. Fedus, B. Zoph, and N. Shazeer. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. arXiv:2101.03961, 2021."},{"issue":"1794","key":"e_1_3_2_1_9_1","first-page":"1807","article-title":"Lightweight pyramid networks for image deraining","volume":"31","author":"Fu X.","year":"2019","unstructured":"X. Fu , B. Liang , Y. Huang , X. Ding , and J. Paisley . Lightweight pyramid networks for image deraining . TNNLS , 31 : 1794 -- 1807 , 2019 . X. Fu, B. Liang, Y. Huang, X. Ding, and J. Paisley. Lightweight pyramid networks for image deraining. TNNLS, 31:1794 -- 1807, 2019.","journal-title":"TNNLS"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.186"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2691802"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i2.16224"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108124"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00745"},{"key":"e_1_3_2_1_15_1","volume-title":"Multi-scale progressive fusion network for single image deraining. CVPR, page 8346--8355","author":"Jiang K.","year":"2020","unstructured":"K. Jiang , Z.y. Wang , P. Yi , C. Chen , B.j. Huang , Y.m. Luo , J.y. Ma , and J.j. Jiang . Multi-scale progressive fusion network for single image deraining. CVPR, page 8346--8355 , 2020 . K. Jiang, Z.y. Wang, P. Yi, C. Chen, B.j. Huang, Y.m. Luo, J.y. Ma, and J.j. Jiang. Multi-scale progressive fusion network for single image deraining. CVPR, page 8346--8355, 2020."},{"key":"e_1_3_2_1_16_1","volume-title":"Unsupervised single image deraining with self-supervised constraints. ICIP, page 2761--2765","author":"Jin X.","year":"2019","unstructured":"X. Jin , Z.b. Chen , J.x. Lin , Z.k. Chen , and W. Zhou . Unsupervised single image deraining with self-supervised constraints. ICIP, page 2761--2765 , 2019 . X. Jin, Z.b. Chen, J.x. Lin, Z.k. Chen, and W. Zhou. Unsupervised single image deraining with self-supervised constraints. ICIP, page 2761--2765, 2019."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2011.2179057"},{"key":"e_1_3_2_1_18_1","volume-title":"Transformers in vision: A survey. arXiv:2101.01169","author":"Khan S.","year":"2021","unstructured":"S. Khan , M. Naseer , M. Hayat , S. W. Zamir , F. S. Khan , and M. Shah . Transformers in vision: A survey. arXiv:2101.01169 , 2021 . S. Khan, M. Naseer, M. Hayat, S. W. Zamir, F. S. Khan, and M. Shah. Transformers in vision: A survey. arXiv:2101.01169, 2021."},{"key":"e_1_3_2_1_19_1","volume-title":"ICLR","author":"Kumar M.","year":"2021","unstructured":"M. Kumar , D. Weissenborn , and N. Kalchbrenner . Colorization transformer . In ICLR , 2021 . M. Kumar, D. Weissenborn, and N. Kalchbrenner. Colorization transformer. In ICLR, 2021."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01234-2_16"},{"key":"e_1_3_2_1_21_1","volume-title":"Rain streak removal using layer priors. CVPR, page 2736--2744","author":"Li Y.","year":"2016","unstructured":"Y. Li , R. T. Tan , X. Guo , J. Lu , and M. S. Brown . Rain streak removal using layer priors. CVPR, page 2736--2744 , 2016 . Y. Li, R. T. Tan, X. Guo, J. Lu, and M. S. Brown. Rain streak removal using layer priors. CVPR, page 2736--2744, 2016."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"e_1_3_2_1_23_1","volume-title":"RoBERTa: A robustly optimized bert pretraining approach. arXiv:1907.11692","author":"Liu Y.","year":"2019","unstructured":"Y. Liu , M. Ott , N. Goyal , J. Du , M. Joshi , D. Chen , O. Levy , M. Lewis , L. Zettlemoyer , and V. Stoyanov . RoBERTa: A robustly optimized bert pretraining approach. arXiv:1907.11692 , 2019 . Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, and V. Stoyanov. RoBERTa: A robustly optimized bert pretraining approach. arXiv:1907.11692, 2019."},{"key":"e_1_3_2_1_24_1","volume-title":"Swin transformer: Hierarchical vision transformer using shifted windows. arXiv:2103.14030","author":"Liu Z.","year":"2021","unstructured":"Z. Liu , Y. Lin , Y. Cao , H. Hu , Y. Wei , Z. Zhang , S. Lin , and B. Guo . Swin transformer: Hierarchical vision transformer using shifted windows. arXiv:2103.14030 , 2021 . Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo. Swin transformer: Hierarchical vision transformer using shifted windows. arXiv:2103.14030, 2021."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"e_1_3_2_1_26_1","volume-title":"Removing rain from a single image via discriminative sparse coding. ICCV, page 3397--3405","author":"Luo Y.","year":"2015","unstructured":"Y. Luo , Y. Xu , and H. Ji . Removing rain from a single image via discriminative sparse coding. ICCV, page 3397--3405 , 2015 . Y. Luo, Y. Xu, and H. Ji. Removing rain from a single image via discriminative sparse coding. ICCV, page 3397--3405, 2015."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2018.2889277"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIEA.2018.8397790"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00231"},{"key":"e_1_3_2_1_30_1","volume-title":"Attentive generative adversarial network for raindrop removal from a single image. CVPR, page 2482--2491","author":"Qian R.","year":"2018","unstructured":"R. Qian , R. T. Tan , W. Yang , J. Su , and J. Liu . Attentive generative adversarial network for raindrop removal from a single image. CVPR, page 2482--2491 , 2018 . R. Qian, R. T. Tan, W. Yang, J. Su, and J. Liu. Attentive generative adversarial network for raindrop removal from a single image. CVPR, page 2482--2491, 2018."},{"key":"e_1_3_2_1_31_1","volume-title":"OpenAI","author":"Radford A.","year":"2018","unstructured":"A. Radford , K. Narasimhan , T. Salimans , and I. Sutskever . Improving language understanding by generative pre-training. Technical report , OpenAI , 2018 . A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever. Improving language understanding by generative pre-training. Technical report, OpenAI, 2018."},{"key":"e_1_3_2_1_32_1","volume-title":"Progressive image deraining networks: a better and simpler baseline. CVPR, page 3937--394","author":"Ren D.w.","year":"2019","unstructured":"D.w. Ren , W.m. Zuo , Q.h. Hu , P.f. Zhu , and D.y. Meng . Progressive image deraining networks: a better and simpler baseline. CVPR, page 3937--394 , 2019 . D.w. Ren, W.m. Zuo, Q.h. Hu, P.f. Zhu, and D.y. Meng. Progressive image deraining networks: a better and simpler baseline. CVPR, page 3937--394, 2019."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"e_1_3_2_1_34_1","volume-title":"ICML","author":"Touvron H.","year":"2021","unstructured":"H. Touvron , M. Cord , M. Douze , F. Massa , A. Sablayrolles , and H. Jegou . Training data-efficient image transformers & distillation through attention . In ICML , 2021 . H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, and H. Jegou. Training data-efficient image transformers & distillation through attention. In ICML, 2021."},{"key":"e_1_3_2_1_35_1","volume-title":"NeurIPS","author":"Vaswani A.","year":"2017","unstructured":"A. Vaswani , N. Shazeer , N. Parmar , J. Uszkoreit , L. Jones , A. N. Gomez , \u0141. Kaiser, and I. Polosukhin . Attention is all you need . In NeurIPS , 2017 . A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, \u0141. Kaiser, and I. Polosukhin. Attention is all you need. In NeurIPS, 2017."},{"key":"e_1_3_2_1_36_1","volume-title":"Erlnet: Entangled representation learning for single image deraining. ICCV, page 5644--5652","author":"Wang G.q.","year":"2019","unstructured":"G.q. Wang , C.m. Sun , and A. Sowmya . Erlnet: Entangled representation learning for single image deraining. ICCV, page 5644--5652 , 2019 . G.q. Wang, C.m. Sun, and A. Sowmya. Erlnet: Entangled representation learning for single image deraining. ICCV, page 5644--5652, 2019."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2895793"},{"key":"e_1_3_2_1_38_1","volume-title":"Rcdnet: An interpretable rain convolutional dictionary network for single image deraining. arXiv:2107.06808","author":"Wang H.","year":"2021","unstructured":"H. Wang , Q. Xie , Q. Zhao , Y. Liang , and D.Y. Meng . Rcdnet: An interpretable rain convolutional dictionary network for single image deraining. arXiv:2107.06808 , 2021 . H.Wang, Q. Xie, Q. Zhao, Y. Liang, and D.Y. Meng. Rcdnet: An interpretable rain convolutional dictionary network for single image deraining. arXiv:2107.06808, 2021."},{"key":"e_1_3_2_1_39_1","volume-title":"Structural residual learning for single image rain removal. Knowledge-Based Systems, page 106595","author":"Wang H.","year":"2020","unstructured":"H. Wang , Y.c.Wu, Q. Xie , Q. Zhao , Y. Liang , S.j. Zhang , and D.y. Meng . Structural residual learning for single image rain removal. Knowledge-Based Systems, page 106595 , 2020 . H.Wang, Y.c.Wu, Q. Xie, Q. Zhao, Y. Liang, S.j. Zhang, and D.y. Meng. Structural residual learning for single image rain removal. Knowledge-Based Systems, page 106595, 2020."},{"key":"e_1_3_2_1_40_1","volume-title":"Spatial attentive single-image deraining with a high quality real rain dataset. CVPR, page 12270--12279","author":"Yang X.","year":"2019","unstructured":"T.y Wang, X. Yang , K. Xu , S.z. Chen , Q. Zhang , and R. WH Lau . Spatial attentive single-image deraining with a high quality real rain dataset. CVPR, page 12270--12279 , 2019 . T.y Wang, X. Yang, K. Xu, S.z. Chen, Q. Zhang, and R. WH Lau. Spatial attentive single-image deraining with a high quality real rain dataset. CVPR, page 12270--12279, 2019."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_3_2_1_43_1","volume-title":"Uformer: A general u-shaped transformer for image restoration. arXiv:2106.03106","author":"Wang Z.","year":"2021","unstructured":"Z. Wang , X. Cun , J. Bao , and J. Liu . Uformer: A general u-shaped transformer for image restoration. arXiv:2106.03106 , 2021 . Z. Wang, X. Cun, J. Bao, and J. Liu. Uformer: A general u-shaped transformer for image restoration. arXiv:2106.03106, 2021."},{"key":"e_1_3_2_1_44_1","volume-title":"Semisupervised transfer learning for image rain removal. CVPR, page 3877--3886","author":"Wei W.","year":"2019","unstructured":"W. Wei , D.y. Meng , Q Zhao , Z.b. Xu , and Y. Wu . Semisupervised transfer learning for image rain removal. CVPR, page 3877--3886 , 2019 . W. Wei, D.y. Meng, Q Zhao, Z.b. Xu, and Y. Wu. Semisupervised transfer learning for image rain removal. CVPR, page 3877--3886, 2019."},{"key":"e_1_3_2_1_45_1","volume-title":"Segformer: Simple and efficient design for semantic segmentation with transformers. arXiv:2105.15203","author":"Xie E.","year":"2021","unstructured":"E. Xie , W. Wang , Z. Yu , A. Anandkumar , J. M. Alvarez , and P. Luo . Segformer: Simple and efficient design for semantic segmentation with transformers. arXiv:2105.15203 , 2021 . E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo. Segformer: Simple and efficient design for semantic segmentation with transformers. arXiv:2105.15203, 2021."},{"key":"e_1_3_2_1_46_1","volume-title":"Mcgkt-net: Multi-level context gating knowledge transfer network for single image deraining. ACCV, page 68--83","author":"Yamamichi K.","year":"2020","unstructured":"K. Yamamichi and X.-H. Han . Mcgkt-net: Multi-level context gating knowledge transfer network for single image deraining. ACCV, page 68--83 , 2020 . K. Yamamichi and X.-H. Han. Mcgkt-net: Multi-level context gating knowledge transfer network for single image deraining. ACCV, page 68--83, 2020."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.183"},{"key":"e_1_3_2_1_49_1","volume-title":"CVPR","author":"Yang W.H.","year":"2020","unstructured":"W.H. Yang , R. T. Tan , J.A. Feng , Z.M. Guo , S.C. Yan , and J.Y. Liu . Rcdnet: a model-driven deep neural network for single image rain removal . CVPR , 2020 . W.H. Yang, R. T. Tan, J.A. Feng, Z.M. Guo, S.C. Yan, and J.Y. Liu. Rcdnet: a model-driven deep neural network for single image rain removal. CVPR, 2020."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.2993406"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICME46284.2020.9102945"},{"key":"e_1_3_2_1_52_1","volume-title":"Uncertainty guided multi-scale residual learning-using a cycle spinning cnn for single image de-raining. CVPR, page 8405--8414","author":"Yasarla R.","year":"2019","unstructured":"R. Yasarla and V. M Patel . Uncertainty guided multi-scale residual learning-using a cycle spinning cnn for single image de-raining. CVPR, page 8405--8414 , 2019 . R. Yasarla and V. M Patel. Uncertainty guided multi-scale residual learning-using a cycle spinning cnn for single image de-raining. CVPR, page 8405--8414, 2019."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00280"},{"key":"e_1_3_2_1_54_1","volume-title":"Tokens-to-token vit: Training vision transformers from scratch on imagenet. arXiv:2101.11986","author":"Yuan L.","year":"2021","unstructured":"L. Yuan , Y. Chen , T. Wang , W. Yu , Y. Shi , Z. Jiang , F. E. Tay , J. Feng , and S. Yan . Tokens-to-token vit: Training vision transformers from scratch on imagenet. arXiv:2101.11986 , 2021 . L. Yuan, Y. Chen, T. Wang, W. Yu, Y. Shi, Z. Jiang, F. E. Tay, J. Feng, and S. Yan. Tokens-to-token vit: Training vision transformers from scratch on imagenet. arXiv:2101.11986, 2021."},{"key":"e_1_3_2_1_55_1","first-page":"14821","volume-title":"ACM Multimedia Conference","author":"Zamir S.W.","year":"2018","unstructured":"S.W. Zamir , A. Arora , S. Khan , M. Hayat , F. S. Khan , M.-H. Yang , and L. Shao . Multi-stage progressive image restoration . ACM Multimedia Conference , page 14821 -- 14831 , 2018 . S.W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, M.-H. Yang, and L. Shao. Multi-stage progressive image restoration. ACM Multimedia Conference, page 14821--14831, 2018."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01458"},{"key":"e_1_3_2_1_57_1","volume-title":"Density-aware single image deraining using a multi-stream dense network. CVPR, page 695--704","author":"Zhang H.","year":"2018","unstructured":"H. Zhang and V. M. Patel . Density-aware single image deraining using a multi-stream dense network. CVPR, page 695--704 , 2018 . H. Zhang and V. M. Patel. Density-aware single image deraining using a multi-stream dense network. CVPR, page 695--704, 2018."},{"key":"e_1_3_2_1_58_1","volume-title":"TCSVT","author":"Zhang H.","year":"2019","unstructured":"H. Zhang , V. Sindagi , and V. M. Patel . Image de-raining using a conditional generative adversarial network . TCSVT , 2019 . H. Zhang, V. Sindagi, and V. M. Patel. Image de-raining using a conditional generative adversarial network. TCSVT, 2019."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"e_1_3_2_1_60_1","volume-title":"BMVC","author":"Zheng Y. P.","year":"2019","unstructured":"Y. P. Zheng , X. Yu , M. M. Liu , and S. L. Zhang . Residual multiscale based single image deraining . BMVC , 2019 . Y. P. Zheng, X. Yu, M. M. Liu, and S. L. Zhang. Residual multiscale based single image deraining. BMVC, 2019."},{"key":"e_1_3_2_1_61_1","volume-title":"Deformable DETR: Deformable transformers for end-to-end object detection. arXiv:2010.04159","author":"Zhu X.","year":"2020","unstructured":"X. Zhu , W. Su , L. Lu , B. Li , X. Wang , and J. Dai . Deformable DETR: Deformable transformers for end-to-end object detection. arXiv:2010.04159 , 2020 . X. Zhu, W. Su, L. Lu, B. Li, X. Wang, and J. Dai. Deformable DETR: Deformable transformers for end-to-end object detection. arXiv:2010.04159, 2020."}],"event":{"name":"MMAsia '22: ACM Multimedia Asia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Tokyo Japan","acronym":"MMAsia '22"},"container-title":["Proceedings of the 4th ACM International Conference on Multimedia in Asia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3551626.3564946","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3551626.3564946","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:00:25Z","timestamp":1750186825000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3551626.3564946"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,13]]},"references-count":61,"alternative-id":["10.1145\/3551626.3564946","10.1145\/3551626"],"URL":"https:\/\/doi.org\/10.1145\/3551626.3564946","relation":{},"subject":[],"published":{"date-parts":[[2022,12,13]]},"assertion":[{"value":"2022-12-13","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}