{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:41:29Z","timestamp":1755823289664,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":41,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,26]],"date-time":"2023-10-26T00:00:00Z","timestamp":1698278400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Anhui Province Key Research and Development Plan","award":["202304a05020045"],"award-info":[{"award-number":["202304a05020045"]}]},{"name":"the Fundamental Research Funds for the Central Universities","award":["WK3490000006"],"award-info":[{"award-number":["WK3490000006"]}]},{"name":"the National Natural Science Foundation of China","award":["61836008"],"award-info":[{"award-number":["61836008"]}]},{"name":"Anhui Province Natural Science Foundation","award":["2208085UD17"],"award-info":[{"award-number":["2208085UD17"]}]},{"name":"the National Natural Science Foundation of China","award":["62176246"],"award-info":[{"award-number":["62176246"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,26]]},"DOI":"10.1145\/3581783.3612021","type":"proceedings-article","created":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T07:27:30Z","timestamp":1698391650000},"page":"800-808","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Sparse Sharing Relation Network for Panoptic Driving Perception"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-0789-9317","authenticated-orcid":false,"given":"Fan","family":"Jiang","sequence":"first","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1822-3731","authenticated-orcid":false,"given":"Zilei","family":"Wang","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]}],"member":"320","published-online":{"date-parts":[[2023,10,27]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00663"},{"key":"e_1_3_2_1_2_1","volume-title":"SegNet: A deep convolutional encoder-decoder architecture for image segmentation","author":"Badrinarayanan Vijay","year":"2017","unstructured":"Vijay Badrinarayanan, Alex Kendall, and Roberto Cipolla. 2017. SegNet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine intelligence, Vol. 39, 12 (2017), 2481--2495."},{"key":"e_1_3_2_1_3_1","volume-title":"Relation Matters: Foreground-aware Graph-based Relational Reasoning for Domain Adaptive Object Detection. arXiv preprint arXiv:2206.02355","author":"Chen Chaoqi","year":"2022","unstructured":"Chaoqi Chen, Jiongcheng Li, Hong-Yu Zhou, Xiaoguang Han, Yue Huang, Xinghao Ding, and Yizhou Yu. 2022. Relation Matters: Foreground-aware Graph-based Relational Reasoning for Domain Adaptive Object Detection. arXiv preprint arXiv:2206.02355 (2022)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00052"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00667"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_21"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01655"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2022.3192473"},{"key":"e_1_3_2_1_9_1","volume-title":"Faster, Stronger for Panoptic Driving Perception. arXiv preprint arXiv:2208.11434","author":"Han Cheng","year":"2022","unstructured":"Cheng Han, Qichao Zhao, Shuyi Zhang, Yinzi Chen, Zhenlin Zhang, and Jinwei Yuan. 2022b. YOLOPv2: Better, Faster, Stronger for Panoptic Driving Perception. arXiv preprint arXiv:2208.11434 (2022)."},{"key":"e_1_3_2_1_10_1","volume-title":"Vision gnn: An image is worth graph of nodes. arXiv preprint arXiv:2206.00272","author":"Han Kai","year":"2022","unstructured":"Kai Han, Yunhe Wang, Jianyuan Guo, Yehui Tang, and Enhua Wu. 2022a. Vision gnn: An image is worth graph of nodes. arXiv preprint arXiv:2206.00272 (2022)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00378"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58520-4_1"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01264-9_45"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00522"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00897"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00197"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.298"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.433"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.12301"},{"key":"e_1_3_2_1_20_1","volume-title":"Multi-task learning as multi-objective optimization. Advances in neural information processing systems","author":"Sener Ozan","year":"2018","unstructured":"Ozan Sener and Vladlen Koltun. 2018. Multi-task learning as multi-objective optimization. Advances in neural information processing systems, Vol. 31 (2018)."},{"key":"e_1_3_2_1_21_1","volume-title":"Attentive Task Interaction Network for Multi-Task Learning. In 2022 26th International Conference on Pattern Recognition (ICPR). IEEE, 2885--2891","author":"Sinodinos Dimitrios","year":"2022","unstructured":"Dimitrios Sinodinos and Narges Armanfard. 2022. Attentive Task Interaction Network for Multi-Task Learning. In 2022 26th International Conference on Pattern Recognition (ICPR). IEEE, 2885--2891."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2023.103622"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00116"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00036"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR48806.2021.9412265"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00972"},{"key":"e_1_3_2_1_27_1","volume-title":"HybridNets: End-to-end perception network. arXiv preprint arXiv:2203.09035","author":"Vu Dat","year":"2022","unstructured":"Dat Vu, Bao Ngo, and Hung Phan. 2022. HybridNets: End-to-end perception network. arXiv preprint arXiv:2203.09035 (2022)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58523-5_42"},{"key":"e_1_3_2_1_29_1","volume-title":"YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv preprint arXiv:2207.02696","author":"Wang Chien-Yao","year":"2022","unstructured":"Chien-Yao Wang, Alexey Bochkovskiy, and Hong-Yuan Mark Liao. 2022. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv preprint arXiv:2207.02696 (2022)."},{"key":"e_1_3_2_1_30_1","volume-title":"You only learn one representation: Unified network for multiple tasks. arXiv preprint arXiv:2105.04206","author":"Wang Chien-Yao","year":"2021","unstructured":"Chien-Yao Wang, I-Hau Yeh, and Hong-Yuan Mark Liao. 2021. You only learn one representation: Unified network for multiple tasks. arXiv preprint arXiv:2105.04206 (2021)."},{"key":"e_1_3_2_1_31_1","volume-title":"Yolop: You only look once for panoptic driving perception. Machine Intelligence Research","author":"Wu Dong","year":"2022","unstructured":"Dong Wu, Man-Wen Liao, Wei-Tian Zhang, Xing-Gang Wang, Xiang Bai, Wen-Qing Cheng, and Wen-Yu Liu. 2022. Yolop: You only look once for panoptic driving perception. Machine Intelligence Research (2022), 1--13."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00952"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00658"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19812-0_18"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-63823-8_58"},{"key":"e_1_3_2_1_36_1","volume-title":"Proceedings, Part VII 16","author":"Yu Changqian","year":"2020","unstructured":"Changqian Yu, Yifan Liu, Changxin Gao, Chunhua Shen, and Nong Sang. 2020. Representative graph neural network. In Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23--28, 2020, Proceedings, Part VII 16. Springer, 379--396."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00968"},{"key":"e_1_3_2_1_38_1","volume-title":"Dual graph convolutional network for semantic segmentation. arXiv preprint arXiv:1909.06121","author":"Zhang Li","year":"2019","unstructured":"Li Zhang, Xiangtai Li, Anurag Arnab, Kuiyuan Yang, Yunhai Tong, and Philip HS Torr. 2019a. Dual graph convolutional network for semantic segmentation. arXiv preprint arXiv:1909.06121 (2019)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i4.16469"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00097"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00867"}],"event":{"name":"MM '23: The 31st ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Ottawa ON Canada","acronym":"MM '23"},"container-title":["Proceedings of the 31st ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581783.3612021","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3581783.3612021","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:03:47Z","timestamp":1755821027000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581783.3612021"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,26]]},"references-count":41,"alternative-id":["10.1145\/3581783.3612021","10.1145\/3581783"],"URL":"https:\/\/doi.org\/10.1145\/3581783.3612021","relation":{},"subject":[],"published":{"date-parts":[[2023,10,26]]},"assertion":[{"value":"2023-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}