{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T20:53:40Z","timestamp":1743022420326,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":32,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819984282"},{"type":"electronic","value":"9789819984299"}],"license":[{"start":{"date-parts":[[2023,12,24]],"date-time":"2023-12-24T00:00:00Z","timestamp":1703376000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,24]],"date-time":"2023-12-24T00:00:00Z","timestamp":1703376000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8429-9_28","type":"book-chapter","created":{"date-parts":[[2023,12,23]],"date-time":"2023-12-23T08:02:17Z","timestamp":1703318537000},"page":"345-356","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["An Effective Dynamic Reweighting Method for\u00a0Unbiased Scene Graph Generation"],"prefix":"10.1007","author":[{"given":"Lingfeng","family":"Hu","sequence":"first","affiliation":[]},{"given":"Si","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Hanzi","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,24]]},"reference":[{"key":"28_CR1","unstructured":"Abedi, A., Karshenas, H., Adibi, P.: Multi-modal reward for visual relationships-based image captioning. arXiv preprint arXiv:2303.10766 (2023)"},{"key":"28_CR2","doi-asserted-by":"crossref","unstructured":"Chen, S., Jin, Q., Wang, P., Wu, Q.: Say as you wish: fine-grained control of image caption generation with abstract scene graphs. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00998"},{"key":"28_CR3","doi-asserted-by":"crossref","unstructured":"Chen, T., Yu, W., Chen, R., Lin, L.: Knowledge-embedded routing network for scene graph generation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00632"},{"key":"28_CR4","doi-asserted-by":"publisher","unstructured":"Deng, Y., et al.: Hierarchical memory learning for fine-grained scene graph generation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV. LNCS, vol. 13687, pp. 266\u2013283. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19812-0_16","DOI":"10.1007\/978-3-031-19812-0_16"},{"key":"28_CR5","doi-asserted-by":"crossref","unstructured":"Dong, X., Gan, T., Song, X., Wu, J., Cheng, Y., Nie, L.: Stacked hybrid-attention and group collaborative learning for unbiased scene graph generation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01882"},{"key":"28_CR6","doi-asserted-by":"crossref","unstructured":"Guo, Y., Chen, J., Zhang, H., Jiang, Y.G.: Visual relations augmented cross-modal retrieval. In: ICMR (2020)","DOI":"10.1145\/3372278.3390709"},{"key":"28_CR7","doi-asserted-by":"crossref","unstructured":"Guo, Y., et al.: From general to specific: informative scene graph generation via balance adjustment. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01607"},{"key":"28_CR8","unstructured":"Hildebrandt, M., Li, H., Koner, R., Tresp, V., G\u00fcnnemann, S.: Scene graph reasoning for visual question answering. arXiv preprint arXiv:2007.01072 (2020)"},{"key":"28_CR9","doi-asserted-by":"crossref","unstructured":"Krishna, R., et al.: Visual genome: connecting language and vision using crowdsourced dense image annotations. IJCV (2017)","DOI":"10.1007\/s11263-016-0981-7"},{"key":"28_CR10","unstructured":"Lertnattee, V., Theeramunkong, T.: Analysis of inverse class frequency in centroid-based text classification. In: ISCIT (2004)"},{"key":"28_CR11","doi-asserted-by":"publisher","unstructured":"Li, M., Qi, Y.: XPNet: cross-domain prototypical network for zero-shot sketch-based image retrieval. In: Yu, S., et al. (eds.) PRCV. LNCS, vol. 13534, pp. 394\u2013410. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-18907-4_31","DOI":"10.1007\/978-3-031-18907-4_31"},{"key":"28_CR12","doi-asserted-by":"crossref","unstructured":"Li, R., Zhang, S., Wan, B., He, X.: Bipartite graph network with adaptive message passing for unbiased scene graph generation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01096"},{"key":"28_CR13","doi-asserted-by":"crossref","unstructured":"Lyu, X., et al.: Fine-grained predicates learning for scene graph generation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01886"},{"key":"28_CR14","doi-asserted-by":"crossref","unstructured":"Pennington, J., Socher, R., Manning, C.D.: GloVe: global vectors for word representation. In: EMNLP (2014)","DOI":"10.3115\/v1\/D14-1162"},{"key":"28_CR15","doi-asserted-by":"crossref","unstructured":"Schroeder, B., Tripathi, S.: Structured query-based image retrieval using scene graphs. In: CVPRW (2020)","DOI":"10.1109\/CVPRW50498.2020.00097"},{"key":"28_CR16","unstructured":"Song, J., Zeng, P., Gao, L., Shen, H.T.: From pixels to objects: cubic visual attention for visual question answering. arXiv preprint arXiv:2206.01923 (2022)"},{"key":"28_CR17","doi-asserted-by":"crossref","unstructured":"Song, X., Chen, J., Wu, Z., Jiang, Y.G.: Spatial-temporal graphs for cross-modal Text2Video retrieval. IEEE T-MM (2021)","DOI":"10.1109\/TMM.2021.3090595"},{"key":"28_CR18","unstructured":"Tang, K.: A scene graph generation codebase in Pytorch (2020). https:\/\/github.com\/KaihuaTang\/Scene-Graph-Benchmark.pytorch"},{"key":"28_CR19","doi-asserted-by":"crossref","unstructured":"Tang, K., Niu, Y., Huang, J., Shi, J., Zhang, H.: Unbiased scene graph generation from biased training. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00377"},{"key":"28_CR20","doi-asserted-by":"crossref","unstructured":"Tang, K., Zhang, H., Wu, B., Luo, W., Liu, W.: Learning to compose dynamic tree structures for visual contexts. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00678"},{"key":"28_CR21","doi-asserted-by":"crossref","unstructured":"Teney, D., Liu, L., van Den Hengel, A.: Graph-structured representations for visual question answering. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.344"},{"key":"28_CR22","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)"},{"key":"28_CR23","doi-asserted-by":"crossref","unstructured":"Wang, J., et al.: Seesaw loss for long-tailed instance segmentation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00957"},{"key":"28_CR24","unstructured":"Xu, P., Chang, X., Guo, L., Huang, P.Y., Chen, X., Hauptmann, A.G.: A survey of scene graph: generation and application. TNNLS (2020)"},{"key":"28_CR25","doi-asserted-by":"crossref","unstructured":"Yan, S., et al.: PCPL: predicate-correlation perception learning for unbiased scene graph generation. In: ACM MM (2020)","DOI":"10.1145\/3394171.3413722"},{"key":"28_CR26","doi-asserted-by":"crossref","unstructured":"Yang, X., et al.: Transforming visual scene graphs to image captions. arXiv preprint arXiv:2305.02177 (2023)","DOI":"10.18653\/v1\/2023.acl-long.694"},{"key":"28_CR27","doi-asserted-by":"crossref","unstructured":"Yang, X., Tang, K., Zhang, H., Cai, J.: Auto-encoding scene graphs for image captioning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01094"},{"key":"28_CR28","doi-asserted-by":"crossref","unstructured":"Yu, J., Chai, Y., Wang, Y., Hu, Y., Wu, Q.: CogTree: cognition tree loss for unbiased scene graph generation. arXiv preprint arXiv:2009.07526 (2020)","DOI":"10.24963\/ijcai.2021\/176"},{"key":"28_CR29","doi-asserted-by":"crossref","unstructured":"Zellers, R., Yatskar, M., Thomson, S., Choi, Y.: Neural motifs: scene graph parsing with global context. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00611"},{"key":"28_CR30","doi-asserted-by":"publisher","unstructured":"Zhang, A., et al.: Fine-grained scene graph generation with data transfer. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds) ECCV. LNCS, vol. 13687, pp. 409\u2013424. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19812-0_24","DOI":"10.1007\/978-3-031-19812-0_24"},{"key":"28_CR31","unstructured":"Zhang, C., Chao, W.L., Xuan, D.: An empirical study on leveraging scene graphs for visual question answering. arXiv preprint arXiv:1907.12133 (2019)"},{"key":"28_CR32","doi-asserted-by":"crossref","unstructured":"Zhou, X., Li, S., Chen, H., Zhu, A.: Disentangled OCR: a more granular information for \u201ctext\u201d-to-image retrieval. In: PRCV (2022)","DOI":"10.1007\/978-3-031-18907-4_40"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8429-9_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,23]],"date-time":"2023-12-23T08:20:21Z","timestamp":1703319621000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8429-9_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,24]]},"ISBN":["9789819984282","9789819984299"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8429-9_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,24]]},"assertion":[{"value":"24 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}