{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:31:46Z","timestamp":1750221106469,"version":"3.41.0"},"reference-count":63,"publisher":"Association for Computing Machinery (ACM)","issue":"3","license":[{"start":{"date-parts":[[2019,8,8]],"date-time":"2019-08-08T00:00:00Z","timestamp":1565222400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"EU H2020 project-AniAge","award":["691215"],"award-info":[{"award-number":["691215"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["61702565"],"award-info":[{"award-number":["61702565"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Multimedia Comput. Commun. Appl."],"published-print":{"date-parts":[[2019,8,31]]},"abstract":"<jats:p>\n            The collection of internet images has been growing in an astonishing speed. It is undoubted that these images contain rich visual information that can be useful in many applications, such as visual media creation and data-driven image synthesis. In this article, we focus on the methodologies for building a visual object database from a collection of internet images. Such database is built to contain a large number of high-quality visual objects that can help with various data-driven image applications. Our method is based on dense proposal generation and objectness-based re-ranking. A novel deep convolutional neural network is designed for the inference of proposal\n            <jats:italic>objectness<\/jats:italic>\n            , the probability of a proposal containing optimally located foreground object. In our work, the\n            <jats:italic>objectness<\/jats:italic>\n            is quantitatively measured in regard of\n            <jats:italic>completeness<\/jats:italic>\n            and\n            <jats:italic>fullness<\/jats:italic>\n            , reflecting two complementary features of an optimal proposal: a complete foreground and relatively small background. Our experiments indicate that object proposals re-ranked according to the output of our network generally achieve higher performance than those produced by other state-of-the-art methods. As a concrete example, a database of over 1.2 million visual objects has been built using the proposed method, and has been successfully used in various data-driven image applications.\n          <\/jats:p>","DOI":"10.1145\/3318463","type":"journal-article","created":{"date-parts":[[2019,8,8]],"date-time":"2019-08-08T12:30:31Z","timestamp":1565267431000},"page":"1-23","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Harvesting Visual Objects from Internet Images via Deep-Learning-Based Objectness Assessment"],"prefix":"10.1145","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0663-3410","authenticated-orcid":false,"given":"Kan","family":"Wu","sequence":"first","affiliation":[{"name":"The University of Hong Kong, Hong Kong S.A.R., China"}]},{"given":"Guanbin","family":"Li","sequence":"additional","affiliation":[{"name":"Sun Yat-Sen University, Guangzhou, Guangdong, China"}]},{"given":"Haofeng","family":"Li","sequence":"additional","affiliation":[{"name":"The University of Hong Kong, Hong Kong S.A.R., China"}]},{"given":"Jianjun","family":"Zhang","sequence":"additional","affiliation":[{"name":"Bournemouth University, Poole, Dorset, United Kingdom"}]},{"given":"Yizhou","family":"Yu","sequence":"additional","affiliation":[{"name":"The University of Hong Kong and Deepwise AI Lab, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2019,8,8]]},"reference":[{"key":"e_1_2_2_1_1","unstructured":"{n.d.}. https:\/\/www.instagram.com. Accessed: 2018-May-20.  {n.d.}. https:\/\/www.instagram.com. Accessed: 2018-May-20."},{"key":"e_1_2_2_2_1","unstructured":"{n.d.}. https:\/\/www.flickr.com. Accessed: 2018-May-20.  {n.d.}. https:\/\/www.flickr.com. Accessed: 2018-May-20."},{"key":"e_1_2_2_3_1","unstructured":"{n.d.}. https:\/\/www.facebook.com. Accessed: 2018-May-20.  {n.d.}. https:\/\/www.facebook.com. Accessed: 2018-May-20."},{"key":"e_1_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.28"},{"key":"e_1_2_2_5_1","doi-asserted-by":"crossref","unstructured":"Pablo Arbel\u00e1ez Jordi Pont-Tuset Jonathan T. Barron Ferran Marques and Jitendra Malik. 2014. Multiscale combinatorial grouping. In Computer Vision and Pattern Recognition.  Pablo Arbel\u00e1ez Jordi Pont-Tuset Jonathan T. Barron Ferran Marques and Jitendra Malik. 2014. Multiscale combinatorial grouping. In Computer Vision and Pattern Recognition.","DOI":"10.1109\/CVPR.2014.49"},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.97"},{"key":"e_1_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2013.18"},{"key":"e_1_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/1618452.1618470"},{"key":"e_1_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.178"},{"key":"e_1_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2070781.2024190"},{"key":"e_1_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298724"},{"key":"e_1_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46466-4_32"},{"key":"e_1_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.412"},{"key":"e_1_2_2_14_1","unstructured":"Mark Everingham Luc Van Gool Christopher K. I. Williams John Winn and Andrew Zisserman. {n.d.}. The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results. http:\/\/www.pascal-network.org\/challenges\/VOC\/voc2007\/workshop\/index.html.  Mark Everingham Luc Van Gool Christopher K. I. Williams John Winn and Andrew Zisserman. {n.d.}. The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results. http:\/\/www.pascal-network.org\/challenges\/VOC\/voc2007\/workshop\/index.html."},{"key":"e_1_2_2_15_1","unstructured":"Mark Everingham Luc Van Gool Christopher K. I. Williams John Winn and Andrew Zisserman. {n.d.}. The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. http:\/\/www.pascal-network.org\/challenges\/VOC\/voc2012\/workshop\/index.html.  Mark Everingham Luc Van Gool Christopher K. I. Williams John Winn and Andrew Zisserman. {n.d.}. The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. http:\/\/www.pascal-network.org\/challenges\/VOC\/voc2012\/workshop\/index.html."},{"key":"e_1_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.296"},{"key":"e_1_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2703148"},{"key":"e_1_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2009.5459257"},{"key":"e_1_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/1276377.1276382"},{"volume-title":"Spatial pyramid pooling in deep convolutional networks for visual recognition. CoRR abs\/1406.4729","year":"2014","author":"He Kaiming","key":"e_1_2_2_20_1"},{"volume-title":"Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167","year":"2015","author":"Ioffe Sergey","key":"e_1_2_2_21_1"},{"volume-title":"Caffe: Convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093","year":"2014","author":"Jia Yangqing","key":"e_1_2_2_22_1"},{"volume-title":"Hinton","year":"2012","author":"Krizhevsky Alex","key":"e_1_2_2_23_1"},{"key":"e_1_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.285"},{"key":"e_1_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/1276377.1276381"},{"key":"e_1_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.34"},{"key":"e_1_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2016.2602079"},{"key":"e_1_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2872595"},{"key":"e_1_2_2_29_1","first-page":"1","article-title":"Quasi-homography warps in image stitching","volume":"99","author":"Li Nan","year":"2017","journal-title":"IEEE Transactions on Multimedia PP"},{"volume-title":"Microsoft COCO: Common objects in context. CoRR abs\/1405.0312","year":"2014","author":"Lin Tsung-Yi","key":"e_1_2_2_30_1"},{"key":"e_1_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.234"},{"key":"e_1_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.5555\/2719259.2719573"},{"key":"e_1_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2703089"},{"key":"e_1_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46448-0_35"},{"volume-title":"Learning to segment object candidates. CoRR abs\/1506.06204","year":"2015","author":"Pinheiro Pedro O.","key":"e_1_2_2_35_1"},{"volume-title":"Learning to refine object segments. CoRR abs\/1603.08695","year":"2016","author":"Pinheiro Pedro O.","key":"e_1_2_2_36_1"},{"volume-title":"You only look once: Unified, real-time object detection. arXiv preprint arXiv:1506.02640","year":"2015","author":"Redmon Joseph","key":"e_1_2_2_37_1"},{"volume-title":"Faster R-CNN: Towards real-time object detection with region proposal networks. arXiv preprint arXiv:1506.01497","year":"2015","author":"Ren Shaoqing","key":"e_1_2_2_38_1"},{"key":"e_1_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/1015706.1015720"},{"key":"e_1_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"e_1_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/2508363.2508425"},{"key":"e_1_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925954"},{"volume-title":"Very deep convolutional networks for large-scale image recognition. CoRR abs\/1409.1556","year":"2014","author":"Simonyan Karen","key":"e_1_2_2_43_1"},{"key":"e_1_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/1179352.1141964"},{"key":"e_1_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.190"},{"key":"e_1_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925942"},{"key":"e_1_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-013-0620-5"},{"key":"e_1_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/2647868.2654948"},{"key":"e_1_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.180"},{"key":"e_1_2_2_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661229.2661278"},{"key":"e_1_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2011.2165053"},{"key":"e_1_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.10"},{"volume-title":"Automatic object extraction from images using deep neural networks and the level-set method. IET Image Processing (February","year":"2018","author":"Wu Kan","key":"e_1_2_2_53_1"},{"key":"e_1_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2010.5539970"},{"key":"e_1_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298678"},{"key":"e_1_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.164"},{"key":"e_1_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2011.2162399"},{"key":"e_1_2_2_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2684626"},{"key":"e_1_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2013.2268051"},{"key":"e_1_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2018.8486523"},{"key":"e_1_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299031"},{"key":"e_1_2_2_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2701641"},{"key":"e_1_2_2_63_1","doi-asserted-by":"crossref","unstructured":"Larry Zitnick and Piotr Dollar. 2014. Edge boxes: Locating object proposals from edges. In ECCV. https:\/\/www.microsoft.com\/en-us\/research\/publication\/edge-boxes-locating-object-proposals-from-edges\/.  Larry Zitnick and Piotr Dollar. 2014. Edge boxes: Locating object proposals from edges. In ECCV. https:\/\/www.microsoft.com\/en-us\/research\/publication\/edge-boxes-locating-object-proposals-from-edges\/.","DOI":"10.1007\/978-3-319-10602-1_26"}],"container-title":["ACM Transactions on Multimedia Computing, Communications, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3318463","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3318463","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T01:01:51Z","timestamp":1750208511000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3318463"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,8,8]]},"references-count":63,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2019,8,31]]}},"alternative-id":["10.1145\/3318463"],"URL":"https:\/\/doi.org\/10.1145\/3318463","relation":{},"ISSN":["1551-6857","1551-6865"],"issn-type":[{"type":"print","value":"1551-6857"},{"type":"electronic","value":"1551-6865"}],"subject":[],"published":{"date-parts":[[2019,8,8]]},"assertion":[{"value":"2018-08-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2019-03-01","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2019-08-08","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}