{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2023,5,19]],"date-time":"2023-05-19T05:02:40Z","timestamp":1684472560160},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,1,9]],"date-time":"2023-01-09T00:00:00Z","timestamp":1673222400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,9]],"date-time":"2023-01-09T00:00:00Z","timestamp":1673222400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2023,7]]},"DOI":"10.1007\/s11760-022-02438-2","type":"journal-article","created":{"date-parts":[[2023,1,9]],"date-time":"2023-01-09T16:05:47Z","timestamp":1673280347000},"page":"2231-2240","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multiple knowledge embedding for few-shot object detection"],"prefix":"10.1007","volume":"17","author":[{"given":"Xiaolin","family":"Gong","sequence":"first","affiliation":[]},{"given":"Youpeng","family":"Cai","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,9]]},"reference":[{"key":"2438_CR1","doi-asserted-by":"publisher","first-page":"9099","DOI":"10.1109\/TIP.2021.3118953","volume":"30","author":"P-Y Chen","year":"2021","unstructured":"Chen, P.-Y., Chang, M.-C., Hsieh, J.-W., Chen, Y.-S.: Parallel residual bi-fusion feature pyramid network for accurate single-shot object detection. IEEE Trans. Image Process. 30, 9099\u20139111 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"2438_CR2","doi-asserted-by":"crossref","unstructured":"Tang, J., Shu, X., Li, Z., Qi, G.-J., Wang, J.: Generalized deep transfer networks for knowledge propagation in heterogeneous domains. ACM Trans. Multimed. Comput. Commun. Appl. 12(4s) (2016)","DOI":"10.1145\/2998574"},{"key":"2438_CR3","doi-asserted-by":"crossref","unstructured":"Shu, X., Qi, G.-J., Tang, J., Wang, J.: Weakly-shared deep transfer networks for heterogeneous-domain knowledge propagation. In Proceedings of the 23rd ACM International Conference on Multimedia, MM\u201915, pp. 35\u201344 (2015)","DOI":"10.1145\/2733373.2806216"},{"key":"2438_CR4","doi-asserted-by":"publisher","first-page":"573","DOI":"10.1109\/LSP.2021.3061978","volume":"28","author":"C Xiong","year":"2021","unstructured":"Xiong, C., Li, W., Liu, Y., Wang, M.: Multi-dimensional edge features graph neural network on few-shot image classification. IEEE Signal Process. Lett. 28, 573\u2013577 (2021)","journal-title":"IEEE Signal Process. Lett."},{"issue":"6","key":"2438_CR5","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2016","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2016)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2438_CR6","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"2438_CR7","doi-asserted-by":"crossref","unstructured":"Kang, B., Liu, Z., Wang, X., Yu, F., Feng, J., Darrell, T.: Few-shot object detection via feature reweighting. In 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 8419\u20138428 (2019)","DOI":"10.1109\/ICCV.2019.00851"},{"key":"2438_CR8","doi-asserted-by":"crossref","unstructured":"Chen, H., Wang, Y., Wang, G., Qiao, Y.: LSTD: a low-shot transfer detector for object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a032 (2018)","DOI":"10.1609\/aaai.v32i1.11716"},{"key":"2438_CR9","doi-asserted-by":"crossref","unstructured":"Wang, Y.-X., Ramanan, D., Hebert, M.: Meta-learning to detect rare objects. In 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9924\u20139933 (2019)","DOI":"10.1109\/ICCV.2019.01002"},{"key":"2438_CR10","doi-asserted-by":"crossref","unstructured":"Karlinsky, L., Shtok, J., Harary, S., Schwartz, E., Aides, A., Feris, R., Giryes, R., Bronstein, A.M.: Repmet: Representative-based metric learning for classification and few-shot object detection. In 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5192\u20135201, (2019)","DOI":"10.1109\/CVPR.2019.00534"},{"key":"2438_CR11","doi-asserted-by":"crossref","unstructured":"Li, B., Yang, B., Liu, C., Liu, F., Ji, R., Ye, Q.: Beyond max-margin: Class margin equilibrium for few-shot object detection. In 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7359\u20137368 (2021)","DOI":"10.1109\/CVPR46437.2021.00728"},{"key":"2438_CR12","unstructured":"Wang, X., Huang, T., Gonzalez, J., Darrell, T., Yu, F.: Frustratingly simple few-shot object detection. In International Conference on Machine Learning, pp. 9919\u20139928. PMLR (2020)"},{"key":"2438_CR13","doi-asserted-by":"crossref","unstructured":"Sun, B., Li, B., Cai, S., Yuan, Y., Zhang, C.: FSCE: Few-shot object detection via contrastive proposal encoding. In 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7348\u20137358 (2021)","DOI":"10.1109\/CVPR46437.2021.00727"},{"key":"2438_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, W., Wang, Y.-X.: Hallucination improves few-shot object detection. In 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13003\u201313012 (2021)","DOI":"10.1109\/CVPR46437.2021.01281"},{"key":"2438_CR15","unstructured":"Li, Y., Yao, T., Pan, Y., Mei, T.: Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. p. 1 (2022)"},{"key":"2438_CR16","first-page":"1","volume":"60","author":"G Cheng","year":"2022","unstructured":"Cheng, G., Wang, J., Li, K., Xie, X., Lang, C., Yao, Y., Han, J.: Anchor-free oriented proposal generator for object detection. IEEE Trans. Geosci. Remote Sens. 60, 1\u201311 (2022)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2438_CR17","doi-asserted-by":"crossref","unstructured":"Mukilan, P., Semunigus, W.: Human and object detection using hybrid deep convolutional neural network. Signal Image Video Process. pp. 1\u201311 (2022)","DOI":"10.1007\/s11760-022-02151-0"},{"issue":"4","key":"2438_CR18","doi-asserted-by":"publisher","first-page":"594","DOI":"10.1109\/TPAMI.2006.79","volume":"28","author":"L Fei-Fei","year":"2006","unstructured":"Fei-Fei, L., Fergus, R., Perona, P.: One-shot learning of object categories. IEEE Trans. Pattern Anal. Mach. Intell. 28(4), 594\u2013611 (2006)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"2438_CR19","doi-asserted-by":"publisher","first-page":"1091","DOI":"10.1109\/TCSVT.2020.2995754","volume":"31","author":"W Jiang","year":"2020","unstructured":"Jiang, W., Huang, K., Geng, J., Deng, X.: Multi-scale metric learning for few-shot learning. IEEE Trans. Circuits Syst. Video Technol. 31(3), 1091\u20131102 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"10","key":"2438_CR20","doi-asserted-by":"publisher","first-page":"2575","DOI":"10.1109\/TMI.2021.3060551","volume":"40","author":"R Feng","year":"2021","unstructured":"Feng, R., Zheng, X., Gao, T., Chen, J., Wang, W., Chen, D.Z., Jian, W.: Interactive few-shot learning: limited supervision, better medical image segmentation. IEEE Trans. Med. Imag. 40(10), 2575\u20132588 (2021)","journal-title":"IEEE Trans. Med. Imag."},{"key":"2438_CR21","first-page":"18661","volume":"33","author":"P Khosla","year":"2020","unstructured":"Khosla, P., Teterwak, P., Wang, C., Sarna, A., Tian, Y., Isola, P., Maschinot, A., Liu, C., Krishnan, D.: Supervised contrastive learning. Adv. Neural. Inf. Process. Syst. 33, 18661\u201318673 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2438_CR22","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597\u20131607. PMLR (2020)"},{"key":"2438_CR23","doi-asserted-by":"crossref","unstructured":"Wang, Y., Wei, Y., Ma, R., Wang, L., Wang, C.: Unsupervised vehicle re-identification based on mixed sample contrastive learning. Signal Image Video Process. 1\u20139 (2022)","DOI":"10.1007\/s11760-022-02170-x"},{"key":"2438_CR24","unstructured":"Parmar, N., Vaswani, A., Uszkoreit, J., Kaiser, L., Shazeer, N., Ku, A., Tran, D.: Image transformer. In International Conference on Machine Learning, pp. 4055\u20134064. PMLR (2018)"},{"key":"2438_CR25","doi-asserted-by":"crossref","unstructured":"Zhou, M., Xueyang, F., Huang, J., Zhao, F., Liu, A., Wang, R.: Effective pan-sharpening with transformer and invertible neural network. IEEE Trans. Geosci. Remote Sens. 60, 1\u201315 (2022)","DOI":"10.1109\/TGRS.2021.3137967"},{"issue":"4","key":"2438_CR26","doi-asserted-by":"publisher","first-page":"408","DOI":"10.1109\/TVCG.2005.62","volume":"11","author":"I Viola","year":"2005","unstructured":"Viola, I., Kanitsar, A., Groller, M.E.: Importance-driven feature enhancement in volume visualization. IEEE Trans. Vis. Comput. Graphics 11(4), 408\u2013418 (2005)","journal-title":"IEEE Trans. Vis. Comput. Graphics"},{"key":"2438_CR27","doi-asserted-by":"crossref","unstructured":"Han, G., Huang, S., Ma, J., He, Y., Chang, S.-F.: Meta Faster R-CNN: towards accurate few-shot object detection with attentive feature alignment. arXiv preprint arXiv:2104.07719 (2021)","DOI":"10.1609\/aaai.v36i1.19959"},{"key":"2438_CR28","doi-asserted-by":"crossref","unstructured":"Ravichandran, A., Bhotika, R., Soatto, S.: Few-shot learning with embedded class models and shot-free meta training. In 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 331\u2013339 (2019)","DOI":"10.1109\/ICCV.2019.00042"},{"key":"2438_CR29","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1109\/LSP.2021.3057548","volume":"28","author":"J Lin","year":"2021","unstructured":"Lin, J., Cai, Q., Lin, M.: Multi-label classification of fundus images with graph convolutional network and self-supervised learning. IEEE Signal Process. Lett. 28, 454\u2013458 (2021)","journal-title":"IEEE Signal Process. Lett."},{"key":"2438_CR30","unstructured":"Balduzzi, D., Frean, M., Leary, L., Lewis, J.P., Ma, K.W.-D., McWilliams, B.: The shattered gradients problem: If resnets are the answer, then what is the question? In International Conference on Machine Learning, pp. 342\u2013350. PMLR (2017)"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-022-02438-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-022-02438-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-022-02438-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,5,18]],"date-time":"2023-05-18T04:26:53Z","timestamp":1684384013000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-022-02438-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,9]]},"references-count":30,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2023,7]]}},"alternative-id":["2438"],"URL":"https:\/\/doi.org\/10.1007\/s11760-022-02438-2","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,1,9]]},"assertion":[{"value":"8 July 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 November 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 December 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 January 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This paper is not applicable for both human or animal studies.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}