{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T21:30:02Z","timestamp":1772919002513,"version":"3.50.1"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2022,2,24]],"date-time":"2022-02-24T00:00:00Z","timestamp":1645660800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,2,24]],"date-time":"2022-02-24T00:00:00Z","timestamp":1645660800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001863","name":"New Energy and Industrial Technology Development Organization","doi-asserted-by":"publisher","award":["JPNP20006"],"award-info":[{"award-number":["JPNP20006"]}],"id":[{"id":"10.13039\/501100001863","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"crossref","award":["19H01134"],"award-info":[{"award-number":["19H01134"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2022,4]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Is it possible to use convolutional neural networks pre-trained without any natural images to assist natural image understanding? The paper proposes a novel concept, Formula-driven Supervised Learning (FDSL). We automatically generate image patterns and their category labels by assigning fractals, which are based on a natural law. Theoretically, the use of automatically generated images instead of natural images in the pre-training phase allows us to generate an infinitely large dataset of labeled images. The proposed framework is similar yet different from Self-Supervised Learning because the FDSL framework enables the creation of image patterns based on any mathematical formulas in addition to self-generated labels. Further, unlike pre-training with a synthetic image dataset, a dataset under the framework of FDSL is not required to define object categories, surface texture, lighting conditions, and camera viewpoint. In the experimental section, we find a better dataset configuration through an exploratory study, e.g., increase of #category\/#instance, patch rendering, image coloring, and training epoch. Although models pre-trained with the proposed Fractal DataBase (FractalDB), a database without natural images, do not necessarily outperform models pre-trained with human annotated datasets in all settings, we are able to partially surpass the accuracy of ImageNet\/Places pre-trained models. The FractalDB pre-trained CNN also outperforms other pre-trained models on auto-generated datasets based on FDSL such as Bezier curves and Perlin noise. This is reasonable since natural objects and scenes existing around us are constructed according to fractal geometry. Image representation with the proposed FractalDB captures a unique feature in the visualization of convolutional layers and attentions.<\/jats:p>","DOI":"10.1007\/s11263-021-01555-8","type":"journal-article","created":{"date-parts":[[2022,2,24]],"date-time":"2022-02-24T09:03:04Z","timestamp":1645693384000},"page":"990-1007","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":44,"title":["Pre-Training Without Natural Images"],"prefix":"10.1007","volume":"130","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8844-165X","authenticated-orcid":false,"given":"Hirokatsu","family":"Kataoka","sequence":"first","affiliation":[]},{"given":"Kazushige","family":"Okayasu","sequence":"additional","affiliation":[]},{"given":"Asato","family":"Matsumoto","sequence":"additional","affiliation":[]},{"given":"Eisuke","family":"Yamagata","sequence":"additional","affiliation":[]},{"given":"Ryosuke","family":"Yamada","sequence":"additional","affiliation":[]},{"given":"Nakamasa","family":"Inoue","sequence":"additional","affiliation":[]},{"given":"Akio","family":"Nakamura","sequence":"additional","affiliation":[]},{"given":"Yutaka","family":"Satoh","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,2,24]]},"reference":[{"key":"1555_CR1","unstructured":"Asano, Y.M., Rupprecht, C., & Vedaldi, A. (2020). A critical analysis of self-supervision, or what we can learn from a single image. In international conference on learning representation (ICLR)."},{"key":"1555_CR2","unstructured":"Asano, Y.M., Rupprecht, C., & Vedaldi, A. (2020). Self-labelling via simultaneous clustering and representation learning. In international conference on learning representation (ICLR)."},{"key":"1555_CR3","volume-title":"Fractals everywhere","author":"MF Barnsley","year":"1988","unstructured":"Barnsley, M. F. (1988). Fractals everywhere. New York: Academic Press."},{"key":"1555_CR4","doi-asserted-by":"crossref","unstructured":"Birhane, A., & Prabhu, V.U. (2021). Large image datasets: A pyrrhic win for computer vision? Winter conference on applications of computer vision (WACV).","DOI":"10.1109\/WACV48630.2021.00158"},{"key":"1555_CR5","doi-asserted-by":"crossref","unstructured":"Bottou, L. (2010). Large-Scale Machine Learning with Stochastic Gradient Descent. In 19th international conference on computational statistics (COMPSTAT), pp. 177\u2013187.","DOI":"10.1007\/978-3-7908-2604-3_16"},{"key":"1555_CR6","doi-asserted-by":"crossref","unstructured":"Caron, M., Bojanowski, P., Joulin, A., & Douze, M. (2018). Deep clustering for unsupervised learning of visual features. In European conference on computer vision (ECCV), pp. 132\u2013149.","DOI":"10.1007\/978-3-030-01264-9_9"},{"key":"1555_CR7","unstructured":"Chen, T., Kornblith, S., Norouzi, M., & Hinton, G. (2020). A simple framework for contrastive learning of visual representations. In International conference on machine learning (ICML)."},{"issue":"3","key":"1555_CR8","doi-asserted-by":"publisher","first-page":"367","DOI":"10.1016\/S0097-8493(97)00014-9","volume":"21","author":"YQ Chen","year":"1997","unstructured":"Chen, Y. Q., & Bi, G. (1997). 3-D IFS fractals as real-time graphics model. Computers and Graphics, 21(3), 367\u2013370.","journal-title":"Computers and Graphics"},{"key":"1555_CR9","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., & Fei-Fei, L. (2009). ImageNet: A large-scale hierarchical image database. In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 248\u2013255.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1555_CR10","doi-asserted-by":"crossref","unstructured":"Doersch, C., Gupta, A., & Efros, A. (2015). Unsupervised visual representation learning by context prediction. In The IEEE international conference on computer vision (ICCV), pp. 1422\u20131430.","DOI":"10.1109\/ICCV.2015.167"},{"key":"1555_CR11","unstructured":"Donahue, J., Jia, Y., Hoffman, J., Zhang, N., Tzeng, E., & Darrell, T. (2014). DeCAF: A deep convolutional activation feature for generic visual recognition. In International conference on machine learning (ICML) pp. 647\u2013655."},{"key":"1555_CR12","unstructured":"Donahue, J., & Simonyan, K. (2019). Large scale adversarial representation learning. In arXiv pre-print arXiv:1907.02544."},{"key":"1555_CR13","doi-asserted-by":"crossref","unstructured":"Dwibedi, D., Misra, I., & Hebert, M. (2017). Cut, paste and learn: Surprisingly easy synthesis for instance detection. In International conference on computer vision (ICCV).","DOI":"10.1109\/ICCV.2017.146"},{"issue":"1","key":"1555_CR14","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1007\/s11263-014-0733-5","volume":"111","author":"M Everingham","year":"2015","unstructured":"Everingham, M., Eslami, S. M. A., Van Gool, L., Williams, C. K. I., Winn, J., & Zisserman, A. (2015). The Pascal visual object classes challenge: A retrospective. International Journal of Computer Vision (IJCV), 111(1), 98\u2013136.","journal-title":"International Journal of Computer Vision (IJCV)"},{"key":"1555_CR15","volume-title":"Fractal geometry: Mathematical foundations and applications","author":"K Falconer","year":"2004","unstructured":"Falconer, K. (2004). Fractal geometry: Mathematical foundations and applications. New Jersey: John Wiley and Sons."},{"key":"1555_CR16","volume-title":"Curves and surfaces for computer aided geometric design: A practical guide","author":"G Farin","year":"1993","unstructured":"Farin, G. (1993). Curves and surfaces for computer aided geometric design: A practical guide. Cambridge: Academic Press."},{"key":"1555_CR17","doi-asserted-by":"crossref","unstructured":"Fellbaum, C. (1998). WordNet: An electronic lexical database. BradfordBooks.","DOI":"10.7551\/mitpress\/7287.001.0001"},{"key":"1555_CR18","unstructured":"Gidaris, S., Singh, P., & Komodakis, N. (2018). Unsupervised representation learning by predicting image rotations. In International conference on learning representation (ICLR)."},{"key":"1555_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., & Girshick, R. (2020). Momentum contrast for unsupervised visual representation learning. In The IEEE international conference on computer vision and pattern recognition (CVPR).","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"1555_CR20","doi-asserted-by":"crossref","unstructured":"He, K., Girshick, R., & Doll\u00e1r, P. (2019). Rethinking ImageNet pre-training. In The IEEE international conference on computer vision (ICCV), pp. 4918\u20134927.","DOI":"10.1109\/ICCV.2019.00502"},{"key":"1555_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 770\u2013778.","DOI":"10.1109\/CVPR.2016.90"},{"key":"1555_CR22","doi-asserted-by":"crossref","unstructured":"Howard, A.G., Sandler, M., Chu, G., Chen, L.C., Chen, B., Tan M.and\u00a0Wang, W., Zhu, Y., Pang, R., Vasudevan, V., Le, Q.V., & Adam, H. (2019). Searching for MobileNetV3. In The IEEE international conference on computer vision (ICCV), pp. 1314\u20131324.","DOI":"10.1109\/ICCV.2019.00140"},{"key":"1555_CR23","unstructured":"Howard, A.G., Zhu M., C.B., Kalenichenko, D., Wang, W., Weyand, T., An-dreetto, M., & Adam, H. (2017). MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv pre-print arXiv:1704.04861."},{"key":"1555_CR24","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1109\/TPAMI.2019.2913372","volume":"42","author":"J Hu","year":"2020","unstructured":"Hu, J., Shen, L., Albanie, S., Sun, G., & Wu, E. (2020). Squeeze-and-excitation networks. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 42, 2011\u20132023.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"},{"key":"1555_CR25","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Maaten, L.V.d., & Weinberger, K.Q. (2017). Densely connected convolutional networks. In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 4700\u20134708.","DOI":"10.1109\/CVPR.2017.243"},{"key":"1555_CR26","unstructured":"Huh, M., Agrawal, P., & Efros, A.A. (2016). What makes ImageNet good for transfer learning? In Advances in neural information processing systems NIPS 2016 Workshop."},{"key":"1555_CR27","unstructured":"Kay, W., Carreira, J., Simonyan, K., Zhang, B., Hillier, C., Vijayanarasimhan, S., Viola, F., Green, T., Back, T., Natsev, P., M., S., & Zisserman, A. (2017). The kinetics human action video dataset. arXiv pre-print arXiv:1705.06950."},{"key":"1555_CR28","doi-asserted-by":"crossref","unstructured":"Kornblith, S., Shlens, J., & Le, Q.V. (2019). Do better imagenet models transfer better? In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 2661\u20132671.","DOI":"10.1109\/CVPR.2019.00277"},{"key":"1555_CR29","unstructured":"Krasin, I., Duerig, T., Alldrin, N., Ferrari, V., Abu-El-Haija, S., Kuznetsova, A., Rom, H., Uijlings, J., Popov, S., Kamali, S., Malloci, M., Pont-Tuset, J., Veit, A., Belongie, S., Gomes, V., Gupta, A., Sun, C., Chechik, G., Cai, D., Feng, Z., Narayanan, D., & Murphy, K. (2017). OpenImages: A public dataset for large-scale multi-label and multi-class image classification."},{"key":"1555_CR30","unstructured":"Krizhevsky, A. (2009). Learning multiple layers of features from tiny images."},{"key":"1555_CR31","unstructured":"Krizhevsky, A., Sutskever, I., & Hinton, G.E. (2012). ImageNet classification with deep convolutional neural networks. In F.\u00a0Pereira, C.J.C. Burges, L.\u00a0Bottou, K.Q. Weinberger (eds.) Advances in neural information processing systems (NIPS) 25, pp. 1097\u20131105."},{"issue":"6266","key":"1555_CR32","doi-asserted-by":"publisher","first-page":"1332","DOI":"10.1126\/science.aab3050","volume":"350","author":"BM Lake","year":"2015","unstructured":"Lake, B. M., & Tenenbaum, J. B. (2015). Human-level concept learning through probabilistic program induction. Science, 350(6266), 1332\u20131338.","journal-title":"Science"},{"key":"1555_CR33","unstructured":"Landini, G., Murry, P.I., & Misson, G.P. (1995). Local connected fractal dimensions and lacunarity analyses of 60 degree fluorescein angiograms. In: Investigative Ophthalmology and Visual Science, pp. 2749\u20132755."},{"key":"1555_CR34","unstructured":"Larsson, G., Maire, M., & Shakhnarovich, G. (2017). FractalNet: Ultra-deep neural networks without residuals. In International conference on learning representation (ICLR)."},{"key":"1555_CR35","doi-asserted-by":"crossref","unstructured":"Lin, T., Maire, M., Belongie, S.J., Hays, J., Perona, P., Ramanan, D., Dollar, P., & Zitnick, C.L. (2014). Microsoft COCO: Common objects in context. In European conference on computer vision (ECCV), pp. 740\u2013755.","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1555_CR36","doi-asserted-by":"crossref","unstructured":"Mahajan, D., Girshick, R., Ramanathan, V., He, K., Paluri, M., Li, Y., Bharambe, A., & Maaten, L.v.d. (2018). Exploring the limits of weakly supervised pretraining. In European conference on computer vision (ECCV), pp. 181\u2013196.","DOI":"10.1007\/978-3-030-01216-8_12"},{"key":"1555_CR37","doi-asserted-by":"crossref","unstructured":"Mandelbrot, B. (1983). The fractal geometry of nature. American Journal of Physics, 51(3)","DOI":"10.1119\/1.13295"},{"key":"1555_CR38","doi-asserted-by":"crossref","unstructured":"Monfort, M., Andonian, A., Zhou, B., Ramakrishnan, K., Adel\u00a0Bargal, S., Yan, T., Brown, L., Fan, Q., Gutfreund, D., Vondrick, C., & Oliva, A. (2019). Moments in time dataset: One million videos for event understanding. In IEEE transactions on pattern analysis and machine intelligence (TPAMI).","DOI":"10.1109\/TPAMI.2019.2901464"},{"key":"1555_CR39","doi-asserted-by":"crossref","unstructured":"Monro, D.M., & Budbridge, F. (1995). Rendering algorithms for deteministic fractals. In IEEE computer graphics and its applications, pp. 32\u201341.","DOI":"10.1109\/38.364961"},{"key":"1555_CR40","doi-asserted-by":"crossref","unstructured":"Movshovitz-Attias, Y., Kanade, T., & Sheikh, Y. (2016). How useful is photo-realistic rendering for visual learning? In European conference on computer vision (ECCV).","DOI":"10.1007\/978-3-319-49409-8_18"},{"key":"1555_CR41","doi-asserted-by":"crossref","unstructured":"Noroozi, M., & Favaro, P. (2016). Unsupervised learning of visual representations by solving jigsaw puzzles. In European conference on computer vision (ECCV).","DOI":"10.1007\/978-3-319-46466-4_5"},{"key":"1555_CR42","doi-asserted-by":"crossref","unstructured":"Noroozi, M., Pirsiavash, H., & Favaro, P. (2017). Representation learning by learning to count. In The IEEE international conference on computer vision (ICCV), pp. 5898\u20135906.","DOI":"10.1109\/ICCV.2017.628"},{"key":"1555_CR43","doi-asserted-by":"crossref","unstructured":"Noroozi, M., Vinjimoor, A., Favaro, P., & Pirsiavash, H. (2018). Boosting self-supervised learning via knowledge transfer. In The IEEE International conference on computer vision and pattern recognition (CVPR), pp. 9359\u20139367.","DOI":"10.1109\/CVPR.2018.00975"},{"key":"1555_CR44","unstructured":"Oord, A.v.d., Li, Y., & Vinyals, O.(2018). Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748."},{"issue":"6","key":"1555_CR45","doi-asserted-by":"publisher","first-page":"661","DOI":"10.1109\/TPAMI.1984.4767591","volume":"6","author":"AP Pentland","year":"1984","unstructured":"Pentland, A. P. (1984). Fractal-based description of natural scenes. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 6(6), 661\u2013674.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"},{"key":"1555_CR46","doi-asserted-by":"crossref","unstructured":"Perez, P., Gangnet, M., & Blake, A. (2003). Poisson image editing. ACM Transactions on Graphics (TOG) 22","DOI":"10.1145\/882262.882269"},{"issue":"3","key":"1555_CR47","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1145\/566654.566636","volume":"21","author":"K Perlin","year":"2002","unstructured":"Perlin, K. (2002). Improving noise. ACM Transactions on Graphics (TOG), 21(3), 681\u2013682.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"1555_CR48","doi-asserted-by":"crossref","unstructured":"Remez, T., Huang, J., & Brown, M. (2018). Learning to segment via cut-and-paste. In European conference on computer vision (ECCV).","DOI":"10.1007\/978-3-030-01234-2_3"},{"key":"1555_CR49","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A.G., Zhu, M., Zhmoginov, A., & Chen, L.C. (2018). MobileNetv2: Inverted residuals and linear bottlenecks. Mobile networks for classification, detection and segmentation. arXiv pre-print arXiv:1801.04381.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1555_CR50","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2017). Grad-CAM: Visual explanations from deep networks via gradient-based localization. In The IEEE international conference on computer vision (ICCV), pp. 618\u2013626.","DOI":"10.1109\/ICCV.2017.74"},{"key":"1555_CR51","doi-asserted-by":"crossref","unstructured":"Shrivastava, A., Pfister, T., Tuzel, O., Susskind, J., Wang, W., & Webb, R. (2017). Learning from simulated and unsupervised images through adversarial training. In IEEE international conference on computer vision and pattern recognition (CVPR).","DOI":"10.1109\/CVPR.2017.241"},{"key":"1555_CR52","unstructured":"Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. In International conference on learning representations (ICLR)."},{"key":"1555_CR53","doi-asserted-by":"crossref","unstructured":"Smith, T. G. J., & Marks, W. B. (1996). Fractal methods and results in cellular morphology - dimentions, lacunarity and multifractals. Journal of Neuroscience Methods, 69(2), 123\u2013136.","DOI":"10.1016\/S0165-0270(96)00080-5"},{"key":"1555_CR54","doi-asserted-by":"crossref","unstructured":"Sun, C., Shrivastava, A., Singh, S., & Gupta, A. (2017). Revisiting unreasonable effectiveness of data in deep learning era. In The IEEE international conference on computer vision (ICCV), pp. 843\u2013852.","DOI":"10.1109\/ICCV.2017.97"},{"key":"1555_CR55","doi-asserted-by":"crossref","unstructured":"Sundermeyer, M., Marton, Z.C., Durner, M., Brucker, M., & Triebel, R. (2018). Implicit 3D orientation learning for 6D object detection from RGB images. In European conference on computer vision (ECCV).","DOI":"10.1007\/978-3-030-01231-1_43"},{"key":"1555_CR56","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., & Rabinovich, A. (2015). Going deeper with convolutions. In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 1\u20139.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"1555_CR57","doi-asserted-by":"crossref","unstructured":"Tobin, J., Fong, R., Ray, A., Schneider, J., Zaremba, W., & Abbeel, P. (2017). Domain randomization for transferring deep neural networks from simulation to the real world. In International conference on intelligent robots and systems (IROS).","DOI":"10.1109\/IROS.2017.8202133"},{"key":"1555_CR58","doi-asserted-by":"crossref","unstructured":"Torralba, A., Fergus, R., & Freeman, W.T. (2008). 80 million tiny images: A large data set for nonparametric object and scene recognition. IEEE transactions on pattern analysis and machine intelligence (TPAMI).","DOI":"10.1109\/TPAMI.2008.128"},{"key":"1555_CR59","doi-asserted-by":"crossref","unstructured":"Varma, M., & Garg, R. (2007). Locally invariant fractal features for statistical texture classification. In The IEEE international conference on computer vision (ICCV), pp. 1\u20138.","DOI":"10.1109\/ICCV.2007.4408876"},{"key":"1555_CR60","doi-asserted-by":"crossref","unstructured":"Varol, G., Romero, J., Martin, X., Mahmood, N., Black, M.J., Laptev, I., & Schmid, C. (2017). Learning from synthetic humans. In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 109\u2013117.","DOI":"10.1109\/CVPR.2017.492"},{"key":"1555_CR61","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Doll\u00e1r, P., Tu, Z., & He, K. (2017). Aggregated residual transformations for deep neural networks. In The IEEE international conference on computer vision and pattern recognition (CVPR), pp. 1492\u20131500.","DOI":"10.1109\/CVPR.2017.634"},{"issue":"1","key":"1555_CR62","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1007\/s11263-009-0220-6","volume":"83","author":"Y Xu","year":"2009","unstructured":"Xu, Y., Ji, H., & Fermuller, C. (2009). Viewpoint invariant texture description using fractal analysis. International Journal of Computer Vision (IJCV), 83(1), 85\u2013100.","journal-title":"International Journal of Computer Vision (IJCV)"},{"key":"1555_CR63","doi-asserted-by":"crossref","unstructured":"Yang, K., Qinami, K., Fei-Fei, L., Deng, J., & Russakovsky, O. (2020). Towards fairer datasets: Filtering and balancing the distribution of the people subtree in the ImageNet hierarchy. In Conference on fairness, accountability and transparency (FAT).","DOI":"10.1145\/3351095.3375709"},{"key":"1555_CR64","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., & Efros, A.A. (2016). Colorful image colorization. In European conference on computer vision (ECCV).","DOI":"10.1007\/978-3-319-46487-9_40"},{"key":"1555_CR65","doi-asserted-by":"publisher","first-page":"1452","DOI":"10.1109\/TPAMI.2017.2723009","volume":"40","author":"B Zhou","year":"2017","unstructured":"Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., & Torralba, A. (2017). Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 40, 1452\u20131464.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-021-01555-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-021-01555-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-021-01555-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,3,25]],"date-time":"2022-03-25T10:41:45Z","timestamp":1648204905000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-021-01555-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,2,24]]},"references-count":65,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2022,4]]}},"alternative-id":["1555"],"URL":"https:\/\/doi.org\/10.1007\/s11263-021-01555-8","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,2,24]]},"assertion":[{"value":"29 March 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 November 2021","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 February 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}