{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:12:36Z","timestamp":1750219956879,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":45,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,10,17]],"date-time":"2022-10-17T00:00:00Z","timestamp":1665964800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,10,17]]},"DOI":"10.1145\/3511808.3557259","type":"proceedings-article","created":{"date-parts":[[2022,10,16]],"date-time":"2022-10-16T01:22:22Z","timestamp":1665883342000},"page":"2413-2422","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Cognize Yourself"],"prefix":"10.1145","author":[{"given":"Tao","family":"Yu","sequence":"first","affiliation":[{"name":"Hikvision Research Institute, Hangzhou, China"}]},{"given":"Yao","family":"Fu","sequence":"additional","affiliation":[{"name":"Hikvision Research Institute, Hangzhou, China"}]},{"given":"Linghui","family":"Hu","sequence":"additional","affiliation":[{"name":"Hikvision Research Institute, Hangzhou, China"}]},{"given":"Huizhao","family":"Wang","sequence":"additional","affiliation":[{"name":"Hikvision Research Institute, Hangzhou, China"}]},{"given":"Weihao","family":"Jiang","sequence":"additional","affiliation":[{"name":"Hikvision Research Institute, Hangzhou, China"}]},{"given":"Shiliang","family":"Pu","sequence":"additional","affiliation":[{"name":"Hikvision Research Institute, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2022,10,17]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171--4186.","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171--4186. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171--4186."},{"key":"e_1_3_2_2_2_1","volume-title":"Decaf: A deep convolutional activation feature for generic visual recognition. In ICML. 647--655.","author":"Donahue Jeff","year":"2014","unstructured":"Jeff Donahue , Yangqing Jia , Oriol Vinyals , Judy Hoffman , Ning Zhang , Eric Tzeng , and Trevor Darrell . 2014 . Decaf: A deep convolutional activation feature for generic visual recognition. In ICML. 647--655. Jeff Donahue, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor Darrell. 2014. Decaf: A deep convolutional activation feature for generic visual recognition. In ICML. 647--655."},{"key":"e_1_3_2_2_3_1","unstructured":"Chelsea Finn Pieter Abbeel and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML. 1126--1135. Chelsea Finn Pieter Abbeel and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML. 1126--1135."},{"key":"e_1_3_2_2_4_1","volume-title":"On cognitive development. Child development","author":"Flavell John H","year":"1982","unstructured":"John H Flavell . 1982. On cognitive development. Child development ( 1982 ), 1--10. John H Flavell. 1982. On cognitive development. Child development (1982), 1--10."},{"key":"e_1_3_2_2_5_1","volume-title":"Metacognition: ideas and insights from neuro-and educational sciences. npj Science of Learning","author":"Fleur Damien S","year":"2021","unstructured":"Damien S Fleur , Bert Bredeweg , and Wouter van den Bos . 2021. Metacognition: ideas and insights from neuro-and educational sciences. npj Science of Learning , Vol. 6 , 1 ( 2021 ), 1--11. Damien S Fleur, Bert Bredeweg, and Wouter van den Bos. 2021. Metacognition: ideas and insights from neuro-and educational sciences. npj Science of Learning, Vol. 6, 1 (2021), 1--11."},{"key":"e_1_3_2_2_6_1","unstructured":"Justin Gilmer Samuel S Schoenholz Patrick F Riley Oriol Vinyals and George E Dahl. 2017. Neural message passing for quantum chemistry. In ICML. 1263--1272. Justin Gilmer Samuel S Schoenholz Patrick F Riley Oriol Vinyals and George E Dahl. 2017. Neural message passing for quantum chemistry. In ICML. 1263--1272."},{"key":"e_1_3_2_2_7_1","unstructured":"Will Hamilton Zhitao Ying and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS. 1024--1034. Will Hamilton Zhitao Ying and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS. 1024--1034."},{"key":"e_1_3_2_2_8_1","unstructured":"R Devon Hjelm Alex Fedorov Samuel Lavoie-Marchildon Karan Grewal Phil Bachman Adam Trischler and Yoshua Bengio. 2019. Learning deep representations by mutual information estimation and maximization. In ICLR. R Devon Hjelm Alex Fedorov Samuel Lavoie-Marchildon Karan Grewal Phil Bachman Adam Trischler and Yoshua Bengio. 2019. Learning deep representations by mutual information estimation and maximization. In ICLR."},{"key":"e_1_3_2_2_9_1","first-page":"22118","article-title":"Open graph benchmark: Datasets for machine learning on graphs","volume":"33","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu , Matthias Fey , Marinka Zitnik , Yuxiao Dong , Hongyu Ren , Bowen Liu , Michele Catasta , and Jure Leskovec . 2020 c. Open graph benchmark: Datasets for machine learning on graphs . In NeurIPS , Vol. 33. 22118 -- 22133 . Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020c. Open graph benchmark: Datasets for machine learning on graphs. In NeurIPS, Vol. 33. 22118--22133.","journal-title":"NeurIPS"},{"key":"e_1_3_2_2_10_1","unstructured":"Weihua Hu Bowen Liu Joseph Gomes Marinka Zitnik Percy Liang Vijay Pande and Jure Leskovec. 2020d. Strategies for pre-training graph neural networks. In ICLR. Weihua Hu Bowen Liu Joseph Gomes Marinka Zitnik Percy Liang Vijay Pande and Jure Leskovec. 2020d. Strategies for pre-training graph neural networks. In ICLR."},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403237"},{"key":"e_1_3_2_2_12_1","unstructured":"Ziniu Hu Yuxiao Dong Kuansan Wang and Yizhou Sun. 2020a. Heterogeneous graph transformer. In WWW. 2704--2710. Ziniu Hu Yuxiao Dong Kuansan Wang and Yizhou Sun. 2020a. Heterogeneous graph transformer. In WWW. 2704--2710."},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"crossref","unstructured":"Hong Huang Zixuan Fang Xiao Wang Youshan Miao and Hai Jin. 2020. Motif-Preserving Temporal Network Embedding. In IJCAI. 1237--1243. Hong Huang Zixuan Fang Xiao Wang Youshan Miao and Hai Jin. 2020. Motif-Preserving Temporal Network Embedding. In IJCAI. 1237--1243.","DOI":"10.24963\/ijcai.2020\/172"},{"key":"e_1_3_2_2_14_1","unstructured":"Eric Jang Shixiang Gu and Ben Poole. 2016. Categorical reparameterization with gumbel-softmax. In ICLR. Eric Jang Shixiang Gu and Ben Poole. 2016. Categorical reparameterization with gumbel-softmax. In ICLR."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"crossref","unstructured":"Xunqiang Jiang Yuanfu Lu Yuan Fang and Chuan Shi. 2021. Contrastive Pre-Training of GNNs on Heterogeneous Graphs. In CIKM. 803--812. Xunqiang Jiang Yuanfu Lu Yuan Fang and Chuan Shi. 2021. Contrastive Pre-Training of GNNs on Heterogeneous Graphs. In CIKM. 803--812.","DOI":"10.1145\/3459637.3482332"},{"key":"e_1_3_2_2_16_1","volume-title":"Issues in Interdisciplinary Studies","volume":"35","author":"Keestra Machiel","year":"2017","unstructured":"Machiel Keestra . 2017 . Metacognition and reflection by interdisciplinary experts: Insights from cognitive science and philosophy . Issues in Interdisciplinary Studies , Vol. 35 (2017). Machiel Keestra. 2017. Metacognition and reflection by interdisciplinary experts: Insights from cognitive science and philosophy. Issues in Interdisciplinary Studies, Vol. 35 (2017)."},{"volume-title":"CoRR","author":"Kipf Thomas N","key":"e_1_3_2_2_17_1","unstructured":"Thomas N Kipf and Max Welling . 2016. Variational graph auto-encoders . In CoRR , Vol. abs\/ 1611 .07308. Thomas N Kipf and Max Welling. 2016. Variational graph auto-encoders. In CoRR, Vol. abs\/1611.07308."},{"key":"e_1_3_2_2_18_1","unstructured":"Thomas N Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. In ICLR. Thomas N Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. In ICLR."},{"key":"e_1_3_2_2_19_1","unstructured":"John Boaz Lee Ryan Rossi and Xiangnan Kong. 2018. Graph classification using structural attention. In SIGKDD. 1666--1674. John Boaz Lee Ryan Rossi and Xiangnan Kong. 2018. Graph classification using structural attention. In SIGKDD. 1666--1674."},{"volume-title":"The analects of Confucius","author":"Leys Simon","key":"e_1_3_2_2_20_1","unstructured":"Simon Leys . 1997. The analects of Confucius . WW Norton & Company . Simon Leys. 1997. The analects of Confucius. WW Norton & Company."},{"key":"e_1_3_2_2_21_1","first-page":"3546","article-title":"Adaptive graph convolutional neural networks","volume":"32","author":"Li Ruoyu","year":"2018","unstructured":"Ruoyu Li , Sheng Wang , Feiyun Zhu , and Junzhou Huang . 2018 . Adaptive graph convolutional neural networks . In AAAI , Vol. 32. 3546 -- 3553 . Ruoyu Li, Sheng Wang, Feiyun Zhu, and Junzhou Huang. 2018. Adaptive graph convolutional neural networks. In AAAI, Vol. 32. 3546--3553.","journal-title":"AAAI"},{"key":"e_1_3_2_2_22_1","volume-title":"Self-supervised learning: Generative or contrastive. TKDE","author":"Liu Xiao","year":"2021","unstructured":"Xiao Liu , Fanjin Zhang , Zhenyu Hou , Li Mian , Zhaoyu Wang , Jing Zhang , and Jie Tang . 2021b. Self-supervised learning: Generative or contrastive. TKDE ( 2021 ). Xiao Liu, Fanjin Zhang, Zhenyu Hou, Li Mian, Zhaoyu Wang, Jing Zhang, and Jie Tang. 2021b. Self-supervised learning: Generative or contrastive. TKDE (2021)."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"crossref","unstructured":"Zhijun Liu Chao Huang Yanwei Yu and Junyu Dong. 2021a. Motif-preserving dynamic attributed network embedding. In WWW. 1629--1638. Zhijun Liu Chao Huang Yanwei Yu and Junyu Dong. 2021a. Motif-preserving dynamic attributed network embedding. In WWW. 1629--1638.","DOI":"10.1145\/3442381.3449821"},{"key":"e_1_3_2_2_24_1","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. In ICLR. Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. In ICLR."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"crossref","unstructured":"Yuanfu Lu Xunqiang Jiang Yuan Fang and Chuan Shi. 2021. Learning to pre-train graph neural networks. In AAAI. 4276--4284. Yuanfu Lu Xunqiang Jiang Yuan Fang and Chuan Shi. 2021. Learning to pre-train graph neural networks. In AAAI. 4276--4284.","DOI":"10.1609\/aaai.v35i5.16552"},{"volume-title":"CoRR","author":"Navarin Nicol\u00f2","key":"e_1_3_2_2_26_1","unstructured":"Nicol\u00f2 Navarin , Dinh V Tran , and Alessandro Sperduti . 2018. Pre-training graph neural networks with kernels . In CoRR , Vol. abs\/ 1811 .06930. Nicol\u00f2 Navarin, Dinh V Tran, and Alessandro Sperduti. 2018. Pre-training graph neural networks with kernels. In CoRR, Vol. abs\/1811.06930."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"crossref","unstructured":"Zhen Peng Wenbing Huang Minnan Luo Qinghua Zheng Yu Rong Tingyang Xu and Junzhou Huang. 2020. Graph representation learning via graphical mutual information maximization. In WWW. 259--270. Zhen Peng Wenbing Huang Minnan Luo Qinghua Zheng Yu Rong Tingyang Xu and Junzhou Huang. 2020. Graph representation learning via graphical mutual information maximization. In WWW. 259--270.","DOI":"10.1145\/3366423.3380112"},{"key":"e_1_3_2_2_28_1","volume-title":"The role of metacognitive knowledge in learning, teaching, and assessing. Theory into practice","author":"Pintrich Paul R","year":"2002","unstructured":"Paul R Pintrich . 2002. The role of metacognitive knowledge in learning, teaching, and assessing. Theory into practice , Vol. 41 , 4 ( 2002 ), 219--225. Paul R Pintrich. 2002. The role of metacognitive knowledge in learning, teaching, and assessing. Theory into practice, Vol. 41, 4 (2002), 219--225."},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403168"},{"volume-title":"CoRR","author":"Shchur Oleksandr","key":"e_1_3_2_2_30_1","unstructured":"Oleksandr Shchur , Maximilian Mumme , Aleksandar Bojchevski , and Stephan G\u00fcnnemann . 2018. Pitfalls of graph neural network evaluation . In CoRR , Vol. abs\/ 1811 .05868. Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan G\u00fcnnemann. 2018. Pitfalls of graph neural network evaluation. In CoRR, Vol. abs\/1811.05868."},{"key":"e_1_3_2_2_31_1","volume-title":"Super-convergence: Very fast training of neural networks using large learning rates. In Artificial intelligence and machine learning for multi-domain operations applications","author":"Smith Leslie N","year":"2019","unstructured":"Leslie N Smith and Nicholay Topin . 2019 . Super-convergence: Very fast training of neural networks using large learning rates. In Artificial intelligence and machine learning for multi-domain operations applications , Vol. 11006 . 1100612. Leslie N Smith and Nicholay Topin. 2019. Super-convergence: Very fast training of neural networks using large learning rates. In Artificial intelligence and machine learning for multi-domain operations applications, Vol. 11006. 1100612."},{"volume-title":"Cognitive science: An introduction","author":"Stillings Neil A","key":"e_1_3_2_2_32_1","unstructured":"Neil A Stillings , Christopher H Chase , Steven E Weisler , Mark H Feinstein , Jay L Garfield , and Edwina L Rissland . 1995. Cognitive science: An introduction . MIT press . Neil A Stillings, Christopher H Chase, Steven E Weisler, Mark H Feinstein, Jay L Garfield, and Edwina L Rissland. 1995. Cognitive science: An introduction. MIT press."},{"key":"e_1_3_2_2_33_1","volume-title":"Infograph: Unsupervised and semi-supervised graph-level representation learning via mutual information maximization. In ICLR.","author":"Sun Fan-Yun","year":"2020","unstructured":"Fan-Yun Sun , Jordan Hoffmann , Vikas Verma , and Jian Tang . 2020 . Infograph: Unsupervised and semi-supervised graph-level representation learning via mutual information maximization. In ICLR. Fan-Yun Sun, Jordan Hoffmann, Vikas Verma, and Jian Tang. 2020. Infograph: Unsupervised and semi-supervised graph-level representation learning via mutual information maximization. In ICLR."},{"key":"e_1_3_2_2_34_1","volume-title":"Mind: Introduction to cognitive science","author":"Thagard Paul","year":"2005","unstructured":"Paul Thagard . 2005 . Mind: Introduction to cognitive science . MIT press . Paul Thagard. 2005. Mind: Introduction to cognitive science. MIT press."},{"volume-title":"CoRR","author":"den Oord Aaron Van","key":"e_1_3_2_2_35_1","unstructured":"Aaron Van den Oord , Yazhe Li , and Oriol Vinyals . 2018. Representation learning with contrastive predictive coding . In CoRR , Vol. abs\/ 1807 .03748. Aaron Van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding. In CoRR, Vol. abs\/1807.03748."},{"key":"e_1_3_2_2_36_1","unstructured":"Petar Veli\u010dkovi\u0107 Guillem Cucurull Arantxa Casanova Adriana Romero Pietro Lio and Yoshua Bengio. 2017. Graph attention networks. In ICLR. Petar Veli\u010dkovi\u0107 Guillem Cucurull Arantxa Casanova Adriana Romero Pietro Lio and Yoshua Bengio. 2017. Graph attention networks. In ICLR."},{"key":"e_1_3_2_2_37_1","unstructured":"Petar Veli\u010dkovi\u0107 William Fedus William L Hamilton Pietro Li\u00f2 Yoshua Bengio and R Devon Hjelm. 2019. Deep Graph Infomax. In ICLR. Petar Veli\u010dkovi\u0107 William Fedus William L Hamilton Pietro Li\u00f2 Yoshua Bengio and R Devon Hjelm. 2019. Deep Graph Infomax. In ICLR."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"crossref","unstructured":"Barbara Von Eckardt. 1995. What is cognitive science? MIT press. Barbara Von Eckardt. 1995. What is cognitive science? MIT press.","DOI":"10.7551\/mitpress\/7209.001.0001"},{"key":"e_1_3_2_2_39_1","unstructured":"Felix Wu Amauri Souza Tianyi Zhang Christopher Fifty Tao Yu and Kilian Weinberger. 2019. Simplifying graph convolutional networks. In ICML. 6861--6871. Felix Wu Amauri Souza Tianyi Zhang Christopher Fifty Tao Yu and Kilian Weinberger. 2019. Simplifying graph convolutional networks. In ICML. 6861--6871."},{"key":"e_1_3_2_2_40_1","unstructured":"Keyulu Xu Weihua Hu Jure Leskovec and Stefanie Jegelka. 2019. How powerful are graph neural networks?. In ICLR. Keyulu Xu Weihua Hu Jure Leskovec and Stefanie Jegelka. 2019. How powerful are graph neural networks?. In ICLR."},{"key":"e_1_3_2_2_41_1","unstructured":"Yuning You Tianlong Chen Zhangyang Wang and Yang Shen. 2020. When does self-supervision help graph convolutional networks?. In ICML. 10871--10880. Yuning You Tianlong Chen Zhangyang Wang and Yang Shen. 2020. When does self-supervision help graph convolutional networks?. In ICML. 10871--10880."},{"key":"e_1_3_2_2_42_1","first-page":"11960","article-title":"Graph transformer networks","volume":"32","author":"Yun Seongjun","year":"2019","unstructured":"Seongjun Yun , Minbyul Jeong , Raehyun Kim , Jaewoo Kang , and Hyunwoo J Kim . 2019 . Graph transformer networks . In NeurIPS , Vol. 32. 11960 -- 11970 . Seongjun Yun, Minbyul Jeong, Raehyun Kim, Jaewoo Kang, and Hyunwoo J Kim. 2019. Graph transformer networks. In NeurIPS, Vol. 32. 11960--11970.","journal-title":"NeurIPS"},{"key":"e_1_3_2_2_43_1","volume-title":"Graphsaint: Graph sampling based inductive learning method. In ICLR.","author":"Zeng Hanqing","year":"2020","unstructured":"Hanqing Zeng , Hongkuan Zhou , Ajitesh Srivastava , Rajgopal Kannan , and Viktor Prasanna . 2020 . Graphsaint: Graph sampling based inductive learning method. In ICLR. Hanqing Zeng, Hongkuan Zhou, Ajitesh Srivastava, Rajgopal Kannan, and Viktor Prasanna. 2020. Graphsaint: Graph sampling based inductive learning method. In ICLR."},{"key":"e_1_3_2_2_44_1","volume-title":"Graph-bert: Only attention is needed for learning graph representations. In CoRR","author":"Zhang Jiawei","year":"2020","unstructured":"Jiawei Zhang , Haopeng Zhang , Congying Xia , and Li Sun . 2020 . Graph-bert: Only attention is needed for learning graph representations. In CoRR , Vol. abs\/ 2001 .05140. Jiawei Zhang, Haopeng Zhang, Congying Xia, and Li Sun. 2020. Graph-bert: Only attention is needed for learning graph representations. In CoRR, Vol. abs\/2001.05140."},{"key":"e_1_3_2_2_45_1","first-page":"5171","article-title":"Link prediction based on graph neural networks","volume":"31","author":"Zhang Muhan","year":"2018","unstructured":"Muhan Zhang and Yixin Chen . 2018 . Link prediction based on graph neural networks . In NeurIPS , Vol. 31. 5171 -- 5181 . Muhan Zhang and Yixin Chen. 2018. Link prediction based on graph neural networks. In NeurIPS, Vol. 31. 5171--5181.","journal-title":"NeurIPS"}],"event":{"name":"CIKM '22: The 31st ACM International Conference on Information and Knowledge Management","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web","SIGIR ACM Special Interest Group on Information Retrieval"],"location":"Atlanta GA USA","acronym":"CIKM '22"},"container-title":["Proceedings of the 31st ACM International Conference on Information &amp; Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3511808.3557259","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3511808.3557259","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T17:49:07Z","timestamp":1750182547000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3511808.3557259"}},"subtitle":["Graph Pre-Training via Core Graph Cognizing and Differentiating"],"short-title":[],"issued":{"date-parts":[[2022,10,17]]},"references-count":45,"alternative-id":["10.1145\/3511808.3557259","10.1145\/3511808"],"URL":"https:\/\/doi.org\/10.1145\/3511808.3557259","relation":{},"subject":[],"published":{"date-parts":[[2022,10,17]]},"assertion":[{"value":"2022-10-17","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}