{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:07:26Z","timestamp":1765544846400,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,8,14]],"date-time":"2022-08-14T00:00:00Z","timestamp":1660435200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,8,14]]},"DOI":"10.1145\/3534678.3539118","type":"proceedings-article","created":{"date-parts":[[2022,8,12]],"date-time":"2022-08-12T19:06:41Z","timestamp":1660331201000},"page":"3219-3229","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Design Domain Specific Neural Network via Symbolic Testing"],"prefix":"10.1145","author":[{"given":"Hui","family":"Li","sequence":"first","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Xing","family":"Fu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Ruofan","family":"Wu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Jinyu","family":"Xu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Kai","family":"Xiao","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Xiaofu","family":"Chang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Weiqiang","family":"Wang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Shuai","family":"Chen","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Leilei","family":"Shi","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Tao","family":"Xiong","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"given":"Yuan","family":"Qi","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2022,8,14]]},"reference":[{"key":"e_1_3_2_2_1_1","first-page":"3351","article-title":"Higher-order factorization machines","volume":"29","author":"Blondel M.","year":"2016","unstructured":"Blondel, M., Fujino, A., Ueda, N., and Ishihata, M. Higher-order factorization machines. NeurIPS, 29:3351--3359, 2016.","journal-title":"NeurIPS"},{"key":"e_1_3_2_2_2_1","volume-title":"ICML","author":"Blondel M.","year":"2016","unstructured":"Blondel, M., Ishihata, M., Fujino, A., and Ueda, N. Polynomial networks and factorization machines: New insights and efficient training algorithms. ICML, 2016."},{"key":"e_1_3_2_2_3_1","volume-title":"Neural model-based reinforcement learning for recommendation","author":"Chen X.","year":"2018","unstructured":"Chen, X., Li, S., Li, H., Jiang, S., and Song, L. Neural model-based reinforcement learning for recommendation. 2018."},{"key":"e_1_3_2_2_4_1","volume-title":"ICML","author":"Chen X.","year":"2019","unstructured":"Chen, X., Li, S., Li, H., Jiang, S., Qi, Y., and Song, L. Generative adversarial user model for reinforcement learning based recommendation system. In ICML, 2019."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/2988450.2988454"},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5768"},{"key":"e_1_3_2_2_7_1","volume-title":"Rethinking attention with performers. arXiv","author":"Choromanski K.","year":"2020","unstructured":"Choromanski, K., Likhosherstov, V., Dohan, D., Song, X., Gane, A., Sarlos, T., Hawkins, P., Davis, J., Mohiuddin, A., Kaiser, L., et al. Rethinking attention with performers. arXiv, 2020."},{"key":"e_1_3_2_2_8_1","first-page":"1115","volume-title":"International conference on machine learning (ICML)","author":"Dai H.","year":"2018","unstructured":"Dai, H., Li, H., Tian, T., Huang, X., Wang, L., Zhu, J., and Song, L. Adversarial attack on graph structured data. In International conference on machine learning (ICML), pp. 1115--1124. PMLR, 2018."},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.5120\/ijca2017913081"},{"key":"e_1_3_2_2_10_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv","author":"Devlin J.","year":"2018","unstructured":"Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv, 2018."},{"key":"e_1_3_2_2_11_1","volume-title":"Deep session interest network for click-through rate prediction. arXiv","author":"Feng Y.","year":"2019","unstructured":"Feng, Y., Lv, F., Shen, W., Wang, M., Sun, F., Zhu, Y., and Yang, K. Deep session interest network for click-through rate prediction. arXiv, 2019."},{"key":"e_1_3_2_2_12_1","volume-title":"Neural-symbolic learning and reasoning: contributions and challenges","author":"Garcez A.","year":"2015","unstructured":"Garcez, A., Besold, T. R., Raedt, L., Foldiak, P., Hitzler, P., Icard, T., Kuhnberger, K.-U., Lamb, L. C., Miikkulainen, R., and Silver, D. L. Neural-symbolic learning and reasoning: contributions and challenges. 2015."},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10710-017-9314-z"},{"key":"e_1_3_2_2_14_1","volume-title":"JMLR","author":"Gretton A.","year":"2012","unstructured":"Gretton, A., Borgwardt, K. M., Rasch, M. J., Sch\u00f6lkopf, B., and Smola, A. A kernel two-sample test. JMLR, 2012."},{"key":"e_1_3_2_2_15_1","volume-title":"Deepfm: a factorization-machine based neural network for ctr prediction. arXiv","author":"Guo H.","year":"2017","unstructured":"Guo, H., Tang, R., Ye, Y., Li, Z., and He, X. Deepfm: a factorization-machine based neural network for ctr prediction. arXiv, 2017."},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3077136.3080777"},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICNN.1996.549076"},{"key":"e_1_3_2_2_19_1","first-page":"1735","article-title":"Long short-term memory","volume":"8","author":"Hochreiter S.","unstructured":"Hochreiter, S. and Schmidhuber, J. Long short-term memory. Number 8, pp. 1735--1780. Neural computation, 1997.","journal-title":"Number"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3415959.3416001"},{"key":"e_1_3_2_2_21_1","first-page":"5562","volume-title":"International Conference on Machine Learning","author":"Kim H.","year":"2021","unstructured":"Kim, H., Papamakarios, G., and Mnih, A. The lipschitz constant of self-attention. In International Conference on Machine Learning, pp. 5562--5571. PMLR, 2021."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2013.6738123"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1117\/1.OE.52.6.067003"},{"key":"e_1_3_2_2_24_1","volume-title":"ICLR","author":"Li H.","year":"2019","unstructured":"Li, H., Hu, K., Zhang, S., Qi, Y., and Song, L. Double neural counterfactual regret minimization. In ICLR, 2019."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220023"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.5120\/ijca2017915495"},{"key":"e_1_3_2_2_27_1","volume-title":"Intention propagation for multi-agent reinforcement learning. arXiv preprint arXiv:2004.08883","author":"Qu C.","year":"2020","unstructured":"Qu, C., Li, H., Liu, C., Xiong, J., Zhang, J., Chu, W., Wang, W., Qi, Y., and Song, L. Intention propagation for multi-agent reinforcement learning. arXiv preprint arXiv:2004.08883, 2020."},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2016.0151"},{"key":"e_1_3_2_2_29_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog","author":"Radford A.","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., and Sutskever, I. Language models are unsupervised multitask learners. OpenAI blog, 2019."},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2010.127"},{"key":"e_1_3_2_2_31_1","volume-title":"Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. arXiv","author":"Saxe A. M.","year":"2013","unstructured":"Saxe, A. M., McClelland, J. L., and Ganguli, S. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. arXiv, 2013."},{"key":"e_1_3_2_2_32_1","volume-title":"Mastering the game of Go with deep neural networks and tree search. Nature, (7587)","author":"Silver D.","year":"2016","unstructured":"Silver, D., Huang, A., Maddison, C. J., Guez, A., Sifre, L., Driessche, G. V. D., and et al., J. S. Mastering the game of Go with deep neural networks and tree search. Nature, (7587), 2016."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357925"},{"key":"e_1_3_2_2_34_1","volume-title":"Neural Information Processing Systems (NeurIPS) on Learning Meets Combinatorial Algorithms","author":"Sun H.","year":"2020","unstructured":"Sun, H., Chen, W., Li, H., and Song, L. Improving learning to branch via reinforcement learning. Neural Information Processing Systems (NeurIPS) on Learning Meets Combinatorial Algorithms, 2020."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNN.1998.712192"},{"key":"e_1_3_2_2_36_1","volume-title":"Long range arena: A benchmark for efficient transformers. arXiv preprint arXiv:2011.04006","author":"Tay Y.","year":"2020","unstructured":"Tay, Y., Dehghani, M., Abnar, S., Shen, Y., Bahri, D., Pham, P., Rao, J., Yang, L., Ruder, S., and Metzler, D. Long range arena: A benchmark for efficient transformers. arXiv preprint arXiv:2011.04006, 2020."},{"key":"e_1_3_2_2_37_1","volume-title":"NeurIPS","author":"Vaswani A.","year":"2017","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., and Polosukhin, I. Attention is all you need. In NeurIPS, 2017."},{"key":"e_1_3_2_2_38_1","volume-title":"Tcl: Transformer-based dynamic graph modelling via contrastive learning. arXiv preprint arXiv:2105.07944","author":"Wang L.","year":"2021","unstructured":"Wang, L., Chang, X., Li, S., Chu, Y., Li, H., Zhang, W., He, X., Song, L., Zhou, J., and Yang, H. Tcl: Transformer-based dynamic graph modelling via contrastive learning. arXiv preprint arXiv:2105.07944, 2021."},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3124749.3124754"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401307"},{"key":"e_1_3_2_2_41_1","volume-title":"-S. Attentional factorization machines: Learning the weight of feature interactions via attention networks. arXiv","author":"Xiao J.","year":"2017","unstructured":"Xiao, J., Ye, H., He, X., Zhang, H.,Wu, F., and Chua, T.-S. Attentional factorization machines: Learning the weight of feature interactions via attention networks. arXiv, 2017."},{"key":"e_1_3_2_2_42_1","volume-title":"International Conference on Learning Representations","author":"Xu K.","year":"2019","unstructured":"Xu, K., Hu, W., Leskovec, J., and Jegelka, S. How powerful are graph neural networks? In International Conference on Learning Representations, 2019. URL https:\/\/openreview.net\/forum?id=ryGs6iA5Km."},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-30671-1_4"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2016.12.006"},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219823"},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380172"}],"event":{"name":"KDD '22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"],"location":"Washington DC USA","acronym":"KDD '22"},"container-title":["Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3534678.3539118","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3534678.3539118","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:58Z","timestamp":1750186978000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3534678.3539118"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,14]]},"references-count":46,"alternative-id":["10.1145\/3534678.3539118","10.1145\/3534678"],"URL":"https:\/\/doi.org\/10.1145\/3534678.3539118","relation":{},"subject":[],"published":{"date-parts":[[2022,8,14]]},"assertion":[{"value":"2022-08-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}