{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:34:37Z","timestamp":1742913277510,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":39,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819755547"},{"type":"electronic","value":"9789819755554"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-5555-4_31","type":"book-chapter","created":{"date-parts":[[2025,1,11]],"date-time":"2025-01-11T05:40:20Z","timestamp":1736574020000},"page":"435-450","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CF-CAD: A Contrastive Fusion Network For 3D Computer-Aided Design Generative Modeling"],"prefix":"10.1007","author":[{"given":"Xueyang","family":"Li","sequence":"first","affiliation":[]},{"given":"Haotian","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yunzhong","family":"Lou","sequence":"additional","affiliation":[]},{"given":"Xiangdong","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,12]]},"reference":[{"unstructured":"Achlioptas, P., Diamanti, O., Mitliagkas, I., Guibas, L.: Learning representations and generative models for 3d point clouds. In: ICML. pp. 40\u201349 (2018)","key":"31_CR1"},{"doi-asserted-by":"crossref","unstructured":"Bagad, P., Tapaswi, M., Snoek, C.G.: Test of time: Instilling video-language models with a sense of time. In: CVPR. pp. 2503\u20132516 (2023)","key":"31_CR2","DOI":"10.1109\/CVPR52729.2023.00247"},{"doi-asserted-by":"crossref","unstructured":"Cao, W., Robinson, T., Hua, Y., Boussuge, F., Colligan, A.R., Pan, W.: Graph representation of 3d cad models for machining feature recognition with deep learning. In: International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. vol. 84003, p. V11AT11A003. American Society of Mechanical Engineers (2020)","key":"31_CR3","DOI":"10.1115\/DETC2020-22355"},{"doi-asserted-by":"crossref","unstructured":"Cascante-Bonilla, P., Shehada, K., Smith, J.S., Doveh, S., Kim, D., Panda, R., Varol, G., Oliva, A., Ordonez, V., Feris, R., et\u00a0al.: Going beyond nouns with vision & language models using synthetic data. In: ICCV. pp. 20155\u201320165 (2023)","key":"31_CR4","DOI":"10.1109\/ICCV51070.2023.01844"},{"doi-asserted-by":"crossref","unstructured":"Cole, E., Yang, X., Wilber, K., Mac\u00a0Aodha, O., Belongie, S.: When does contrastive visual representation learning work? In: CVPR. pp. 14755\u201314764 (2022)","key":"31_CR5","DOI":"10.1109\/CVPR52688.2022.01434"},{"doi-asserted-by":"crossref","unstructured":"Das, S.S.S., Katiyar, A., Passonneau, R.J., Zhang, R.: Container: Few-shot named entity recognition via contrastive learning. In: ACL (2022)","key":"31_CR6","DOI":"10.18653\/v1\/2022.acl-long.439"},{"doi-asserted-by":"crossref","unstructured":"Gao, T., Yao, X., Chen, D.: SimCSE: Simple contrastive learning of sentence embeddings. In: EMNLP (2021)","key":"31_CR7","DOI":"10.18653\/v1\/2021.emnlp-main.552"},{"doi-asserted-by":"crossref","unstructured":"Ge, S., Mishra, S., Kornblith, S., Li, C.L., Jacobs, D.: Hyperbolic contrastive learning for visual representations beyond objects. In: CVPR. pp. 6840\u20136849 (2023)","key":"31_CR8","DOI":"10.1109\/CVPR52729.2023.00661"},{"unstructured":"Grill, J.B., Strub, F., Altch\u00e9, F., Tallec, C., Richemond, P., Buchatskaya, E., Doersch, C., Avila\u00a0Pires, B., Guo, Z., Gheshlaghi\u00a0Azar, M., et\u00a0al.: Bootstrap your own latent-a new approach to self-supervised learning. NeurIPS 33, 21271\u201321284 (2020)","key":"31_CR9"},{"doi-asserted-by":"crossref","unstructured":"Guo, H., Liu, S., Pan, H., Liu, Y., Tong, X., Guo, B.: Complexgen: Cad reconstruction by b-rep chain complex generation. ACM Transactions on Graphics (TOG) 41(4), 1\u201318 (2022)","key":"31_CR10","DOI":"10.1145\/3528223.3530078"},{"doi-asserted-by":"crossref","unstructured":"Hadsell, R., Chopra, S., LeCun, Y.: Dimensionality reduction by learning an invariant mapping. In: CVPR. pp. 1735\u20131742 (2006)","key":"31_CR11","DOI":"10.1109\/CVPR.2006.100"},{"doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR. pp. 770\u2013778 (2016)","key":"31_CR12","DOI":"10.1109\/CVPR.2016.90"},{"doi-asserted-by":"crossref","unstructured":"Jones, B.T., Hu, M., Kodnongbua, M., Kim, V.G., Schulz, A.: Self-supervised representation learning for cad. In: CVPR. pp. 21327\u201321336 (2023)","key":"31_CR13","DOI":"10.1109\/CVPR52729.2023.02043"},{"doi-asserted-by":"crossref","unstructured":"Kamath, A., Anderson, P., Wang, S., Koh, J.Y., Ku, A., Waters, A., Yang, Y., Baldridge, J., Parekh, Z.: A new path: Scaling vision-and-language navigation with synthetic instructions and imitation learning. In: CVPR. pp. 10813\u201310823 (2023)","key":"31_CR14","DOI":"10.1109\/CVPR52729.2023.01041"},{"doi-asserted-by":"crossref","unstructured":"Kinakh, V., Taran, O., Voloshynovskiy, S.: Scatsimclr: self-supervised contrastive learning with pretext task regularization for small-scale datasets. In: ICCV (2021)","key":"31_CR15","DOI":"10.1109\/ICCVW54120.2021.00129"},{"doi-asserted-by":"crossref","unstructured":"Lambourne, J.G., Willis, K.D., Jayaraman, P.K., Sanghi, A., Meltzer, P., Shayani, H.: Brepnet: A topological message passing system for solid models. In: CVPR. pp. 12773\u201312782 (2021)","key":"31_CR16","DOI":"10.1109\/CVPR46437.2021.01258"},{"doi-asserted-by":"crossref","unstructured":"Li, P., Guo, J., Zhang, X., Yan, D.M.: Secad-net: Self-supervised cad reconstruction by learning sketch-extrude operations. In: CVPR. pp. 16816\u201316826 (2023)","key":"31_CR17","DOI":"10.1109\/CVPR52729.2023.01613"},{"doi-asserted-by":"crossref","unstructured":"Li, X., Xu, M., Zhou, X.: Twins-mix: Self mixing in latent space for reasonable data augmentation of 3d computer-aided design generative modeling. In: IEEE International Conference on Multimedia and Expo (ICME). pp. 906\u2013911 (2023)","key":"31_CR18","DOI":"10.1109\/ICME55011.2023.00160"},{"doi-asserted-by":"crossref","unstructured":"Lin, Y.B., Sung, Y.L., Lei, J., Bansal, M., Bertasius, G.: Vision transformers are parameter-efficient audio-visual learners. In: CVPR. pp. 2299\u20132309 (2023)","key":"31_CR19","DOI":"10.1109\/CVPR52729.2023.00228"},{"doi-asserted-by":"crossref","unstructured":"Liu, V., Vermeulen, J., Fitzmaurice, G., Matejka, J.: 3dall-e: Integrating text-to-image ai in 3d design workflows. In: Proceedings of the 2023 ACM Designing Interactive Systems Conference. pp. 1955\u20131977 (2023)","key":"31_CR20","DOI":"10.1145\/3563657.3596098"},{"doi-asserted-by":"crossref","unstructured":"Liu, Z., Ning, J., Cao, Y., Wei, Y., Zhang, Z., Lin, S., Hu, H.: Video swin transformer. In: CVPR. pp. 3202\u20133211 (2022)","key":"31_CR21","DOI":"10.1109\/CVPR52688.2022.00320"},{"unstructured":"Moreno, P., Ho, P., Vasconcelos, N.: A kullback-leibler divergence based kernel for svm classification in multimedia applications. NeurIPS 16 (2003)","key":"31_CR22"},{"doi-asserted-by":"crossref","unstructured":"Peng, X., Wang, K., Zhu, Z., Wang, M., You, Y.: Crafting better contrastive views for siamese representation learning. In: CVPR. pp. 16031\u201316040 (2022)","key":"31_CR23","DOI":"10.1109\/CVPR52688.2022.01556"},{"unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: Pointnet++: Deep hierarchical feature learning on point sets in a metric space. NeurIPS 30 (2017)","key":"31_CR24"},{"doi-asserted-by":"crossref","unstructured":"Ren, D., Zheng, J., Cai, J., Li, J., Zhang, J.: Extrudenet: Unsupervised inverse sketch-and-extrude for shape parsing. In: ECCV. pp. 482\u2013498 (2022)","key":"31_CR25","DOI":"10.1007\/978-3-031-20086-1_28"},{"unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: A simple way to prevent neural networks from overfitting 15(1), 1929\u20131958 (jan 2014)","key":"31_CR26"},{"doi-asserted-by":"crossref","unstructured":"Uy, M.A., Chang, Y.Y., Sung, M., Goel, P., Lambourne, J.G., Birdal, T., Guibas, L.J.: Point2cyl: Reverse engineering 3d objects from point clouds to extrusion cylinders. In: CVPR. pp. 11850\u201311860 (2022)","key":"31_CR27","DOI":"10.1109\/CVPR52688.2022.01155"},{"unstructured":"Van Den\u00a0Oord, A., Vinyals, O., et\u00a0al.: Neural discrete representation learning. NeurIPS 30 (2017)","key":"31_CR28"},{"unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. NeurIPS 30 (2017)","key":"31_CR29"},{"unstructured":"Wang, T., Isola, P.: Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In: ICML. pp. 9929\u20139939 (2020)","key":"31_CR30"},{"doi-asserted-by":"crossref","unstructured":"Willis, K.D.D., Pu, Y., Luo, J., Chu, H., Du, T., Lambourne, J.G., Solar-Lezama, A., Matusik, W.: Fusion 360 gallery: A dataset and environment for programmatic cad construction from human design sequences. ACM Transactions on Graphics (TOG) 40(4), 1\u201324 (2021)","key":"31_CR31","DOI":"10.1145\/3450626.3459818"},{"doi-asserted-by":"crossref","unstructured":"Wu, R., Xiao, C., Zheng, C.: DeepCAD: A Deep Generative Network for Computer-Aided Design Models. In: ICCV. pp. 6752\u20136762 (2021)","key":"31_CR32","DOI":"10.1109\/ICCV48922.2021.00670"},{"doi-asserted-by":"crossref","unstructured":"Xie, Z., Lin, Y., Zhang, Z., Cao, Y., Lin, S., Hu, H.: Propagate yourself: Exploring pixel-level consistency for unsupervised visual representation learning. In: CVPR. pp. 16684\u201316693 (2021)","key":"31_CR33","DOI":"10.1109\/CVPR46437.2021.01641"},{"unstructured":"Xu, X., Willis, K.D., Lambourne, J.G., Cheng, C.Y., Jayaraman, P.K., Furukawa, Y.: Skexgen: Autoregressive generation of cad construction sequences with disentangled codebooks. In: International Conference on Machine Learning (2022)","key":"31_CR34"},{"doi-asserted-by":"crossref","unstructured":"Xu, X., Peng, W., Cheng, C.Y., Willis, K.D., Ritchie, D.: Inferring CAD Modeling Sequences Using Zone Graphs. In: CVPR. pp. 6058\u20136066 (2021)","key":"31_CR35","DOI":"10.1109\/CVPR46437.2021.00600"},{"doi-asserted-by":"crossref","unstructured":"Yu, F., Chen, Z., Li, M., Sanghi, A., Shayani, H., Mahdavi-Amiri, A., Zhang, H.: Capri-net: Learning compact cad shapes with adaptive primitive assembly. In: CVPR. pp. 11768\u201311778 (2022)","key":"31_CR36","DOI":"10.1109\/CVPR52688.2022.01147"},{"unstructured":"Yu, J., Wang, Z., Vasudevan, V., Yeung, L., Seyedhosseini, M., Wu, Y.: Coca: Contrastive captioners are image-text foundation models. TMLR (2022)","key":"31_CR37"},{"doi-asserted-by":"crossref","unstructured":"Zellers, R., Lu, J., Lu, X., Yu, Y., Zhao, Y., Salehi, M., Kusupati, A., Hessel, J., Farhadi, A., Choi, Y.: Merlot reserve: Neural script knowledge through vision and language and sound. In: CVPR. pp. 16375\u201316387 (2022)","key":"31_CR38","DOI":"10.1109\/CVPR52688.2022.01589"},{"doi-asserted-by":"crossref","unstructured":"Zhang, J., Ma, K.: Rethinking the augmentation module in contrastive learning: Learning hierarchical augmentation invariance with expanded views. In: CVPR. pp. 16650\u201316659 (2022)","key":"31_CR39","DOI":"10.1109\/CVPR52688.2022.01615"}],"container-title":["Lecture Notes in Computer Science","Database Systems for Advanced Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-5555-4_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,11]],"date-time":"2025-01-11T06:08:12Z","timestamp":1736575692000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-5555-4_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819755547","9789819755554"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-5555-4_31","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 January 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DASFAA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Database Systems for Advanced Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Gifu","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 July 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 July 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dasfaa2024a","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.dasfaa2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}