{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T11:32:07Z","timestamp":1769945527981,"version":"3.49.0"},"publisher-location":"Cham","reference-count":36,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198175","type":"print"},{"value":"9783031198182","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19818-2_34","type":"book-chapter","created":{"date-parts":[[2022,10,21]],"date-time":"2022-10-21T16:21:10Z","timestamp":1666369270000},"page":"596-611","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Point MixSwap: Attentional Point Cloud Mixing via\u00a0Swapping Matched Structural Divisions"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7819-1941","authenticated-orcid":false,"given":"Ardian","family":"Umam","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7355-5467","authenticated-orcid":false,"given":"Cheng-Kun","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1383-0017","authenticated-orcid":false,"given":"Yung-Yu","family":"Chuang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4934-4811","authenticated-orcid":false,"given":"Jen-Hui","family":"Chuang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7183-6070","authenticated-orcid":false,"given":"Yen-Yu","family":"Lin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,22]]},"reference":[{"key":"34_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision \u2013 ECCV 2020","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"34_CR2","doi-asserted-by":"crossref","unstructured":"Chen, N., et al.: Unsupervised learning of intrinsic structural representation points. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00914"},{"key":"34_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"330","DOI":"10.1007\/978-3-030-58580-8_20","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Chen","year":"2020","unstructured":"Chen, Y., et al.: PointMixup: augmentation for point clouds. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12348, pp. 330\u2013345. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58580-8_20"},{"key":"34_CR4","doi-asserted-by":"crossref","unstructured":"Choi, J., Song, Y., Kwak, N.: Part-aware data augmentation for 3D object detection in point cloud (2021)","DOI":"10.1109\/IROS51168.2021.9635887"},{"key":"34_CR5","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Mane, D., Vasudevan, V., Le, Q.V.: Autoaugment: learning augmentation strategies from data. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00020"},{"key":"34_CR6","unstructured":"Cubuk, E.D., Zoph, B., Shlens, J., Le, Q.: Randaugment: practical automated data augmentation with a reduced search space. In: NIPS"},{"key":"34_CR7","unstructured":"Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp. 249\u2013256. JMLR Workshop and Conference Proceedings (2010)"},{"key":"34_CR8","unstructured":"Kim, J.H., Choo, W., Song, H.O.: Puzzle mix: exploiting saliency and local statistics for optimal mixup. In: ICLR (2020)"},{"key":"34_CR9","doi-asserted-by":"crossref","unstructured":"Kim, S., Lee, S., Hwang, D., Lee, J., Hwang, S.J., Kim, H.J.: Point cloud augmentation with weighted local transformations. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00059"},{"key":"34_CR10","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks"},{"key":"34_CR11","doi-asserted-by":"crossref","unstructured":"Lee, D., et al.: Regularization strategy for point cloud via rigidly mixed sample. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01564"},{"key":"34_CR12","doi-asserted-by":"crossref","unstructured":"Li, R., Li, X., Heng, P.A., Fu, C.W.: Pointaugment: an auto-augmentation framework for point cloud classification. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00641"},{"key":"34_CR13","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)"},{"key":"34_CR14","doi-asserted-by":"crossref","unstructured":"Qi, C.R., Litany, O., He, K., Guibas, L.J.: Deep hough voting for 3D object detection in point clouds. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00937"},{"key":"34_CR15","unstructured":"Qi, C.R., Su, H., Mo, K., Guibas, L.J.: Pointnet: deep learning on point sets for 3D classification and segmentation. In: CVPR (2017)"},{"key":"34_CR16","unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: Pointnet++: deep hierarchical feature learning on point sets in a metric space. NIPS (2017)"},{"key":"34_CR17","doi-asserted-by":"crossref","unstructured":"Shi, S., et al.: PV-RCNN: point-voxel feature set abstraction for 3D object detection. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01054"},{"key":"34_CR18","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"34_CR19","doi-asserted-by":"crossref","unstructured":"Sixt, L., Wild, B., Landgraf, T.: RenderGAN: generating realistic labeled data. Frontiers in Robotics and AI (2018)","DOI":"10.3389\/frobt.2018.00066"},{"key":"34_CR20","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: ICCV (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"34_CR21","doi-asserted-by":"crossref","unstructured":"Uy, M.A., Pham, Q.H., Hua, B.S., Nguyen, D.T., Yeung, S.K.: Revisiting point cloud classification: a new benchmark dataset and classification model on real-world data. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00167"},{"key":"34_CR22","doi-asserted-by":"crossref","unstructured":"Uy, M.A., Pham, Q.H., Hua, B.S., Nguyen, T., Yeung, S.K.: Revisiting point cloud classification: A new benchmark dataset and classification model on real-world data. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 1588\u20131597 (2019)","DOI":"10.1109\/ICCV.2019.00167"},{"key":"34_CR23","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NIPS (2017)"},{"key":"34_CR24","unstructured":"Verma, V., et al.: Manifold Mixup: better representations by interpolating hidden states. In: ICML (2019)"},{"key":"34_CR25","doi-asserted-by":"crossref","unstructured":"Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph CNN for learning on point clouds. TOG (2019)","DOI":"10.1145\/3326362"},{"key":"34_CR26","unstructured":"Wu, Z., et al.: 3D shapeNets: a deep representation for volumetric shapes. In: CVPR (2015)"},{"key":"34_CR27","doi-asserted-by":"crossref","unstructured":"Xiang, T., Zhang, C., Song, Y., Yu, J., Cai, W.: Walk in the cloud: learning curves for point clouds shape analysis. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00095"},{"key":"34_CR28","doi-asserted-by":"crossref","unstructured":"Yang, C.K., Chuang, Y.Y., Lin, Y.Y.: Unsupervised point cloud object co-segmentation by co-contrastive learning and mutual attention sampling. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00724"},{"key":"34_CR29","doi-asserted-by":"crossref","unstructured":"Yang, C.K., Wu, J.J., Chen, K.S., Chuang, Y.Y., Lin, Y.Y.: An mil-derived transformer for weakly supervised point cloud segmentation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01153"},{"key":"34_CR30","doi-asserted-by":"crossref","unstructured":"Yoo, J., Ahn, N., Sohn, K.A.: Rethinking data augmentation for image super-resolution: a comprehensive analysis and a new strategy. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00840"},{"key":"34_CR31","doi-asserted-by":"crossref","unstructured":"Yun, S., Han, D., Oh, S.J., Chun, S., Choe, J., Yoo, Y.: CutMix: regularization strategy to train strong classifiers with localizable features. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00612"},{"key":"34_CR32","unstructured":"Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: Mixup: beyond empirical risk minimization. ICLR (2018)"},{"key":"34_CR33","doi-asserted-by":"crossref","unstructured":"Zhao, H., Jiang, L., Jia, J., Torr, P.H., Koltun, V.: Point transformer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01595"},{"key":"34_CR34","doi-asserted-by":"crossref","unstructured":"Zhu, C., Xu, K., Chaudhuri, S., Yi, L., Guibas, L.J., Zhang, H.: AdaCoSeg: adaptive shape co-segmentation with group consistency loss. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00857"},{"key":"34_CR35","doi-asserted-by":"crossref","unstructured":"Zhu, X., Liu, Y., Li, J., Wan, T., Qin, Z.: Emotion classification with data augmentation using generative adversarial networks. In: KDD (2018)","DOI":"10.1007\/978-3-319-93040-4_28"},{"key":"34_CR36","unstructured":"Zhu, Y., Aoun, M., Krijn, M., Vanschoren, J., Campus, H.T.: Data augmentation using conditional generative adversarial networks for leaf counting in arabidopsis plants. In: BMVC (2018)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19818-2_34","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T14:34:53Z","timestamp":1710340493000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19818-2_34"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198175","9783031198182"],"references-count":36,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19818-2_34","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"22 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}