{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T19:22:23Z","timestamp":1768591343388,"version":"3.49.0"},"reference-count":41,"publisher":"MDPI AG","issue":"21","license":[{"start":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T00:00:00Z","timestamp":1729900800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100000038","name":"Natural Sciences and Engineering Research Council (NSERC) of Canada","doi-asserted-by":"publisher","award":["RGPIN-2018-04046"],"award-info":[{"award-number":["RGPIN-2018-04046"]}],"id":[{"id":"10.13039\/501100000038","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>SegContrast first paved the way for contrastive learning on outdoor point clouds. Its original formulation targeted individual scans in applications like autonomous driving and object detection. However, mobile mapping purposes such as digital twin cities and urban planning require large-scale dense datasets to capture the full complexity and diversity present in outdoor environments. In this paper, the SegContrast method is revisited and adapted to overcome its limitations associated with mobile mapping datasets, namely the scarcity of contrastive pairs and memory constraints. To overcome the scarcity of contrastive pairs, we propose the merging of heterogeneous datasets. However, this merging is not a straightforward procedure due to the variety of size and number of points in the point clouds of these datasets. Therefore, a data augmentation approach is designed to create a vast number of segments while optimizing the size of the point cloud samples to the allocated memory. This methodology, called CLOUDSPAM, guarantees the performance of the self-supervised model for both small- and large-scale mobile mapping point clouds. Overall, the results demonstrate the benefits of utilizing datasets with a wide range of densities and class diversity. CLOUDSPAM matched the state of the art on the KITTI-360 dataset, with a 63.6% mIoU, and came in second place on the Toronto-3D dataset. Finally, CLOUDSPAM achieved competitive results against its fully supervised counterpart with only 10% of labeled data.<\/jats:p>","DOI":"10.3390\/rs16213984","type":"journal-article","created":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T07:04:04Z","timestamp":1730099044000},"page":"3984","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["CLOUDSPAM: Contrastive Learning On Unlabeled Data for Segmentation and Pre-Training Using Aggregated Point Clouds and MoCo"],"prefix":"10.3390","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-6539-8749","authenticated-orcid":false,"given":"Reza","family":"Mahmoudi Kouhi","sequence":"first","affiliation":[{"name":"Department of Geomatics Sciences, Universit\u00e9 Laval, Qu\u00e9bec City, QC G1V 0A6, Canada"}]},{"given":"Olivier","family":"Stocker","sequence":"additional","affiliation":[{"name":"Department of Geomatics Sciences, Universit\u00e9 Laval, Qu\u00e9bec City, QC G1V 0A6, Canada"}]},{"given":"Philippe","family":"Gigu\u00e8re","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Software Engineering, Universit\u00e9 Laval, Qu\u00e9bec City, QC G1V 0A6, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2383-9442","authenticated-orcid":false,"given":"Sylvie","family":"Daniel","sequence":"additional","affiliation":[{"name":"Department of Geomatics Sciences, Universit\u00e9 Laval, Qu\u00e9bec City, QC G1V 0A6, Canada"}]}],"member":"1968","published-online":{"date-parts":[[2024,10,26]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Griffiths, D., and Boehm, J. (2019). A Review on Deep Learning Techniques for 3D Sensed Data Classification. Remote Sens., 11.","DOI":"10.3390\/rs11121499"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"4338","DOI":"10.1109\/TPAMI.2020.3005434","article-title":"Deep learning for 3d point clouds: A survey","volume":"43","author":"Guo","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"38","DOI":"10.1109\/MGRS.2019.2937630","article-title":"Linking points with labels in 3D: A review of point cloud semantic segmentation","volume":"8","author":"Xie","year":"2020","journal-title":"IEEE Geosci. Remote Sens. Mag."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"11321","DOI":"10.1109\/TPAMI.2023.3262786","article-title":"Unsupervised point cloud representation learning with deep neural networks: A survey","volume":"45","author":"Xiao","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_5","unstructured":"Arora, S., Khandeparkar, H., Khodak, M., Plevrakis, O., and Saunshi, N. (2019). A theoretical analysis of contrastive unsupervised representation learning. arXiv."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Xiao, A., Zhang, X., Shao, L., and Lu, S. (2023). A Survey of Label-Efficient Deep Learning for 3D Point Clouds. arXiv.","DOI":"10.1109\/TPAMI.2024.3416302"},{"key":"ref_7","unstructured":"Behley, J., Garbade, M., Milioto, A., Quenzel, J., Behnke, S., Stachniss, C., and Gall, J. (November, January 27). SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"3292","DOI":"10.1109\/TPAMI.2022.3179507","article-title":"KITTI-360: A Novel Dataset and Benchmarks for Urban Scene Understanding in 2D and 3D","volume":"45","author":"Liao","year":"2022","journal-title":"Pattern Anal. Mach. Intell. (PAMI)"},{"key":"ref_9","unstructured":"Gui, J., Chen, T., Cao, Q., Sun, Z., Luo, H., and Tao, D. (2023). A Survey of Self-Supervised Learning from Multiple Perspectives: Algorithms, Theory, Applications and Future Trends. arXiv."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Hou, J., Graham, B., Nie\u00dfner, M., and Xie, S. (2021). Exploring Data-Efficient 3D Scene Understanding with Contrastive Scene Contexts. arXiv.","DOI":"10.1109\/CVPR46437.2021.01533"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., and Girshick, R. (2020). Momentum Contrast for Unsupervised Visual Representation Learning. arXiv.","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Xie, S., Gu, J., Guo, D., Qi, C.R., Guibas, L.J., and Litany, O. (2020, January 23\u201328). PointContrast: Unsupervised Pre-training for 3D Point Cloud Understanding. Proceedings of the IEEE\/CVF Europian Conference Computer Vision (ECCV), Glasgow, UK.","DOI":"10.1007\/978-3-030-58580-8_34"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"2116","DOI":"10.1109\/LRA.2022.3142440","article-title":"SegContrast: 3D Point Cloud Feature Representation Learning through Self-supervised Segment Discrimination","volume":"7","author":"Nunes","year":"2022","journal-title":"IEEE Robot. Autom. Lett. (RA-L)"},{"key":"ref_14","first-page":"21271","article-title":"Bootstrap your own latent-a new approach to self-supervised learning","volume":"33","author":"Grill","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Jiang, L., Shi, S., Tian, Z., Lai, X., Liu, S., Fu, C.W., and Jia, J. (2021, January 11\u201317). Guided point contrastive learning for semi-supervised point cloud semantic segmentation. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, BC, Canada.","DOI":"10.1109\/ICCV48922.2021.00636"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Li, L., Shum, H.P., and Breckon, T.P. (2023, January 17\u201324). Less is more: Reducing task and model complexity for 3d point cloud semantic segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","DOI":"10.1109\/CVPR52729.2023.00903"},{"key":"ref_17","unstructured":"Fei, B., Yang, W., Liu, L., Luo, T., Zhang, R., Li, Y., and He, Y. (2023). Self-supervised Learning for Pre-Training 3D Point Clouds: A Survey. arXiv."},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Tan, W., Qin, N., Ma, L., Li, Y., Du, J., Cai, G., Yang, K., and Li, J. (2020, January 14\u201319). Toronto-3D: A large-scale mobile LiDAR dataset for semantic segmentation of urban roadways. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA.","DOI":"10.1109\/CVPRW50498.2020.00109"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"545","DOI":"10.1177\/0278364918767506","article-title":"Paris-Lille-3D: A large and high-quality ground-truth urban point cloud dataset for automatic segmentation and classification","volume":"37","author":"Roynard","year":"2018","journal-title":"Int. J. Robot. Res."},{"key":"ref_20","unstructured":"van den Oord, A., Li, Y., and Vinyals, O. (2018). Representation Learning with Contrastive Predictive Coding. arXiv."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"509","DOI":"10.1109\/34.993558","article-title":"Shape Matching and Object Recognition Using Shape Contexts","volume":"24","author":"Belongie","year":"2002","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_22","unstructured":"K\u00f6rtgen, M., Park, G.J., Novotni, M., and Klein, R. (2003, January 7\u20139). 3D Shape Matching with 3D Shape Contexts. Proceedings of the 7th Central European Seminar on Computer Graphics, Vienna, Austria."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Xie, S., Liu, S., Chen, Z., and Tu, Z. (2018, January 18\u201323). Attentional ShapeContextNet for Point Cloud Recognition. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00484"},{"key":"ref_24","first-page":"381","article-title":"Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography","volume":"24","author":"Fischler","year":"1981","journal-title":"Assoc. Comput. Mach. Commun. ACM"},{"key":"ref_25","unstructured":"Ester, M., Kriegel, H.P., Sander, J., and Xu, X. (1996, January 2\u20134). A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Narksri, P., Takeuchi, E., Ninomiya, Y., Morales, Y., Akai, N., and Kawaguchi, N. (2018, January 4\u20137). A Slope-robust Cascaded Ground Segmentation in 3D Point Cloud for Autonomous Vehicles. Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA.","DOI":"10.1109\/ITSC.2018.8569534"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Mahmoudi Kouhi, R., Daniel, S., and Gigu\u00e8re, P. (2023). Data Preparation Impact on Semantic Segmentation of 3D Mobile LiDAR Point Clouds Using Deep Neural Networks. Remote Sens., 15.","DOI":"10.3390\/rs15040982"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Choy, C., Gwak, J., and Savarese, S. (2019, January 15\u201320). 4d Spatio-Temporal Convnets: Minkowski Convolutional Neural Networks. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00319"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Robert, D., Vallet, B., and Landrieu, L. (2022, January 19\u201320). Learning Multi-View Aggregation In the Wild for Large-Scale 3D Semantic Segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LO, USA.","DOI":"10.1109\/CVPR52688.2022.00549"},{"key":"ref_30","unstructured":"Qi, C.R., Yi, L., Su, H., and Guibas, L.J. (2017, January 4\u20139). Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA."},{"key":"ref_31","first-page":"1","article-title":"Dynamic graph cnn for learning on point clouds","volume":"38","author":"Wang","year":"2019","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"ref_32","unstructured":"Thomas, H., Qi, C.R., Deschaud, J.E., Marcotegui, B., Goulette, F., and Guibas, L.J. (November, January 27). Kpconv: Flexible and deformable convolution for point clouds. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Republic of Korea."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"821","DOI":"10.1109\/TITS.2019.2961060","article-title":"Multi-scale point-wise convolutional neural networks for 3D object segmentation from LiDAR point clouds in large-scale environments","volume":"22","author":"Ma","year":"2019","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"3588","DOI":"10.1109\/TGRS.2019.2958517","article-title":"TGNet: Geometric graph CNN on 3-D point cloud segmentation","volume":"58","author":"Li","year":"2019","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_35","first-page":"8338","article-title":"Learning semantic segmentation of large-scale point clouds with random sampling","volume":"44","author":"Hu","year":"2021","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Thomas, H., Goulette, F., Deschaud, J.E., Marcotegui, B., and LeGall, Y. (2018, January 5\u20138). Semantic classification of 3D point clouds with multiscale spherical neighborhoods. Proceedings of the 2018 International conference on 3D vision (3DV), Verona, Italy.","DOI":"10.1109\/3DV.2018.00052"},{"key":"ref_37","unstructured":"Roynard, X., Deschaud, J.E., and Goulette, F. (2018, January 1\u20135). Classification of point cloud for road scene understanding with multiscale voxel deep network. Proceedings of the 10th Workshop on Planning, Perceptionand Navigation for Intelligent Vehicules PPNIV\u20192018, Madrid, Spain."},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Liang, Z., Yang, M., Deng, L., Wang, C., and Wang, B. (2019, January 20\u201324). Hierarchical depthwise graph convolutional neural network for 3D semantic segmentation of point clouds. Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada.","DOI":"10.1109\/ICRA.2019.8794052"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"8301","DOI":"10.1109\/TGRS.2020.2985695","article-title":"MS-RRFSegNet: Multiscale regional relation feature segmentation network for semantic segmentation of urban scene point clouds","volume":"58","author":"Luo","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_40","first-page":"1","article-title":"Unstructured point cloud semantic labeling using deep segmentation networks","volume":"3","author":"Boulch","year":"2017","journal-title":"3dor@ Eurographics"},{"key":"ref_41","unstructured":"Boulch, A., Puy, G., and Marlet, R. (December, January 30). FKAConv: Feature-kernel alignment for point cloud convolution. Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/21\/3984\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T16:21:24Z","timestamp":1760113284000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/21\/3984"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,26]]},"references-count":41,"journal-issue":{"issue":"21","published-online":{"date-parts":[[2024,11]]}},"alternative-id":["rs16213984"],"URL":"https:\/\/doi.org\/10.3390\/rs16213984","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,26]]}}}