{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T00:38:06Z","timestamp":1760143086348,"version":"build-2065373602"},"reference-count":31,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T00:00:00Z","timestamp":1705881600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"\u201cPioneer\u201d R&amp;D Program of Zhejiang","award":["2023C03011","#U23A2037","#U22B2012","#42176177","#LDT23D06021D06"],"award-info":[{"award-number":["2023C03011","#U23A2037","#U22B2012","#42176177","#LDT23D06021D06"]}]},{"name":"National Natural Science Foundation of China","award":["2023C03011","#U23A2037","#U22B2012","#42176177","#LDT23D06021D06"],"award-info":[{"award-number":["2023C03011","#U23A2037","#U22B2012","#42176177","#LDT23D06021D06"]}]},{"name":"Zhejiang Provincial Natural Science Foundation of China","award":["2023C03011","#U23A2037","#U22B2012","#42176177","#LDT23D06021D06"],"award-info":[{"award-number":["2023C03011","#U23A2037","#U22B2012","#42176177","#LDT23D06021D06"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>In this study, we developed an innovative and self-supervised pretraining approach using Sentinel-2\/MSI satellite imagery specifically designed for the intelligent identification of drainage at sea discharge outlets. By integrating the geographical information from remote sensing images into our proposed methodology, we surpassed the classification accuracy of conventional models, such as MoCo (momentum contrast) and BYOL (bootstrap your own latent). Using Sentinel-2\/MSI remote sensing imagery, we developed our model through an unsupervised dataset comprising 25,600 images. The model was further refined using a supervised dataset composed of 1100 images. After supervised fine-tuning, the resulting framework yielded an adept model that was capable of classifying outfall drainage with an accuracy rate of 90.54%, facilitating extensive outfall monitoring. A series of ablation experiments affirmed the effectiveness of our enhancement of the training framework, showing a 10.81% improvement in accuracy compared to traditional models. Furthermore, the authenticity of the learned features was further validated using visualization techniques. This study contributes an efficient approach to large-scale monitoring of coastal outfalls, with implications for augmenting environmental protection measures and reducing manual inspection efforts.<\/jats:p>","DOI":"10.3390\/rs16020423","type":"journal-article","created":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T06:49:31Z","timestamp":1705906171000},"page":"423","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Intelligent Recognition of Coastal Outfall Drainage Based on Sentinel-2\/MSI Imagery"],"prefix":"10.3390","volume":"16","author":[{"given":"Hongzhe","family":"Li","sequence":"first","affiliation":[{"name":"Southern Marine Science and Engineering Guangdong Laboratory (Guangzhou), Guangzhou 511458, China"},{"name":"School of Oceanography, Shanghai Jiao Tong University, Shanghai 201100, China"}]},{"given":"Xianqiang","family":"He","sequence":"additional","affiliation":[{"name":"School of Oceanography, Shanghai Jiao Tong University, Shanghai 201100, China"},{"name":"State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China"},{"name":"Donghai Laboratory, Zhoushan 316021, China"}]},{"given":"Yan","family":"Bai","sequence":"additional","affiliation":[{"name":"School of Oceanography, Shanghai Jiao Tong University, Shanghai 201100, China"},{"name":"State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China"}]},{"given":"Fang","family":"Gong","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China"}]},{"given":"Teng","family":"Li","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7747-3082","authenticated-orcid":false,"given":"Difeng","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Satellite Ocean Environment Dynamics, Second Institute of Oceanography, Ministry of Natural Resources, Hangzhou 310012, China"}]}],"member":"1968","published-online":{"date-parts":[[2024,1,22]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"358","DOI":"10.1016\/j.envpol.2016.07.011","article-title":"Industrial water pollution, water environment treatment, and health risks in China","volume":"218","author":"Wang","year":"2016","journal-title":"Environ. Pollut."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"127640","DOI":"10.1016\/j.jclepro.2021.127640","article-title":"Novel method for industrial sewage outfall detection: Water pollution monitoring based on web crawler and remote sensing interpretation techniques","volume":"312","author":"Zhang","year":"2021","journal-title":"J. Clean. Prod."},{"key":"ref_3","first-page":"1","article-title":"An Improved Deep Learning Approach for Retrieving Outfalls Into Rivers From UAS Imagery","volume":"60","author":"Huang","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Xu, H., Huang, Q., Yang, Y., Li, J., Chen, X., Han, W., and Wang, L. (2022, January 28\u201330). UAV-ODS: A Real-time Outfall Detection System Based on UAV Remote Sensing and Edge Computing. Proceedings of the 2022 IEEE International Conference on Unmanned Systems (ICUS), Guangzhou, China.","DOI":"10.1109\/ICUS55513.2022.9987107"},{"key":"ref_5","unstructured":"Ballasiotes, A.D. (2020). Mapping Untreated and Semi-Treated Wastewater Effluent off the Coast of Gaza with Sentinel-1 Time Series Data. [Master\u2019s Thesis, Oregon State University]."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"158374","DOI":"10.1016\/j.scitotenv.2022.158374","article-title":"Automatic detection of suspected sewage discharge from coastal outfalls based on Sentinel-2 imagery","volume":"853","author":"Wang","year":"2022","journal-title":"Sci. Total Environ."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"979","DOI":"10.1134\/S0001433820090066","article-title":"Registering from space the features of deep wastewater outfalls into coastal water areas due to discharge collector breaks","volume":"56","author":"Bondur","year":"2020","journal-title":"Izv. Atmos. Ocean. Phys."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"111716","DOI":"10.1016\/j.rse.2020.111716","article-title":"Deep learning in environmental remote sensing: Achievements and challenges","volume":"241","author":"Yuan","year":"2020","journal-title":"Remote Sens. Environ."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"539","DOI":"10.1016\/j.patcog.2016.07.001","article-title":"Towards better exploiting convolutional neural networks for remote sensing scene classification","volume":"61","author":"Nogueira","year":"2017","journal-title":"Pattern Recognit."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"14078","DOI":"10.1109\/ACCESS.2021.3051085","article-title":"Classification of remote sensing images using EfficientNet-B3 CNN model with attention","volume":"9","author":"Alhichri","year":"2021","journal-title":"IEEE Access"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Berg, P., Pham, M.-T., and Courty, N. (2022). Self-Supervised Learning for Scene Classification in Remote Sensing: Current State of the Art and Perspectives. Remote Sens., 14.","DOI":"10.3390\/rs14163995"},{"key":"ref_12","unstructured":"Goyal, P., Caron, M., Lefaudeux, B., Xu, M., Wang, P., Pai, V., Singh, M., Liptchinsky, V., Misra, I., and Joulin, A. (2021). Self-supervised pretraining of visual features in the wild. arXiv."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., and Girshick, R. (2020, January 13\u201319). Momentum contrast for unsupervised visual representation learning. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Chen, X., and He, K. (2021, January 19\u201325). Exploring simple siamese representation learning. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01549"},{"key":"ref_15","first-page":"21271","article-title":"Bootstrap your own latent-a new approach to self-supervised learning","volume":"33","author":"Grill","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Wang, Y., Albrecht, C.M., Braham, N.A.A., Mou, L., and Zhu, X.X. (2022). Self-supervised learning in remote sensing: A review. arXiv.","DOI":"10.1109\/MGRS.2022.3198244"},{"key":"ref_17","unstructured":"Mai, G., Lao, N., He, Y., Song, J., and Ermon, S. (2023). CSP: Self-Supervised Contrastive Spatial Pre-Training for Geospatial-Visual Representations. arXiv."},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Ayush, K., Uzkent, B., Meng, C., Tanmay, K., Burke, M., Lobell, D., and Ermon, S. (2021, January 19\u201325). Geography-aware self-supervised learning. Proceedings of the IEEE\/CVF International Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/ICCV48922.2021.01002"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Phiri, D., Simwanda, M., Salekin, S., Nyirenda, V.R., Murayama, Y., and Ranagalage, M. (2020). Sentinel-2 data for land cover\/use mapping: A review. Remote Sens., 12.","DOI":"10.3390\/rs12142291"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"164862","DOI":"10.1016\/j.scitotenv.2023.164862","article-title":"Monitoring and spatial traceability of river water quality using Sentinel-2 satellite images","volume":"894","author":"Zhang","year":"2023","journal-title":"Sci. Total Environ."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"8743","DOI":"10.1038\/s41598-020-65600-1","article-title":"New capabilities of Sentinel-2A\/B satellites combined with in situ data for monitoring small harmful algal blooms in complex coastal waters","volume":"10","author":"Caballero","year":"2020","journal-title":"Sci. Rep."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Hafeez, S., Wong, M.S., Abbas, S., and Asim, M. (2022). Evaluating landsat-8 and sentinel-2 data consistency for high spatiotemporal inland and coastal water quality monitoring. Remote Sens., 14.","DOI":"10.3390\/rs14133155"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Gao, T., Yao, X., and Chen, D. (2021). Simcse: Simple contrastive learning of sentence embeddings. arXiv.","DOI":"10.18653\/v1\/2021.emnlp-main.552"},{"key":"ref_24","unstructured":"Oord, A.v.d., Li, Y., and Vinyals, O. (2018). Representation learning with contrastive predictive coding. arXiv."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., and Balasubramanian, V.N. (2018, January 12\u201315). Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA.","DOI":"10.1109\/WACV.2018.00097"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017, January 22\u201329). Grad-cam: Visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref_27","unstructured":"Springenberg, J.T., Dosovitskiy, A., Brox, T., and Riedmiller, M. (2014). Striving for simplicity: The all convolutional net. arXiv."},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_29","unstructured":"Chen, X., Fan, H., Girshick, R., and He, K. (2020). Improved baselines with momentum contrastive learning. arXiv."},{"key":"ref_30","unstructured":"Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. (2020, January 12\u201318). A simple framework for contrastive learning of visual representations. Proceedings of the International Conference on Machine Learning, Online."},{"key":"ref_31","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An image is worth 16\u00d716 words: Transformers for image recognition at scale. arXiv."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/2\/423\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T13:46:56Z","timestamp":1760104016000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/2\/423"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,22]]},"references-count":31,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2024,1]]}},"alternative-id":["rs16020423"],"URL":"https:\/\/doi.org\/10.3390\/rs16020423","relation":{},"ISSN":["2072-4292"],"issn-type":[{"type":"electronic","value":"2072-4292"}],"subject":[],"published":{"date-parts":[[2024,1,22]]}}}