{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T01:49:39Z","timestamp":1760147379608,"version":"build-2065373602"},"reference-count":47,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2023,1,31]],"date-time":"2023-01-31T00:00:00Z","timestamp":1675123200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National NSF of China","award":["U19A2058","41971362","41871248","62106276"],"award-info":[{"award-number":["U19A2058","41971362","41871248","62106276"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Semantic segmentation of remote sensing (RS) images, which is a fundamental research topic, classifies each pixel in an image. It plays an essential role in many downstream RS areas, such as land-cover mapping, road extraction, traffic monitoring, and so on. Recently, although deep-learning-based methods have shown their dominance in automatic semantic segmentation of RS imagery, the performance of these existing methods has relied heavily on large amounts of high-quality training data, which are usually hard to obtain in practice. Moreover, human-in-the-loop semantic segmentation of RS imagery cannot be completely replaced by automatic segmentation models, since automatic models are prone to error in some complex scenarios. To address these issues, in this paper, we propose an improved, smart, and interactive segmentation model, DRE-Net, for RS images. The proposed model facilitates humans\u2019 performance of segmentation by simply clicking a mouse. Firstly, a dynamic radius-encoding (DRE) algorithm is designed to distinguish the purpose of each click, such as a click for the selection of a segmentation outline or for fine-tuning. Secondly, we propose an incremental training strategy to cause the proposed model not only to converge quickly, but also to obtain refined segmentation results. Finally, we conducted comprehensive experiments on the Potsdam and Vaihingen datasets and achieved 9.75% and 7.03% improvements in NoC95 compared to the state-of-the-art results, respectively. In addition, our DRE-Net can improve the convergence and generalization of a network with a fast inference speed.<\/jats:p>","DOI":"10.3390\/rs15030801","type":"journal-article","created":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T05:33:53Z","timestamp":1675229633000},"page":"801","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["DRE-Net: A Dynamic Radius-Encoding Neural Network with an Incremental Training Strategy for Interactive Segmentation of Remote Sensing Images"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7894-3170","authenticated-orcid":false,"given":"Liangzhe","family":"Yang","sequence":"first","affiliation":[{"name":"Department of Cognitive Communication, College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China"}]},{"given":"Wenjie","family":"Zi","sequence":"additional","affiliation":[{"name":"Department of Cognitive Communication, College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7880-3394","authenticated-orcid":false,"given":"Hao","family":"Chen","sequence":"additional","affiliation":[{"name":"Department of Cognitive Communication, College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China"}]},{"given":"Shuang","family":"Peng","sequence":"additional","affiliation":[{"name":"Department of Cognitive Communication, College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,1,31]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"3681","DOI":"10.1109\/TKDE.2020.3025580","article-title":"Deep Learning for Spatio-Temporal Data Mining: A Survey","volume":"34","author":"Wang","year":"2022","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Martinez, R.P., Schiopu, I., Cornelis, B., and Munteanu, A. (2021). Real-Time Instance Segmentation of Traffic Videos for Embedded Devices. Sensors, 21.","DOI":"10.3390\/s21010275"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Chen, H., and Peng, S. (2022). SW-GAN: Road Extraction from Remote Sensing Imagery Using Semi-Weakly Supervised Adversarial Learning. Remote Sens., 14.","DOI":"10.3390\/rs14174145"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"4673","DOI":"10.1109\/TGRS.2020.3016086","article-title":"Road Segmentation for Remote Sensing Images Using Adversarial Spatial Pyramid Networks","volume":"59","author":"Shamsolmoali","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"173","DOI":"10.1016\/j.rse.2018.11.014","article-title":"Joint deep learning for land cover and land use classification","volume":"221","author":"Zhang","year":"2019","journal-title":"Remote Sens. Environ."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"196","DOI":"10.1016\/j.isprsjprs.2022.06.008","article-title":"UNetFormer: An UNet-like Transformer for Efficient Semantic Segmentation of Remote Sensing Urban Scene Imagery","volume":"190","author":"Wang","year":"2022","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_7","first-page":"1","article-title":"A Novel Transformer based Semantic Segmentation Scheme for Fine-Resolution Remote Sensing Images","volume":"19","author":"Wang","year":"2022","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Xu, N., Price, B., Cohen, S., Yang, J., and Huang, T.S. (2016, January 27\u201330). Deep interactive object selection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.47"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"110","DOI":"10.1007\/s11390-017-1681-7","article-title":"Intelligent Visual Media Processing: When Graphics Meets Vision","volume":"32","author":"Cheng","year":"2017","journal-title":"J. Comput. Sci. Technol."},{"key":"ref_10","first-page":"1","article-title":"RepFinder: Finding approximately repeated scene elements for image editing","volume":"29","author":"Cheng","year":"2010","journal-title":"Int. Conf. Comput. Graph. Interact. Tech."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Lin, Z., Zhang, Z., Chen, L.Z., Cheng, M.M., and Lu, S.P. (2020, January 14\u201319). Interactive Image Segmentation with First Click Attention. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.01335"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Sofiiuk, K., Petrov, I.A., Barinova, O., and Konushin, A. (2020, January 14\u201319). F-BRS: Rethinking Backpropagating Refinement for Interactive Segmentation. Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00865"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Dupont, C., Ouakrim, Y., and Pham, Q.C. (2021, January 17\u201320). UCP-Net: Unstructured Contour Points for Instance Segmentation. Proceedings of the 2021 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Melbourne, Australia.","DOI":"10.1109\/SMC52423.2021.9658754"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"1559","DOI":"10.1109\/TPAMI.2018.2840695","article-title":"DeepIGeoS: A Deep Interactive Geodesic Framework for Medical Image Segmentation","volume":"41","author":"Wang","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"193","DOI":"10.1111\/cgf.12758","article-title":"DenseCut: Densely Connected CRFs for Realtime GrabCut","volume":"34","author":"Cheng","year":"2015","journal-title":"Comput. Graph. Forum"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"309","DOI":"10.1145\/1015706.1015720","article-title":"\u201cGrabCut\u201d interactive foreground extraction using iterated graph cuts","volume":"23","author":"Rother","year":"2004","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Bai, J., and Wu, X. (2014, January 23\u201328). Error-Tolerant Scribbles Based Interactive Image Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.57"},{"key":"ref_18","unstructured":"Freedman, D., and Zhang, T. (2005, January 20\u201325). Interactive graph cut based segmentation with shape priors. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), San Diego, CA, USA."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Maninis, K.K., Caelles, S., Pont-Tuset, J., and Gool, L.V. (2018, January 18\u201322). Deep Extreme Cut: From Extreme Points to Object Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00071"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Papadopoulos, D.P., Uijlings, J., Keller, F., and Ferrari, V. (2017, January 22\u201329). Extreme clicking for efficient object annotation. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.528"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Hao, Y., Liu, Y., Wu, Z., Han, L., Chen, Y., Chen, G., Chu, L., Tang, S., Yu, Z., and Chen, Z. (2021, January 11\u201317). EdgeFlow: Achieving Practical Interactive Segmentation with Edge-Guided Flow. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, QC, Canada.","DOI":"10.1109\/ICCVW54120.2021.00180"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Sofiiuk, K., Petrov, I.A., and Konushin, A. (2022, January 16\u201319). Reviving Iterative Training with Mask Guidance for Interactive Segmentation. Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France.","DOI":"10.1109\/ICIP46576.2022.9897365"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Jang, W.D., and Kim, C.S. (2019, January 15\u201320). Interactive image segmentation via backpropagating refinement scheme. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00544"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Chen, X., Zhao, Z., Zhang, Y., Duan, M., Qi, D., and Zhao, H. (2022, January 19\u201324). FocalClick: Towards Practical Interactive Image Segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.00136"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Liew, J.H., Wei, Y., Xiong, W., Ong, S.H., and Feng, J. (2017, January 22\u201329). Regional Interactive Image Segmentation Networks. Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.297"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Gleicher, M. (1995, January 15). Image snapping. Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques, New York, NY, USA.","DOI":"10.1145\/218380.218441"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"1768","DOI":"10.1109\/TPAMI.2006.233","article-title":"Random walks for image segmentation","volume":"28","author":"Grady","year":"2006","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"113","DOI":"10.1007\/s11263-008-0191-z","article-title":"Geodesic matting: A framework for fast interactive image and video segmentation and matting","volume":"82","author":"Bai","year":"2009","journal-title":"Int. J. Comput. Vis."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Mortensen, E.N., and Barrett, W.A. (1995, January 15). Intelligent scissors for image composition. Proceedings of the 22nd Annual conference on Computer Graphics and Interactive Techniques, New York, NY, USA.","DOI":"10.1145\/218380.218442"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T., and Berkeley, U. (2015, January 7\u201312). Fully Convolutional Networks for Semantic Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Lee, K.M., Myeong, H., and Song, G. (2018, January 18\u201322). SeedNet: Automatic Seed Generation with Deep Reinforcement Learning for Robust Interactive Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00189"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"31","DOI":"10.1016\/j.neunet.2018.10.009","article-title":"A Fully Convolutional Two-Stream Fusion Network for Interactive Image Segmentation","volume":"109","author":"Hu","year":"2019","journal-title":"Neural Netw. Off. J. Int. Neural Netw. Soc."},{"key":"ref_33","unstructured":"Forte, M., Price, B., Cohen, S., Xu, N., and Piti\u00e9, F. (2020). Getting to 99% accuracy in interactive segmentation. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"3521","DOI":"10.1073\/pnas.1611835114","article-title":"Overcoming catastrophic forgetting in neural networks","volume":"114","author":"Kirkpatrick","year":"2017","journal-title":"Proc. Natl. Acad. Sci. USA."},{"key":"ref_35","first-page":"647","article-title":"DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition","volume":"32","author":"Donahue","year":"2014","journal-title":"Int. Conf. Mach. Learn."},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Razavian, A.S., Azizpour, H., Sullivan, J., and Carlsson, S. (2014, January 23\u201328). CNN Features off-the-shelf: An Astounding Baseline for Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA.","DOI":"10.1109\/CVPRW.2014.131"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J., and Berkeley, U. (2014, January 23\u201328). Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.81"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"1790","DOI":"10.1109\/TPAMI.2015.2500224","article-title":"Factors of Transferability for a Generic ConvNet Representation","volume":"38","author":"Azizpour","year":"2016","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1023\/A:1007379606734","article-title":"Multitask learning","volume":"28","author":"Caruana","year":"1997","journal-title":"Mach. Learn."},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"149","DOI":"10.1007\/s10994-010-5231-6","article-title":"Boosted multi-task learning","volume":"85","author":"Chapelle","year":"2011","journal-title":"Mach. Learn."},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"2935","DOI":"10.1109\/TPAMI.2017.2773081","article-title":"Learning without Forgetting","volume":"40","author":"Li","year":"2018","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_42","unstructured":"Duan, X.C.Z.Y.Z. (2021, January 10\u201317). Conditional Diffusion for Interactive Segmentation. Proceedings of the 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Benenson, R., Popov, S., and Ferrari, V. (2019, January 15\u201320). Large-Scale Interactive Object Segmentation With Human Annotators. Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.01197"},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D., and Wang, J. (2019, January 15\u201320). Deep high-resolution representation learning for human pose estimation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00584"},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Zi, W., Xiong, W., Hao, C., Li, J., and Jing, N. (2021). SGA-Net: Self-Constructing Graph Attention Neural Network for Semantic Segmentation of Remote Sensing Images. Remote Sens., 13.","DOI":"10.3390\/rs13214201"},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018, January 8\u201314). Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"ref_47","unstructured":"Sofiiuk, K., Barinova, O., and Konushin, A. (November, January 27). AdaptIS: Adaptive Instance Selection Network. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/3\/801\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T18:19:37Z","timestamp":1760120377000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/3\/801"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,31]]},"references-count":47,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2023,2]]}},"alternative-id":["rs15030801"],"URL":"https:\/\/doi.org\/10.3390\/rs15030801","relation":{},"ISSN":["2072-4292"],"issn-type":[{"type":"electronic","value":"2072-4292"}],"subject":[],"published":{"date-parts":[[2023,1,31]]}}}