{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T15:50:16Z","timestamp":1743090616703,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":30,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819985425"},{"type":"electronic","value":"9789819985432"}],"license":[{"start":{"date-parts":[[2023,12,29]],"date-time":"2023-12-29T00:00:00Z","timestamp":1703808000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,29]],"date-time":"2023-12-29T00:00:00Z","timestamp":1703808000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8543-2_38","type":"book-chapter","created":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T10:03:03Z","timestamp":1703757783000},"page":"473-484","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["SADD: Generative Adversarial Networks via\u00a0Self-attention and\u00a0Dual Discriminator in\u00a0Unsupervised Domain Adaptation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0838-6318","authenticated-orcid":false,"given":"Zaiyan","family":"Dai","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0971-6593","authenticated-orcid":false,"given":"Jun","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3664-2323","authenticated-orcid":false,"given":"Anfei","family":"Fan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9029-0378","authenticated-orcid":false,"given":"Jinyin","family":"Jia","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-4199-6147","authenticated-orcid":false,"given":"Junfan","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,29]]},"reference":[{"key":"38_CR1","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60, 84\u201390 (2012)","journal-title":"Commun. ACM"},{"key":"38_CR2","unstructured":"Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR, abs\/1409.1556 (2014)"},{"key":"38_CR3","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Ioffe, S., Vanhoucke, V., Alemi, A.A.: Inception-V4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv, abs\/1602.07261 (2016)","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"38_CR4","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2015)","DOI":"10.1109\/CVPR.2016.90"},{"key":"38_CR5","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1016\/j.neucom.2020.07.005","volume":"415","author":"Y Li","year":"2020","unstructured":"Li, Y., Lin, C., Li, H., et al.: Unsupervised domain adaptation with self-attention for post-disaster building damage detection. Neurocomputing 415, 27\u201339 (2020)","journal-title":"Neurocomputing"},{"key":"38_CR6","doi-asserted-by":"crossref","unstructured":"Sankaranarayanan, S., Balaji, Y., Castillo, C.D., et al.: Generate to adapt: aligning domains using generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8503\u20138512 (2018)","DOI":"10.1109\/CVPR.2018.00887"},{"key":"38_CR7","unstructured":"Long, M., Zhu, H., Wang, J., et al.: Deep transfer learning with joint adaptation networks. In: International Conference on Machine Learning, pp. 2208\u20132217. PMLR (2017)"},{"key":"38_CR8","doi-asserted-by":"crossref","unstructured":"Shen, J., Qu, Y., Zhang, W., et al.: Wasserstein distance guided representation learning for domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1 (2018)","DOI":"10.1609\/aaai.v32i1.11784"},{"key":"38_CR9","unstructured":"Tzeng, E., Hoffman, J., Zhang, N., et al.: Deep domain confusion: maximizing for domain invariance. arXiv preprint arXiv:1412.3474 (2014)"},{"key":"38_CR10","unstructured":"Long, M., Cao, Y., Wang, J., et al.: Learning transferable features with deep adaptation networks. In: International Conference on Machine Learning, pp. 97\u2013105. PMLR (2015)"},{"key":"38_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"443","DOI":"10.1007\/978-3-319-49409-8_35","volume-title":"Computer Vision \u2013 ECCV 2016 Workshops","author":"B Sun","year":"2016","unstructured":"Sun, B., Saenko, K.: Deep CORAL: correlation alignment for deep domain adaptation. In: Hua, G., J\u00e9gou, H. (eds.) ECCV 2016. LNCS, vol. 9915, pp. 443\u2013450. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-49409-8_35"},{"key":"38_CR12","doi-asserted-by":"publisher","first-page":"504","DOI":"10.1016\/j.isatra.2021.12.037","volume":"129","author":"P Chen","year":"2022","unstructured":"Chen, P., Zhao, R., He, T., Wei, K., Qidong, Y.: Unsupervised domain adaptation of bearing fault diagnosis based on join sliced Wasserstein distance. ISA Trans. 129, 504\u2013519 (2022)","journal-title":"ISA Trans."},{"key":"38_CR13","unstructured":"Nguyen, A., Tran, T., Gal, Y., Torr, P.H., Baydin, A.G.: KL Guided Domain Adaptation. arXiv, abs\/2106.07780 (2021)"},{"key":"38_CR14","unstructured":"Ganin, Y., Ustinova, E., Ajakan, H., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(1), 2096-2030 (2016)"},{"key":"38_CR15","doi-asserted-by":"crossref","unstructured":"Tzeng, E., Hoffman, J., Saenko, K., et al.: Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7167\u20137176 2017","DOI":"10.1109\/CVPR.2017.316"},{"key":"38_CR16","doi-asserted-by":"crossref","unstructured":"Lee, J., Hwang, K., Kwak, M., et al.: Domain adaptation training of a transformer. In: 2022 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), pp. 1\u20135. IEEE (2022)","DOI":"10.1109\/ICCE-Asia57006.2022.9954860"},{"key":"38_CR17","doi-asserted-by":"crossref","unstructured":"Zhang, J., Huang, J., Tian, Z., et al.: Spectral unsupervised domain adaptation for visual recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9829\u20139840 (2022)","DOI":"10.1109\/CVPR52688.2022.00960"},{"key":"38_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"597","DOI":"10.1007\/978-3-319-46493-0_36","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Ghifary","year":"2016","unstructured":"Ghifary, M., Kleijn, W.B., Zhang, M., Balduzzi, D., Li, W.: Deep reconstruction-classification networks for unsupervised domain adaptation. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 597\u2013613. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_36"},{"key":"38_CR19","doi-asserted-by":"crossref","unstructured":"Saito, K., Watanabe, K., Ushiku, Y., et al.: Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3723\u20133732 (2018)","DOI":"10.1109\/CVPR.2018.00392"},{"key":"38_CR20","doi-asserted-by":"crossref","unstructured":"Bousmalis, K., Silberman, N., Dohan, D., et al.: Unsupervised pixel-level domain adaptation with generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3722\u20133731 (2017)","DOI":"10.1109\/CVPR.2017.18"},{"key":"38_CR21","doi-asserted-by":"crossref","unstructured":"Tran, L., Sohn, K., Yu, X., Liu, X., Chandraker, M.: Gotta adapt \u2019em all: joint pixel and feature-level domain adaptation for recognition in the wild. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2667\u20132676 (2018)","DOI":"10.1109\/CVPR.2019.00278"},{"key":"38_CR22","unstructured":"Hoffman, J., et al.: CyCADA: cycle-consistent adversarial domain adaptation. In: International Conference on Machine Learning (2017)"},{"key":"38_CR23","doi-asserted-by":"crossref","unstructured":"Zhu, H., Yin, H., Xia, D., Wang, D., Liu, X., Zhu, S.: Joint pixel-level and feature-level unsupervised domain adaptation for surveillance face recognition. In: Chinese Conference on Pattern Recognition and Computer Vision (2022)","DOI":"10.1007\/978-3-031-18913-5_36"},{"key":"38_CR24","first-page":"1","volume":"19","author":"Z Chen","year":"2022","unstructured":"Chen, Z., Zhao, L., He, Q., Kuang, G.: Pixel-level and feature-level domain adaptation for heterogeneous SAR target recognition. IEEE Geosci. Remote Sens. Lett. 19, 1\u20135 (2022)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"38_CR25","unstructured":"Poojary, A., Phapale, A., Salpekar, R., Balpande, S.: Self-Attention Generative Adversarial Network: The Latest Advancement in GAN (2020)"},{"key":"38_CR26","unstructured":"Odena, A., Olah, C., Shlens, J.: Conditional Image Synthesis with Auxiliary Classifier GANs. Presented at the (2016)"},{"key":"38_CR27","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9992\u201310002 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"38_CR28","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"6","key":"38_CR29","first-page":"1190","volume":"48","author":"GG Wang","year":"2020","unstructured":"Wang, G.G., Guo, T., Yu, Y., Su, H.: Unsupervised domain adaptation classification model based on generative adversarial network. Acta Electonica Sinica 48(6), 1190 (2020)","journal-title":"Acta Electonica Sinica"},{"key":"38_CR30","first-page":"55","volume":"6","author":"D Poobathy","year":"2014","unstructured":"Poobathy, D., Chezian, R.M.: Edge detection operators: peak signal to noise ratio based comparison. Int. J. Image Graph. Sig. Process. 6, 55\u201361 (2014)","journal-title":"Int. J. Image Graph. Sig. Process."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8543-2_38","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T10:11:58Z","timestamp":1703758318000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8543-2_38"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,29]]},"ISBN":["9789819985425","9789819985432"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8543-2_38","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,29]]},"assertion":[{"value":"29 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}