{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T07:34:48Z","timestamp":1743060888806,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":20,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819703753"},{"type":"electronic","value":"9789819703760"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-0376-0_9","type":"book-chapter","created":{"date-parts":[[2024,2,11]],"date-time":"2024-02-11T09:02:09Z","timestamp":1707642129000},"page":"110-123","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MobileNet-SA: Lightweight CNN with\u00a0Self Attention for\u00a0Sketch Classification"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8537-1331","authenticated-orcid":false,"given":"Viet-Tham","family":"Huynh","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7729-2927","authenticated-orcid":false,"given":"Trong-Thuan","family":"Nguyen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0236-7992","authenticated-orcid":false,"given":"Tam V.","family":"Nguyen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3046-3041","authenticated-orcid":false,"given":"Minh-Triet","family":"Tran","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,12]]},"reference":[{"key":"9_CR1","doi-asserted-by":"crossref","unstructured":"Bhunia, A.K., et al.: Sketch2Saliency: learning to detect salient objects from human drawings. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00268"},{"key":"9_CR2","doi-asserted-by":"crossref","unstructured":"Chowdhury, P.N., et al.: What can human sketches do for object detection? In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01448"},{"key":"9_CR3","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Minneapolis, Minnesota: Association for Computational Linguistics, June 2019, pp. 4171\u20134186 (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1423. https:\/\/aclanthology.org\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Gao, C., et al.: SketchyCOCO: image generation from freehand scene sketches (2020). arXiv: 2003.02683 [cs.CV]","DOI":"10.1109\/CVPR42600.2020.00522"},{"key":"9_CR5","doi-asserted-by":"crossref","unstructured":"Huynh, V.-T., Nguyen, and T.V., Tran, M.-T.: Light-weight sketch recognition with knowledge distillation. In: International Conference on Multimedia Analysis and Pattern Recognition (2023)","DOI":"10.1109\/MAPR59823.2023.10289002"},{"key":"9_CR6","doi-asserted-by":"publisher","first-page":"10859","DOI":"10.1007\/s11042-020-10283-z","volume":"80","author":"H Kim","year":"2021","unstructured":"Kim, H., Yeo, C., Cha, M., Mun, D.: A method of generating depth images for view-based shape retrieval of 3D CAD models from partial point clouds. Multimedia Tools Appl. 80, 10859\u201310880 (2021)","journal-title":"Multimedia Tools Appl."},{"key":"9_CR7","doi-asserted-by":"crossref","unstructured":"Le, T.-N., et al.: SketchANIMAR: sketch-based 3D animal fine- grained retrieval. Comput. Graphics (2023)","DOI":"10.1016\/j.cag.2023.07.035"},{"key":"9_CR8","unstructured":"Le, T.-N., et al.: TextANIMAR: text-based 3D animal fine-grained retrieval. Comput. Graph. (2023)"},{"key":"9_CR9","doi-asserted-by":"publisher","first-page":"157322","DOI":"10.1109\/ACCESS.2020.3018696","volume":"8","author":"Z Li","year":"2020","unstructured":"Li, Z., Junyu, X., Zhao, Y., Li, W., Nie, W.: MPAN: multipart attention network for point cloud based 3D shape retrieval. IEEE Access 8, 157322\u2013157332 (2020)","journal-title":"IEEE Access"},{"key":"9_CR10","doi-asserted-by":"crossref","unstructured":"Lin, H., Fu, Y., Jiang, Y.-G., Xue, X.: Sketch- BERT: learning sketch bidirectional encoder representation from transformers by self-supervised learning of sketch gestalt (2020). arXiv: 2005.09159 [cs.CV]","DOI":"10.1109\/CVPR42600.2020.00679"},{"key":"9_CR11","unstructured":"Liu, X., et al.: Self-supervised learning: generative or contrastive. en. In: arXiv:2006.08218 [cs, stat] (July 2020). arXiv:2006.08218. http:\/\/arxiv.org\/abs\/2006.08218 (visited on 10\/30\/2020)"},{"key":"9_CR12","unstructured":"O\u2019Shea, K., Nash, R.: An introduction to convolutional neural networks (2015). arXiv: 1511.08458 [cs.NE]"},{"key":"9_CR13","unstructured":"Ribeiro, L.S.F., Bui, T., Collomosse, J., Ponti, M.: Sketchformer: transformer-based Representation for Sketched Structure. In: Proceedings of the CVPR (2020)"},{"key":"9_CR14","doi-asserted-by":"crossref","unstructured":"Sain, A., et al.: Exploiting Unlabelled Photos for Stronger Fine-Grained SBIR (2023). arXiv: 2303.13779 [cs.CV]","DOI":"10.1109\/CVPR52729.2023.00664"},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: MobileNetV2: inverted residuals and linear bottlenecks (2019). arXiv: 1801.04381 [cs.CV]","DOI":"10.1109\/CVPR.2018.00474"},{"key":"9_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1007\/978-3-030-58539-6_32","volume-title":"Computer Vision \u2013 ECCV 2020","author":"A Tripathi","year":"2020","unstructured":"Tripathi, A., Dani, R.R., Mishra, A., Chakraborty, A.: Sketch-guided object localization in\u00a0natural images. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12351, pp. 532\u2013547. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58539-6_32"},{"key":"9_CR17","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"9_CR18","doi-asserted-by":"crossref","unstructured":"Xie, X., et al.: Artistic style discovery with independent components. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 19870\u201319879 (2022)","DOI":"10.1109\/CVPR52688.2022.01925"},{"key":"9_CR19","doi-asserted-by":"publisher","unstructured":"Yang, L., Pang, K., Zhang, H., Song, Y.-Z.: SketchAA: abstract representation for abstract sketches. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 10077\u201310086 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00994","DOI":"10.1109\/ICCV48922.2021.00994"},{"key":"9_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848\u20136856 (2018)","DOI":"10.1109\/CVPR.2018.00716"}],"container-title":["Lecture Notes in Computer Science","Image and Video Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-0376-0_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,11]],"date-time":"2024-02-11T09:03:10Z","timestamp":1707642190000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-0376-0_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819703753","9789819703760"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-0376-0_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"12 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PSIVT","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific-Rim Symposium on Image and Video Technology","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"psivt2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/psivt2023.aut.ac.nz\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"75","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}