{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:38:03Z","timestamp":1742913483130,"version":"3.40.3"},"publisher-location":"Cham","reference-count":15,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030379681"},{"type":"electronic","value":"9783030379698"}],"license":[{"start":{"date-parts":[[2019,12,20]],"date-time":"2019-12-20T00:00:00Z","timestamp":1576800000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2019,12,20]],"date-time":"2019-12-20T00:00:00Z","timestamp":1576800000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-37969-8_5","type":"book-chapter","created":{"date-parts":[[2019,12,19]],"date-time":"2019-12-19T09:07:51Z","timestamp":1576746471000},"page":"35-42","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Feature Pyramid Based Attention for Cervical Image Classification"],"prefix":"10.1007","author":[{"given":"Hongfeng","family":"Li","sequence":"first","affiliation":[]},{"given":"Jian","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Li","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Li","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Quanzheng","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,12,20]]},"reference":[{"key":"5_CR1","doi-asserted-by":"crossref","unstructured":"Anavi, Y., Kogan, I., Gelbart, E., Geva, O., Greenspan, H.: Visualizing and enhancing a deep learning framework using patients age and gender for chest x-ray image retrieval. In: Medical Imaging 2016: Computer-Aided Diagnosis, vol. 9785, p. 978510. International Society for Optics and Photonics (2016)","DOI":"10.1117\/12.2217587"},{"issue":"10071","key":"5_CR2","doi-asserted-by":"publisher","first-page":"847","DOI":"10.1016\/S0140-6736(16)31392-7","volume":"389","author":"O Ginsburg","year":"2017","unstructured":"Ginsburg, O., et al.: The global burden of women\u2019s cancers: a grand challenge in global health. Lancet 389(10071), 847\u2013860 (2017)","journal-title":"Lancet"},{"issue":"3","key":"5_CR3","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1109\/TMI.2008.2007823","volume":"28","author":"H Greenspan","year":"2009","unstructured":"Greenspan, H., et al.: Automatic detection of anatomical landmarks in uterine cervix images. IEEE Trans. Med. Imaging 28(3), 454\u2013468 (2009)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"5_CR4","unstructured":"Jetley, S., Lord, N.A., Lee, N., Torr, P.H.: Learn to pay attention. arXiv preprint arXiv:1804.02391 (2018)"},{"key":"5_CR5","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1016\/j.media.2016.10.004","volume":"36","author":"K Kamnitsas","year":"2017","unstructured":"Kamnitsas, K., et al.: Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 36, 61\u201378 (2017)","journal-title":"Med. Image Anal."},{"key":"5_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-94-007-5389-1_1","volume-title":"Color Medical Image analysis","author":"E Kim","year":"2013","unstructured":"Kim, E., Huang, X.: A data driven approach to cervigram image analysis and classification. In: Celebi, M., Schaefer, G. (eds.) Color Medical Image analysis, pp. 1\u201313. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-94-007-5389-1_1"},{"key":"5_CR7","unstructured":"Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images. Technical report, Citeseer (2009)"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"5_CR9","unstructured":"Liu, Y., et al.: Detecting cancer metastases on gigapixel pathology images. arXiv preprint arXiv:1703.02442 (2017)"},{"issue":"3","key":"5_CR10","first-page":"3518","volume":"15","author":"M Sato","year":"2018","unstructured":"Sato, M., et al.: Application of deep learning to the classification of images from colposcopy. Oncol. Lett. 15(3), 3518\u20133523 (2018)","journal-title":"Oncol. Lett."},{"key":"5_CR11","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"5_CR12","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"5_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1007\/978-3-319-46723-8_14","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2016","author":"T Xu","year":"2016","unstructured":"Xu, T., Zhang, H., Huang, X., Zhang, S., Metaxas, D.N.: Multimodal deep learning for cervical dysplasia diagnosis. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.) MICCAI 2016. LNCS, vol. 9901, pp. 115\u2013123. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46723-8_14"},{"issue":"3","key":"5_CR14","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1109\/MCI.2018.2840738","volume":"13","author":"T Young","year":"2018","unstructured":"Young, T., Hazarika, D., Poria, S., Cambria, E.: Recent trends in deep learning based natural language processing. IEEE Comput. Intell. Mag. 13(3), 55\u201375 (2018)","journal-title":"IEEE Comput. Intell. Mag."},{"key":"5_CR15","doi-asserted-by":"crossref","unstructured":"Zhang, J., Xie, Y., Xia, Y., Shen, C.: Attention residual learning for skin lesion classification. IEEE Trans. Med. Imaging (2019)","DOI":"10.1109\/TMI.2019.2893944"}],"container-title":["Lecture Notes in Computer Science","Multiscale Multimodal Medical Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-37969-8_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:02:54Z","timestamp":1734566574000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-37969-8_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,12,20]]},"ISBN":["9783030379681","9783030379698"],"references-count":15,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-37969-8_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2019,12,20]]},"assertion":[{"value":"20 December 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMMI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Multiscale Multimodal Medical Imaging","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenzhen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmmi2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/mmmi2019.github.io","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"18","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"13","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"72% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.6","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1.6","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}