{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T02:51:13Z","timestamp":1775098273647,"version":"3.50.1"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031185229","type":"print"},{"value":"9783031185236","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-18523-6_17","type":"book-chapter","created":{"date-parts":[[2022,10,8]],"date-time":"2022-10-08T10:04:48Z","timestamp":1665223488000},"page":"178-189","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Verifiable and\u00a0Energy Efficient Medical Image Analysis with\u00a0Quantised Self-attentive Deep Neural Networks"],"prefix":"10.1007","author":[{"given":"Rakshith","family":"Sathish","sequence":"first","affiliation":[]},{"given":"Swanand","family":"Khare","sequence":"additional","affiliation":[]},{"given":"Debdoot","family":"Sheet","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,7]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Aji, A.F., Heafield, K.: Compressing neural machine translation models with 4-bit precision. In: Workshop on Neural Generation and Translation, pp. 35\u201342 (2020)","DOI":"10.18653\/v1\/2020.ngt-1.4"},{"key":"17_CR2","unstructured":"Antonelli, M., et al.: The medical segmentation decathlon. arXiv preprint arXiv:2106.05735 (2021)"},{"key":"17_CR3","unstructured":"Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)"},{"key":"17_CR4","unstructured":"Chakravarty, A., Ghosh, N., Sheet, D., Sarkar, T., Sethuraman, R.: Radiologist validated systematic search over deep neural networks for screening musculoskeletal radiographs (2019)"},{"key":"17_CR5","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A., et al.: UNETR: transformers for 3D medical image segmentation. In: IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 574\u2013584 (2022)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"17_CR7","doi-asserted-by":"crossref","unstructured":"Horowitz, M.: 1.1 computing\u2019s energy problem (and what we can do about it). In: IEEE International Solid-State Circuits Conference Digest of Technical Papers, pp. 10\u201314 (2014)","DOI":"10.1109\/ISSCC.2014.6757323"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"17_CR9","doi-asserted-by":"crossref","unstructured":"Jacob, B., et al.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"17_CR10","unstructured":"Khudia, D., et al.: FBGEMM: enabling high-performance low-precision deep learning inference. arXiv preprint arXiv:2101.05615 (2021)"},{"key":"17_CR11","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"17_CR12","doi-asserted-by":"publisher","unstructured":"Milletari, F., Navab, N., Ahmadi, S.A.: V-Net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV), pp. 565\u2013571 (2016). https:\/\/doi.org\/10.1109\/3DV.2016.79","DOI":"10.1109\/3DV.2016.79"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Nandamuri, S., China, D., Mitra, P., Sheet, D.: SUMNet: fully convolutional model for fast segmentation of anatomical structures in ultrasound volumes. In: IEEE International Symposium on Biomedical Imaging, pp. 1729\u20131732 (2019)","DOI":"10.1109\/ISBI.2019.8759210"},{"issue":"1","key":"17_CR14","doi-asserted-by":"publisher","first-page":"134","DOI":"10.3390\/electronics9010134","volume":"9","author":"SS Park","year":"2020","unstructured":"Park, S.S., Chung, K.S.: CENNA: cost-effective neural network accelerator. Electronics 9(1), 134 (2020)","journal-title":"Electronics"},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"Paupamah, K., James, S., Klein, R.: Quantisation and pruning for neural network compression and regularisation. In: International SAUPEC\/RobMech\/PRASA Conference, pp. 1\u20136. IEEE (2020)","DOI":"10.1109\/SAUPEC\/RobMech\/PRASA48453.2020.9041096"},{"key":"17_CR16","unstructured":"Petsiuk, V., Das, A., Saenko, K.: RISE: randomized input sampling for explanation of black-box models. In: British Machine Vision Conference (2018)"},{"key":"17_CR17","unstructured":"Ramachandran, P., Parmar, N., Vaswani, A., Bello, I., Levskaya, A., Shlens, J.: Stand-alone self-attention in vision models. arXiv preprint arXiv:1906.05909 (2019)"},{"key":"17_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Shaw, P., Uszkoreit, J., Vaswani, A.: Self-attention with relative position representations. arXiv preprint arXiv:1803.02155 (2018)","DOI":"10.18653\/v1\/N18-2074"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Vaswani, A., Ramachandran, P., Srinivas, A., Parmar, N., Hechtman, B., Shlens, J.: Scaling local self-attention for parameter efficient visual backbones. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12894\u201312904 (2021)","DOI":"10.1109\/CVPR46437.2021.01270"},{"key":"17_CR21","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Wang, X., Peng, Y., Lu, L., Lu, Z., Bagheri, M., Summers, R.M.: ChestX-ray8: hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2097\u20132106 (2017)","DOI":"10.1109\/CVPR.2017.369"},{"key":"17_CR23","unstructured":"Wu, S., Li, G., Chen, F., Shi, L.: Training and inference with integers in deep neural networks. arXiv preprint arXiv:1802.04680 (2018)"},{"key":"17_CR24","doi-asserted-by":"publisher","first-page":"3679","DOI":"10.1109\/TASLP.2021.3129357","volume":"29","author":"J Xu","year":"2021","unstructured":"Xu, J., Yu, J., Hu, S., Liu, X., Meng, H.M.: Mixed precision low-bit quantization of neural network language models for speech recognition. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 3679\u20133693 (2021)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Yang, F., Yang, H., Fu, J., Lu, H., Guo, B.: Learning texture transformer network for image super-resolution. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5791\u20135800 (2020)","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"17_CR26","doi-asserted-by":"crossref","unstructured":"Ye, L., Rochan, M., Liu, Z., Wang, Y.: Cross-modal self-attention network for referring image segmentation. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10502\u201310511 (2019)","DOI":"10.1109\/CVPR.2019.01075"},{"key":"17_CR27","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable DETR: deformable transformers for end-to-end object detection. arXiv preprint arXiv:2010.04159 (2020)"}],"container-title":["Lecture Notes in Computer Science","Distributed, Collaborative, and Federated Learning, and Affordable AI and Healthcare for Resource Diverse Global Health"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-18523-6_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,8]],"date-time":"2022-10-08T10:07:34Z","timestamp":1665223654000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-18523-6_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031185229","9783031185236"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-18523-6_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"7 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"FAIR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Workshop on Affordable Healthcare and AI for Resource Diverse Global Health","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"fair22022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/fair-workshop.github.io\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"9","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}