{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,30]],"date-time":"2025-10-30T07:14:22Z","timestamp":1761808462694,"version":"3.40.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030872335"},{"type":"electronic","value":"9783030872342"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-87234-2_1","type":"book-chapter","created":{"date-parts":[[2021,9,23]],"date-time":"2021-09-23T06:19:41Z","timestamp":1632377981000},"page":"3-13","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Learning More for Free - A Multi Task Learning Approach for Improved Pathology Classification in Capsule Endoscopy"],"prefix":"10.1007","author":[{"given":"Anuja","family":"Vats","sequence":"first","affiliation":[]},{"given":"Marius","family":"Pedersen","sequence":"additional","affiliation":[]},{"given":"Ahmed","family":"Mohammed","sequence":"additional","affiliation":[]},{"given":"\u00d8istein","family":"Hovde","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,9,21]]},"reference":[{"key":"1_CR1","unstructured":"Ali, S., et al.: Endoscopy artifact detection (EAD 2019) challenge dataset. arXiv preprint arXiv:1905.03209 (2019)"},{"issue":"2","key":"1_CR2","doi-asserted-by":"publisher","first-page":"33","DOI":"10.37126\/aige.v1.i2.33","volume":"1","author":"A Atsawarungruangkit","year":"2020","unstructured":"Atsawarungruangkit, A., Elfanagely, Y., Asombang, A.W., Rupawala, A., Rich, H.G.: Understanding deep learning in capsule endoscopy: can artificial intelligence enhance clinical practice? Artif. Intell. Gastrointest. Endosc. 1(2), 33\u201343 (2020)","journal-title":"Artif. Intell. Gastrointest. Endosc."},{"issue":"1","key":"1_CR3","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1023\/A:1007327622663","volume":"28","author":"J Baxter","year":"1997","unstructured":"Baxter, J.: A bayesian\/information theoretic model of learning to learn via multiple task sampling. Mach. Learn. 28(1), 7\u201339 (1997)","journal-title":"Mach. Learn."},{"key":"1_CR4","doi-asserted-by":"crossref","unstructured":"Benton, A., Mitchell, M., Hovy, D.: Multitask learning for mental health conditions with limited social media data. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, vol. 1, Long Papers, pp. 152\u2013162 (2017)","DOI":"10.18653\/v1\/E17-1015"},{"key":"1_CR5","doi-asserted-by":"crossref","unstructured":"Bingel, J., S\u00f8gaard, A.: Identifying beneficial task relations for multi-task learning in deep neural networks. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, vol. 2, Short Papers, pp. 164\u2013169. Association for Computational Linguistics, Valencia, Spain, April 2017. https:\/\/www.aclweb.org\/anthology\/E17-2026","DOI":"10.18653\/v1\/E17-2026"},{"issue":"1","key":"1_CR6","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1023\/A:1007379606734","volume":"28","author":"R Caruana","year":"1997","unstructured":"Caruana, R.: Multitask learning. Mach. Learn. 28(1), 41\u201375 (1997). https:\/\/doi.org\/10.1023\/A:1007379606734","journal-title":"Mach. Learn."},{"key":"1_CR7","doi-asserted-by":"crossref","unstructured":"Caruana, R.: Multitask learning: a knowledge-based source of inductive bias. In: Proceedings of the Tenth International Conference on Machine Learning, pp. 41\u201348. Morgan Kaufmann (1993)","DOI":"10.1016\/B978-1-55860-307-3.50012-5"},{"issue":"4","key":"1_CR8","doi-asserted-by":"publisher","first-page":"1044","DOI":"10.1053\/j.gastro.2019.06.025","volume":"157","author":"Z Ding","year":"2019","unstructured":"Ding, Z., et al.: Gastroenterologist-level identification of small-bowel diseases and normal variants by capsule endoscopy using a deep-learning model. Gastroenterology 157(4), 1044\u20131054 (2019)","journal-title":"Gastroenterology"},{"issue":"6","key":"1_CR9","doi-asserted-by":"publisher","first-page":"547","DOI":"10.5946\/ce.2018.173","volume":"51","author":"Y Hwang","year":"2018","unstructured":"Hwang, Y., Park, J., Lim, Y.J., Chun, H.J.: Application of artificial intelligence in capsule endoscopy: where are we now? Clin. Endosc. 51(6), 547\u2013551 (2018)","journal-title":"Clin. Endosc."},{"key":"1_CR10","unstructured":"Kendall, A., Gal, Y., Cipolla, R.: Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7482\u20137491 (2018)"},{"key":"1_CR11","doi-asserted-by":"crossref","unstructured":"Kokkinos, I.: Ubernet: training a universal convolutional neural network for low-, mid-, and high-level vision using diverse datasets and limited memory. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6129\u20136138 (2017)","DOI":"10.1109\/CVPR.2017.579"},{"issue":"6","key":"1_CR12","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017)","journal-title":"Commun. ACM"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Laiz, P., Vitria, J., Segu\u00ed, S.: Using the triplet loss for domain adaptation in WCE. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 399\u2013405 (2019)","DOI":"10.1109\/ICCVW.2019.00051"},{"issue":"3","key":"1_CR14","doi-asserted-by":"publisher","first-page":"E415","DOI":"10.1055\/a-1035-9088","volume":"8","author":"R Leenhardt","year":"2020","unstructured":"Leenhardt, R., Li, C., Le Mouel, J.P., Rahmi, G., Saurin, J.C., Cholet, F., Boureille, A., Amiot, X., Delvaux, M., Duburque, C., et al.: CAD-CAP: a 25,000-image database serving the development of artificial intelligence for capsule endoscopy. Endosc. Int. Open 8(3), E415 (2020)","journal-title":"Endosc. Int. Open"},{"key":"1_CR15","doi-asserted-by":"publisher","unstructured":"Liu, X., Gao, J., He, X., Deng, L., Duh, K., Wang, Y.y.: Representation learning using multi-task deep neural networks for semantic classification and information retrieval. In: Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 912\u2013921. Association for Computational Linguistics, Denver, Colorado, May\u2013June 2015. https:\/\/doi.org\/10.3115\/v1\/N15-1092, https:\/\/www.aclweb.org\/anthology\/N15-1092","DOI":"10.3115\/v1\/N15-1092"},{"issue":"19","key":"1_CR16","doi-asserted-by":"publisher","first-page":"369","DOI":"10.21037\/atm.2016.09.18","volume":"4","author":"ME McAlindon","year":"2016","unstructured":"McAlindon, M.E., Ching, H.L., Yung, D., Sidhu, R., Koulaouzidis, A.: Capsule endoscopy of the small bowel. Ann. Transl. Med. 4(19), 369 (2016)","journal-title":"Ann. Transl. Med."},{"key":"1_CR17","doi-asserted-by":"crossref","unstructured":"McInnes, L., Healy, J., Melville, J.: UMAP: uniform manifold approximation and projection for dimension reduction. arXiv preprint arXiv:1802.03426 (2018)","DOI":"10.21105\/joss.00861"},{"key":"1_CR18","doi-asserted-by":"crossref","unstructured":"Misra, I., Shrivastava, A., Gupta, A., Hebert, M.: Cross-stitch networks for multi-task learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3994\u20134003 (2016)","DOI":"10.1109\/CVPR.2016.433"},{"issue":"6","key":"1_CR19","doi-asserted-by":"publisher","first-page":"75","DOI":"10.3390\/jimaging4060075","volume":"4","author":"A Mohammed","year":"2018","unstructured":"Mohammed, A., Farup, I., Pedersen, M., Hovde, \u00d8., Yildirim Yayilgan, S.: Stochastic capsule endoscopy image enhancement. J. Imaging 4(6), 75 (2018)","journal-title":"J. Imaging"},{"key":"1_CR20","doi-asserted-by":"publisher","first-page":"266","DOI":"10.1016\/j.future.2020.06.048","volume":"113","author":"K Muhammad","year":"2020","unstructured":"Muhammad, K., Khan, S., Kumar, N., Del Ser, J., Mirjalili, S.: Vision-based personalized wireless capsule endoscopy for smart healthcare: taxonomy, literature review, opportunities and challenges. Future Gener. Comput. Syst. 113, 266\u2013280 (2020)","journal-title":"Future Gener. Comput. Syst."},{"issue":"4","key":"1_CR21","doi-asserted-by":"publisher","first-page":"317","DOI":"10.5946\/ce.2018.101","volume":"51","author":"J Park","year":"2018","unstructured":"Park, J., Cho, Y.K., Kim, J.H.: Current and future use of esophageal capsule endoscopy. Clin. Endosc. 51(4), 317\u2013322 (2018)","journal-title":"Clin. Endosc."},{"key":"1_CR22","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1016\/j.compbiomed.2016.10.011","volume":"79","author":"S Segu\u00ed","year":"2016","unstructured":"Segu\u00ed, S., Drozdzal, M., Pascual, G., Radeva, P., Malagelada, C., Azpiroz, F., Vitri\u00e0, J.: Generic feature learning for wireless capsule endoscopy analysis. Comput. Biol. Med. 79, 163\u2013172 (2016)","journal-title":"Comput. Biol. Med."},{"issue":"4","key":"1_CR23","doi-asserted-by":"publisher","first-page":"831","DOI":"10.1016\/j.gie.2020.04.039","volume":"92","author":"S Soffer","year":"2020","unstructured":"Soffer, S., Klang, E., Shimon, O., Nachmias, N., Eliakim, R., Ben-Horin, S., Kopylov, U., Barash, Y.: Deep learning for wireless capsule endoscopy: a systematic review and meta-analysis. Gastrointest. Endosc. 92(4), 831\u2013839 (2020)","journal-title":"Gastrointest. Endosc."},{"issue":"10","key":"1_CR24","doi-asserted-by":"publisher","first-page":"1490","DOI":"10.1093\/ibd\/izaa211","volume":"26","author":"S Syed","year":"2020","unstructured":"Syed, S., Stidham, R.W.: Potential for standardization and automation for pathology and endoscopy in inflammatory bowel disease. Inflamm. Bowel Dis. 26(10), 1490\u20131497 (2020)","journal-title":"Inflamm. Bowel Dis."},{"key":"1_CR25","doi-asserted-by":"publisher","first-page":"637","DOI":"10.1016\/j.procs.2019.12.230","volume":"164","author":"MT Val\u00e9rio","year":"2019","unstructured":"Val\u00e9rio, M.T., Gomes, S., Salgado, M., Oliveira, H.P., Cunha, A.: Lesions multiclass classification in endoscopic capsule frames. Procedia Comput. Sci. 164, 637\u2013645 (2019)","journal-title":"Procedia Comput. Sci."},{"issue":"4","key":"1_CR26","doi-asserted-by":"publisher","first-page":"387","DOI":"10.5946\/ce.2020.133","volume":"53","author":"YJ Yang","year":"2020","unstructured":"Yang, Y.J.: The future of capsule endoscopy: the role of artificial intelligence and other technical advancements. Clin. Endosc. 53(4), 387\u2013394 (2020)","journal-title":"Clin. Endosc."},{"key":"1_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1007\/978-3-319-10599-4_7","volume-title":"Computer Vision \u2013 ECCV 2014","author":"Z Zhang","year":"2014","unstructured":"Zhang, Z., Luo, P., Loy, C.C., Tang, X.: Facial landmark detection by deep multi-task learning. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8694, pp. 94\u2013108. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10599-4_7"},{"issue":"4","key":"1_CR28","doi-asserted-by":"publisher","first-page":"554","DOI":"10.1038\/ajg.2011.461","volume":"107","author":"Y Zheng","year":"2012","unstructured":"Zheng, Y., Hawkins, L., Wolff, J., Goloubeva, O., Goldberg, E.: Detection of lesions during capsule endoscopy: physician performance is disappointing. Am. J. Gastroenterol. 107(4), 554\u2013560 (2012)","journal-title":"Am. J. Gastroenterol."}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-87234-2_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,4]],"date-time":"2022-05-04T05:09:09Z","timestamp":1651640949000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-87234-2_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030872335","9783030872342"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-87234-2_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"21 September 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Strasbourg","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 October 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/miccai2021.org\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1622","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"531","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}