{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2022,3,30]],"date-time":"2022-03-30T18:14:14Z","timestamp":1648664054138},"reference-count":0,"publisher":"IOS Press","license":[{"start":{"date-parts":[[2021,10,14]],"date-time":"2021-10-14T00:00:00Z","timestamp":1634169600000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,10,14]]},"abstract":"<jats:p>Quality of retinal image is vital for screening of ailments pertaining to eye such as glaucoma, diabetic retinopathy (DR) and age related macular degeneration. Therefore, assessing quality of retinal image prior to any kind of diagnosis has assumed significance in Computer Aided Desgin (CAD) applications. The rationale behind this is that reliability of retinal image is to be guaranteed to have dependable diagnosis. In this paper, we propose a novel retinal fundus image quality assessment (RIQA) method based on autoencoder network to assess retinal images if the image is acceptable for screening or not. The autoencoder network architecture is well suited to precisely to properly represent the key features of the image quality, especially when the network can correctly reconstruct the input image. The proposed model consists of encoder and decoder successive networks. The encoder will be used for representing the features of the input image. In turn , the decoder will be used for reconstruct the input image. The features get from encoder network will then be fed to a classifier in order to classify the quality of retinal image to two classes: gradable or ungradable. The experimental results revealed more useful assessment and the proposed deep model provides a superior performance for RIQA. Thus, our model can serve real-world Clinical Decision Support Systems in the healthcare domain.<\/jats:p>","DOI":"10.3233\/faia210160","type":"book-chapter","created":{"date-parts":[[2021,10,14]],"date-time":"2021-10-14T12:50:36Z","timestamp":1634215836000},"source":"Crossref","is-referenced-by-count":0,"title":["Efficient Fundus Image Gradeability Approach Based on Deep Reconstruction-Classification Network"],"prefix":"10.3233","author":[{"given":"Saif","family":"Khalid","sequence":"first","affiliation":[{"name":"DEIM, Universitat Rovira i Virgili, Tarragona, Spain"},{"name":"University of Al-Qadisiyah, Iraq"}]},{"given":"Saddam","family":"Abdulwahab","sequence":"additional","affiliation":[{"name":"DEIM, Universitat Rovira i Virgili, Tarragona, Spain"}]},{"given":"Hatem A.","family":"Rashwan","sequence":"additional","affiliation":[{"name":"DEIM, Universitat Rovira i Virgili, Tarragona, Spain"}]},{"given":"Juli\u00e1n","family":"Cristiano","sequence":"additional","affiliation":[{"name":"DEIM, Universitat Rovira i Virgili, Tarragona, Spain"}]},{"given":"Mohamed","family":"Abdel-Nasser","sequence":"additional","affiliation":[{"name":"DEIM, Universitat Rovira i Virgili, Tarragona, Spain"},{"name":"Department of Electrical Engineering, Aswan University, 81528 Aswan, Egypt"}]},{"given":"Domenec","family":"Puig","sequence":"additional","affiliation":[{"name":"DEIM, Universitat Rovira i Virgili, Tarragona, Spain"}]}],"member":"7437","container-title":["Frontiers in Artificial Intelligence and Applications","Artificial Intelligence Research and Development"],"original-title":[],"link":[{"URL":"https:\/\/ebooks.iospress.nl\/pdf\/doi\/10.3233\/FAIA210160","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,10,25]],"date-time":"2021-10-25T13:37:49Z","timestamp":1635169069000},"score":1,"resource":{"primary":{"URL":"https:\/\/ebooks.iospress.nl\/doi\/10.3233\/FAIA210160"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,14]]},"references-count":0,"URL":"https:\/\/doi.org\/10.3233\/faia210160","relation":{},"ISSN":["0922-6389","1879-8314"],"issn-type":[{"value":"0922-6389","type":"print"},{"value":"1879-8314","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,10,14]]}}}