{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T11:47:01Z","timestamp":1769168821878,"version":"3.49.0"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031082764","type":"print"},{"value":"9783031082771","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-08277-1_27","type":"book-chapter","created":{"date-parts":[[2022,6,16]],"date-time":"2022-06-16T12:13:01Z","timestamp":1655381581000},"page":"327-340","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["A Novel Deep Convolutional Neural Network Architecture for Customer Counting in the Retail Environment"],"prefix":"10.1007","author":[{"given":"Almustafa","family":"Abed","sequence":"first","affiliation":[]},{"given":"Belhassen","family":"Akrout","sequence":"additional","affiliation":[]},{"given":"Ikram","family":"Amous","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,6,17]]},"reference":[{"issue":"2","key":"27_CR1","doi-asserted-by":"publisher","first-page":"165","DOI":"10.1007\/s10846-017-0674-7","volume":"91","author":"M Paolanti","year":"2017","unstructured":"Paolanti, M., Liciotti, D., Pietrini, R., Mancini, A., Frontoni, E.: Modelling and forecasting customer navigation in intelligent retail environments. J. Intell. Rob. Syst. 91(2), 165\u2013180 (2017). https:\/\/doi.org\/10.1007\/s10846-017-0674-7","journal-title":"J. Intell. Rob. Syst."},{"key":"27_CR2","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1016\/j.patrec.2014.09.013","volume":"53","author":"J Liu","year":"2015","unstructured":"Liu, J., Liu, Y., Zhang, G., Zhu, P., Chen, Y.Q.: Detecting and tracking people in real time with RGB-D camera. Pattern Recogn. Lett. 53, 16\u201323 (2015). https:\/\/doi.org\/10.1016\/j.patrec.2014.09.013","journal-title":"Pattern Recogn. Lett."},{"key":"27_CR3","doi-asserted-by":"publisher","unstructured":"Liang, B., Zheng, L.: A survey on human action recognition using depth sensors. In: 2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Adelaide, Australia, pp. 1\u20138, November 2015. https:\/\/doi.org\/10.1109\/DICTA.2015.7371223","DOI":"10.1109\/DICTA.2015.7371223"},{"key":"27_CR4","doi-asserted-by":"publisher","unstructured":"Paolanti, M., Sturari, M., Mancini, A., Zingaretti, P., Frontoni, E.: Mobile robot for retail surveying and inventory using visual and textual analysis of monocular pictures based on deep learning. In: 2017 European Conference on Mobile Robots (ECMR), Paris, pp. 1\u20136, September 2017. https:\/\/doi.org\/10.1109\/ECMR.2017.8098666","DOI":"10.1109\/ECMR.2017.8098666"},{"key":"27_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"207","DOI":"10.1007\/978-3-319-70742-6_20","volume-title":"New Trends in Image Analysis and Processing \u2013 ICIAP 2017","author":"D Liciotti","year":"2017","unstructured":"Liciotti, D., Paolanti, M., Frontoni, E., Zingaretti, P.: People detection and tracking from an RGB-D camera in top-view configuration: review of challenges and applications. In: Battiato, S., Farinella, G.M., Leo, M., Gallo, G. (eds.) New Trends in Image Analysis and Processing \u2013 ICIAP 2017. LNCS, vol. 10590, pp. 207\u2013218. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-70742-6_20"},{"key":"27_CR6","doi-asserted-by":"publisher","unstructured":"Liciotti, D.: TVHeads (Top-View Heads) Dataset, vol. 1, January 2018. https:\/\/doi.org\/10.17632\/nz4hy7yrps.1","DOI":"10.17632\/nz4hy7yrps.1"},{"key":"27_CR7","doi-asserted-by":"publisher","first-page":"103055","DOI":"10.1016\/j.micron.2021.103055","volume":"145","author":"B Akrout","year":"2021","unstructured":"Akrout, B.: A new structure of decision tree based on oriented edges gradient map for circles detection and the analysis of nano-particles. Micron 145, 103055 (2021). https:\/\/doi.org\/10.1016\/j.micron.2021.103055","journal-title":"Micron"},{"key":"27_CR8","doi-asserted-by":"publisher","unstructured":"Akrout, B., Mahdi, W.: A novel approach for driver fatigue detection based on visual characteristics analysis. J. Ambient Intell. Hum. Comput. 1\u201326 (2021). https:\/\/doi.org\/10.1007\/s12652-021-03311-9","DOI":"10.1007\/s12652-021-03311-9"},{"key":"27_CR9","doi-asserted-by":"publisher","unstructured":"Bondi, E., Seidenari, L., Bagdanov, A.D., Del Bimbo, A.: Real-time people counting from depth imagery of crowded environments. In: 2014 11th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Seoul, South Korea, pp. 337\u2013342, August 2014. https:\/\/doi.org\/10.1109\/AVSS.2014.6918691","DOI":"10.1109\/AVSS.2014.6918691"},{"key":"27_CR10","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.patrec.2016.05.033","volume":"81","author":"L Del Pizzo","year":"2016","unstructured":"Del Pizzo, L., Foggia, P., Greco, A., Percannella, G., Vento, M.: Counting people by RGB or depth overhead cameras. Pattern Recogn. Lett. 81, 41\u201350 (2016). https:\/\/doi.org\/10.1016\/j.patrec.2016.05.033","journal-title":"Pattern Recogn. Lett."},{"key":"27_CR11","doi-asserted-by":"publisher","unstructured":"Liciotti, D., Paolanti, M., Pietrini, R., Frontoni, E., Zingaretti, P.: Convolutional networks for semantic heads segmentation using top-view depth data in crowded environment. In: 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, pp. 1384\u20131389, August 2018. https:\/\/doi.org\/10.1109\/ICPR.2018.8545397","DOI":"10.1109\/ICPR.2018.8545397"},{"issue":"3","key":"27_CR12","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1109\/TCE.2012.6311323","volume":"58","author":"B Mrazovac","year":"2012","unstructured":"Mrazovac, B., Bjelica, M.Z., Kukolj, D., Todorovi, B.M.: A human detection method for residential smart energy systems based on Zigbee RSSI changes. IEEE Trans. Consum. Electron. 58(3), 6 (2012)","journal-title":"IEEE Trans. Consum. Electron."},{"issue":"9","key":"27_CR13","doi-asserted-by":"publisher","first-page":"3991","DOI":"10.1109\/TIE.2012.2206330","volume":"60","author":"J Garcia","year":"2013","unstructured":"Garcia, J., Gardel, A., Bravo, I., Lazaro, J.L., Martinez, M., Rodriguez, D.: Directional people counter based on head tracking. IEEE Trans. Ind. Electron. 60(9), 3991\u20134000 (2013). https:\/\/doi.org\/10.1109\/TIE.2012.2206330","journal-title":"IEEE Trans. Ind. Electron."},{"key":"27_CR14","doi-asserted-by":"publisher","unstructured":"Wang, C., Zhang, H., Yang, L., Liu, S., Cao, X.: Deep people counting in extremely dense crowds. In: Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, pp. 1299\u20131302, October 2015. https:\/\/doi.org\/10.1145\/2733373.2806337","DOI":"10.1145\/2733373.2806337"},{"key":"27_CR15","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.engappai.2015.04.006","volume":"43","author":"M Fu","year":"2015","unstructured":"Fu, M., Xu, P., Li, X., Liu, Q., Ye, M., Zhu, C.: Fast crowd density estimation with convolutional neural networks. Eng. Appl. Artif. Intell. 43, 81\u201388 (2015). https:\/\/doi.org\/10.1016\/j.engappai.2015.04.006","journal-title":"Eng. Appl. Artif. Intell."},{"key":"27_CR16","doi-asserted-by":"publisher","unstructured":"Zhang, C., Li, H., Wang, X., Yang, X.: Cross-scene crowd counting via deep convolutional neural networks. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, pp. 833\u2013841, June 2015. https:\/\/doi.org\/10.1109\/CVPR.2015.7298684","DOI":"10.1109\/CVPR.2015.7298684"},{"key":"27_CR17","doi-asserted-by":"publisher","unstructured":"Noh, H., Hong, S., Han, B.: Learning deconvolution network for semantic segmentation. In: 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, pp. 1520\u20131528, December 2015. https:\/\/doi.org\/10.1109\/ICCV.2015.178","DOI":"10.1109\/ICCV.2015.178"},{"key":"27_CR18","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder-decoder architecture for image segmentation, October 2016. arXiv:1511.00561, http:\/\/arxiv.org\/abs\/1511.00561. Accessed 01 Apr 2021"},{"key":"27_CR19","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition, April 2015. arXiv:1409.1556, http:\/\/arxiv.org\/abs\/1409.1556. Accessed 02 Apr 2021"},{"key":"27_CR20","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, pp. 770\u2013778, June 2016. https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"27_CR21","unstructured":"Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation, August 2018. arXiv:1802.02611, http:\/\/arxiv.org\/abs\/1802.02611. Accessed 26 Jan 2022"},{"key":"27_CR22","unstructured":"Chen, L.-C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation, December 2017. arXiv:1706.05587, http:\/\/arxiv.org\/abs\/1706.05587. Accessed 22 Sep 2021"},{"issue":"9","key":"27_CR23","doi-asserted-by":"publisher","first-page":"2627","DOI":"10.1109\/TCSVT.2018.2803115","volume":"29","author":"MB Shami","year":"2019","unstructured":"Shami, M.B., Maqbool, S., Sajid, H., Ayaz, Y., Cheung, S.-C.S.: People counting in dense crowd images using sparse head detections. IEEE Trans. Circ. Syst. Video Technol. 29(9), 2627\u20132636 (2019). https:\/\/doi.org\/10.1109\/TCSVT.2018.2803115","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"27_CR24","doi-asserted-by":"publisher","unstructured":"Nogueira, V., Oliveira, H., Augusto Silva, J., Vieira, T., Oliveira, K.: RetailNet: a deep learning approach for people counting and hot spots detection in retail stores. In: 2019 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Rio de Janeiro, Brazil, pp. 155\u2013162, October 2019. https:\/\/doi.org\/10.1109\/SIBGRAPI.2019.00029","DOI":"10.1109\/SIBGRAPI.2019.00029"},{"key":"27_CR25","doi-asserted-by":"crossref","unstructured":"Takikawa, T., Acuna, D., Jampani, V., Fidler, S.: Gated-SCNN: gated shape CNNs for semantic segmentation, July 2019. arXiv:1907.05740, http:\/\/arxiv.org\/abs\/1907.05740. Accessed 06 Dec 2021","DOI":"10.1109\/ICCV.2019.00533"},{"key":"27_CR26","unstructured":"Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, May 2017. arXiv:1606.00915, http:\/\/arxiv.org\/abs\/1606.00915. Accessed 26 Jan 2022"},{"key":"27_CR27","doi-asserted-by":"publisher","unstructured":"Jaccard, P.: \u00c9tude comparative de la distribution florale dans une portion des Alpes et du Jura (1901). https:\/\/doi.org\/10.5169\/SEALS-266450","DOI":"10.5169\/SEALS-266450"},{"key":"27_CR28","doi-asserted-by":"publisher","unstructured":"Akrout, B., Fakhfakh, S.: Three-dimensional head-pose estimation for smart Iris recognition from a calibrated camera. Math. Prob. Eng. 2020 (2020). https:\/\/doi.org\/10.1155\/2020\/9830672","DOI":"10.1155\/2020\/9830672"},{"key":"27_CR29","doi-asserted-by":"publisher","first-page":"511","DOI":"10.13053\/cys-23-2-3013","volume":"23","author":"W Mahdi","year":"2019","unstructured":"Mahdi, W., Akrout, B., Alroobaea, R., Alsufyani, A.: Automated drowsiness detection through facial features analysis. Computaci\u00f3n y Sistemas 23, 511\u2013521 (2019). https:\/\/doi.org\/10.13053\/cys-23-2-3013","journal-title":"Computaci\u00f3n y Sistemas"}],"container-title":["Communications in Computer and Information Science","Intelligent Systems and Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-08277-1_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,16]],"date-time":"2022-06-16T12:18:30Z","timestamp":1655381910000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-08277-1_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031082764","9783031082771"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-08277-1_27","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 June 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ISPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Systems and Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hammamet","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tunisia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 March 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 March 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ispr22022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ispr2022.sciencesconf.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"91","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"22","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"10","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"24% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Due to the COVID-19 pandemic the conference was held online.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}