{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T17:48:12Z","timestamp":1776275292481,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2021,3,9]],"date-time":"2021-03-09T00:00:00Z","timestamp":1615248000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,3,9]],"date-time":"2021-03-09T00:00:00Z","timestamp":1615248000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2021,10]]},"DOI":"10.1007\/s10489-021-02247-z","type":"journal-article","created":{"date-parts":[[2021,3,9]],"date-time":"2021-03-09T13:28:22Z","timestamp":1615296502000},"page":"7427-7439","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["PSigmoid: Improving squeeze-and-excitation block with parametric sigmoid"],"prefix":"10.1007","volume":"51","author":[{"given":"Yao","family":"Ying","sequence":"first","affiliation":[]},{"given":"Nengbo","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3274-3226","authenticated-orcid":false,"given":"Peng","family":"Shan","sequence":"additional","affiliation":[]},{"given":"Ligang","family":"Miao","sequence":"additional","affiliation":[]},{"given":"Peng","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Silong","family":"Peng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,3,9]]},"reference":[{"key":"2247_CR1","doi-asserted-by":"publisher","unstructured":"LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436\u2013444. https:\/\/doi.org\/10.1038\/NATURE14539","DOI":"10.1038\/NATURE14539"},{"key":"2247_CR2","doi-asserted-by":"publisher","unstructured":"Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L (2015) ImageNet Large Scale Visual Recognition Challenge. Int J Comput Vis 115:211\u2013252. https:\/\/doi.org\/10.1007\/s11263-015-0816-y","DOI":"10.1007\/s11263-015-0816-y"},{"key":"2247_CR3","doi-asserted-by":"publisher","unstructured":"Krizhevsky A, Sutskever I, Hinton G (2012) ImageNet classication with deep convolutional neural networks. In: Advances in neural information processing systems, pp 1097\u20131105. https:\/\/doi.org\/10.1145\/3065386","DOI":"10.1145\/3065386"},{"key":"2247_CR4","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Proceedings of the International conference on learning representations"},{"key":"2247_CR5","doi-asserted-by":"publisher","unstructured":"Szegedy C, Liu W, Jia Y, Sermanet PS et al (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1\u20139. https:\/\/doi.org\/10.1109\/CVPR.2015.7298594","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"2247_CR6","doi-asserted-by":"publisher","unstructured":"Bianchini M, Scarselli F (2014) On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures. IEEE Trans Neural Netw Learn Syst 25(8):1553\u20131565. https:\/\/doi.org\/10.1109\/TNNLS.2013.2293637","DOI":"10.1109\/TNNLS.2013.2293637"},{"key":"2247_CR7","doi-asserted-by":"publisher","unstructured":"Huang G, Liu Z, Pleiss G, van der Maaten L, Weinberger KQ (2019) Convolutional Networks with Dense Connectivity. IEEE Trans Pattern Anal Mach Intell:1\u20131. https:\/\/doi.org\/10.1109\/TPAMI.2019.2918284","DOI":"10.1109\/TPAMI.2019.2918284"},{"key":"2247_CR8","doi-asserted-by":"publisher","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778. https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"2247_CR9","doi-asserted-by":"publisher","unstructured":"Zhang K, Sun M, Han TX, Yuan Xf, Guo L, Liu T (2018) Residual Networks of Residual Networks: Multilevel Residual Networks. IEEE Trans Circ Syst Video Technol 28(6):1303\u20131314. https:\/\/doi.org\/10.1109\/TCSVT.2017.2654543","DOI":"10.1109\/TCSVT.2017.2654543"},{"key":"2247_CR10","unstructured":"Ioffe S, Szegedy C (2015) Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: Proceedings of the lnternational Conference on Machine Learning, pp 448\u2013456"},{"key":"2247_CR11","unstructured":"Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. J Mach Learn Res 9:249\u2013256"},{"key":"2247_CR12","unstructured":"Srivastava RK, Greff K, Schmidhuber J, Schmidhuber (2015) Training very deep networks. In: Advances in Neural Information Processing Systems, pp 2377\u20132385"},{"key":"2247_CR13","doi-asserted-by":"publisher","unstructured":"Huang G, Sun Y, Liu Z, Sedra D, Weinberger KQ (2016) Deep networks with stochastic depth. In: Proceedings of the European Conference on Computer Vision, pp 646\u2013661. https:\/\/doi.org\/10.1007\/978-3-319-46493-0_39","DOI":"10.1007\/978-3-319-46493-0_39"},{"key":"2247_CR14","unstructured":"Veit A, Wilber M, Belongie S (2016) Residual networks are exponential ensembles of relatively shallow networks. In: Advances in neural information processing systems, pp 550\u2013558"},{"key":"2247_CR15","doi-asserted-by":"publisher","unstructured":"Xie S, Girshick R, Dollar P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1492\u20131500. https:\/\/doi.org\/10.1109\/CVPR.2017.634","DOI":"10.1109\/CVPR.2017.634"},{"key":"2247_CR16","doi-asserted-by":"publisher","unstructured":"Chollet F (2017) Xception: Deep learning with depthwise separable convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 550\u2013558. https:\/\/doi.org\/10.1109\/CVPR.2017.195","DOI":"10.1109\/CVPR.2017.195"},{"key":"2247_CR17","doi-asserted-by":"publisher","unstructured":"Gao H, Wang Z, Cai L, Ji S (2020) ChannelNets: Compact and Efficient Convolutional Neural Networks via Channel-Wise Convolutions. IEEE Trans Pattern Anal Mach Intell:1\u20131. https:\/\/doi.org\/10.1109\/TPAMI.2020.2975796","DOI":"10.1109\/TPAMI.2020.2975796"},{"key":"2247_CR18","doi-asserted-by":"publisher","unstructured":"Hu J, Shen L, Sun G (2020) Squeeze-and-Excitation Networks. IEEE Trans Pattern Anal Mach Intell 42(8):2011\u20132023. https:\/\/doi.org\/10.1109\/TPAMI.2019.2913372","DOI":"10.1109\/TPAMI.2019.2913372"},{"key":"2247_CR19","doi-asserted-by":"publisher","unstructured":"Li Y, Fan C, Li Y, Wu Q, Ming Y (2018) Improving deep neural network with multiple parametric exponential linear units. Neurocomputing 301:11\u201324. https:\/\/doi.org\/10.1016\/j.neucom.2018.01.084","DOI":"10.1016\/j.neucom.2018.01.084"},{"key":"2247_CR20","doi-asserted-by":"publisher","unstructured":"Zhao H, Liu F, Li L, Luo C (2018) A novel softplus linear unit for deep convolutional neural networks. Appl Intell 48(7):1707\u20131720. https:\/\/doi.org\/10.1007\/s10489-017-1028-7","DOI":"10.1007\/s10489-017-1028-7"},{"key":"2247_CR21","doi-asserted-by":"publisher","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In: Proceedings of the IEEE International Conference on Computer Vision, pp 1026\u20131034. https:\/\/doi.org\/10.1109\/ICCV.2015.123","DOI":"10.1109\/ICCV.2015.123"},{"key":"2247_CR22","doi-asserted-by":"publisher","unstructured":"Njikam ANS, Zhao H (2016) A novel activation function for multilayer feed-forward neural networks. Appl Intell 45(1):75\u201382. https:\/\/doi.org\/10.1007\/s10489-015-0744-0","DOI":"10.1007\/s10489-015-0744-0"},{"key":"2247_CR23","doi-asserted-by":"publisher","unstructured":"Ying Y, Su J, Shan P, Miao L, Wang X, Peng S (2019) Rectified exponential units for convolutional neural networks. IEEE Access 7:101633\u2013101640. https:\/\/doi.org\/10.1109\/ACCESS.2019.2928442","DOI":"10.1109\/ACCESS.2019.2928442"},{"key":"2247_CR24","doi-asserted-by":"publisher","unstructured":"Kim D, Kim J, Kim J (2020) Elastic exponential linear units for convolutional neural networks. Neurocomputing 406:253\u2013266. https:\/\/doi.org\/10.1016\/j.neucom.2020.03.051","DOI":"10.1016\/j.neucom.2020.03.051"},{"key":"2247_CR25","doi-asserted-by":"publisher","unstructured":"Yu X, Ye X, Gao Q (2020) Infrared Handprint Image Restoration Algorithm Based on Apoptotic Mechanism. IEEE Access 8:47334\u201347343. https:\/\/doi.org\/10.1109\/ACCESS.2020.2979018","DOI":"10.1109\/ACCESS.2020.2979018"},{"key":"2247_CR26","doi-asserted-by":"publisher","unstructured":"Liu X, Zhu X, Li M, Wang L, Zhu E, Liu T, Kloft M, Shen D, Yin J, Gao W (2020) Multiple Kernel k k-Means with Incomplete Kernels. IEEE Trans Pattern Anal Mach Intell1191\u20131204. https:\/\/doi.org\/10.1109\/TPAMI.2019.2892416","DOI":"10.1109\/TPAMI.2019.2892416"},{"key":"2247_CR27","doi-asserted-by":"publisher","unstructured":"Chandra P, Singh Y (2004) An activation function adapting training algorithm for sigmoidal feedforward networks. Neurocomputing 61:429\u2013437. https:\/\/doi.org\/10.1016\/J.NEUCOM.2004.04.001","DOI":"10.1016\/J.NEUCOM.2004.04.001"},{"key":"2247_CR28","doi-asserted-by":"publisher","unstructured":"Sharma SK, Chandra P (2010) An adaptive slope sigmoidal function cascading neural networks algorithm. In: International Conference on Emerging Trends in Engineering and Technology, pp 531\u2013536. https:\/\/doi.org\/10.1109\/ICETET.2010.71","DOI":"10.1109\/ICETET.2010.71"},{"key":"2247_CR29","doi-asserted-by":"publisher","unstructured":"LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278\u20132324. https:\/\/doi.org\/10.1109\/5.726791","DOI":"10.1109\/5.726791"},{"key":"2247_CR30","unstructured":"Nair V, Hinton G (2010) Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the lnternational Conference on Machine Learning, pp 807\u2013814"},{"key":"2247_CR31","doi-asserted-by":"publisher","unstructured":"Hahnloser RHR, Sarpeshkar R, Mahowald MA, Douglas RJ, Sebastian Seung H (2000) Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit. Nature 405(6789):947\u2013951. https:\/\/doi.org\/10.1038\/35016072","DOI":"10.1038\/35016072"},{"key":"2247_CR32","doi-asserted-by":"publisher","unstructured":"Zhao M, Zhong S, Fu X, Tang B, Dong S, Pecht M (2020) Deep Residual Networks with Adaptively Parametric Rectifier Linear Units for Fault Diagnosis. IEEE Trans Indust Electron:1-1. https:\/\/doi.org\/10.1109\/TIE.2020.2972458","DOI":"10.1109\/TIE.2020.2972458"},{"key":"2247_CR33","unstructured":"Clevert DA, Unterthiner T, Hochreiter S (2016) Fast and accurate deep network learning by exponential linear units (ELUS). In: Proceedings of the International Conference on Learning Representations"},{"key":"2247_CR34","doi-asserted-by":"publisher","unstructured":"Zagoruyko S., Komodakis N. (2016) Wide residual networks. In: British Machine Vision Conference. https:\/\/doi.org\/10.5244\/C.30.87","DOI":"10.5244\/C.30.87"},{"key":"2247_CR35","unstructured":"Lin M, Chen Q, Yan S (2014) Network in network. In: Proceedings of the International Conference on Learning Representations"},{"key":"2247_CR36","doi-asserted-by":"publisher","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2818\u20132826. https:\/\/doi.org\/10.1109\/CVPR.2016.308","DOI":"10.1109\/CVPR.2016.308"},{"key":"2247_CR37","doi-asserted-by":"crossref","unstructured":"Szegedy C, Ioffe S, Vanhoucke V (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 4278\u20134284","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"2247_CR38","doi-asserted-by":"publisher","unstructured":"Zhang X, Zhou X, Lin M, Sun J (2018) ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848\u20136856. https:\/\/doi.org\/10.1109\/CVPR.2018.00716","DOI":"10.1109\/CVPR.2018.00716"},{"key":"2247_CR39","doi-asserted-by":"publisher","unstructured":"Hou S, Wang Z (2019) Weighted channel dropout for regularization of deep convolutional neural network. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 8425\u20138432. https:\/\/doi.org\/10.1609\/AAAI.V33I01.33018425","DOI":"10.1609\/AAAI.V33I01.33018425"},{"key":"2247_CR40","unstructured":"Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images. Department of Computer Science, University of Toronto, Masters thesis"},{"key":"2247_CR41","unstructured":"Netzer Y, Wang T, Coates A, Bissacco A, Wu B, Ng AY (2011) Reading digits in natural images with unsupervised feature learning. In: Advances in neural information processing systems workshop on deep learning and unsupervised feature learning"},{"key":"2247_CR42","unstructured":"Tiny ImageNet Visual Recognition Challenge. [Online] Available: https:\/\/tinyimagenet.herokuapp.com"},{"key":"2247_CR43","doi-asserted-by":"publisher","unstructured":"Zeiler MD, Fergus R (2014) Visualizing and understanding convolutional networks. In: Proceedings of the European Conference on Computer Vision, pp 818\u2013833. https:\/\/doi.org\/10.1007\/978-3-319-10590-1_53","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"2247_CR44","unstructured":"Alain G, Bengio Y (2017) Understanding intermediate layers using linear classifier probes. In: Proceedings of the International Conference on Learning Representations Workshop"},{"key":"2247_CR45","unstructured":"Yosinski J, Clune J, Bengio Y, Lipson H (2014) How transferable are features in deep neural networks? In: Advances in neural information processing systems, pp 3320\u20133328"},{"key":"2247_CR46","doi-asserted-by":"publisher","unstructured":"Lin T-Y, Maire M, Belongie S et al (2014) Microsoft COCO: Common Objects in Context. In: Proceedings of the European Conference on Computer Vision, pp 740\u2013755. https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2247_CR47","unstructured":"Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The Caltech-UCSD Birds-200-2011 dataset. California Institute of Technology"},{"key":"2247_CR48","doi-asserted-by":"publisher","unstructured":"Song W, Zheng J, Wu Y, Chen C, Liu F (2020) Discriminative feature extraction for video person re-identification via multi-task network. Applied Intelligence. https:\/\/doi.org\/10.1007\/s10489-020-01844-8","DOI":"10.1007\/s10489-020-01844-8"},{"key":"2247_CR49","doi-asserted-by":"publisher","unstructured":"Wu L, Wang Y, Li X, Gao J (2019) Deep Attention-Based Spatially Recursive Networks for Fine-Grained Visual Recognition. IEEE Trans Cybern 49(5):1791\u20131802. https:\/\/doi.org\/10.1109\/TCYB.2018.2813971","DOI":"10.1109\/TCYB.2018.2813971"},{"key":"2247_CR50","doi-asserted-by":"publisher","unstructured":"Zheng Z, An G, Wu D, Ruan Q (2020) Global and Local Knowledge-Aware Attention Network for Action Recognition. IEEE Trans Neural Netw Learn Syst:1\u201314. https:\/\/doi.org\/10.1109\/TNNLS.2020.2978613","DOI":"10.1109\/TNNLS.2020.2978613"},{"key":"2247_CR51","doi-asserted-by":"publisher","unstructured":"Choe J, Lee S, Shim H (2020) Attention-based Dropout Layer for Weakly Supervised Single Object Localization and Semantic Segmentation. IEEE Trans Pattern Anal Mach Intell:1\u20131. https:\/\/doi.org\/10.1109\/TPAMI.2020.2999099","DOI":"10.1109\/TPAMI.2020.2999099"},{"key":"2247_CR52","doi-asserted-by":"publisher","unstructured":"Woo S, Park J, Lee J-Y, So Kweon I (2018) Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision, pp 3\u201319. https:\/\/doi.org\/10.1007\/978-3-030-01234-2_1","DOI":"10.1007\/978-3-030-01234-2_1"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02247-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-021-02247-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02247-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,20]],"date-time":"2022-12-20T21:11:23Z","timestamp":1671570683000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-021-02247-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,3,9]]},"references-count":52,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2021,10]]}},"alternative-id":["2247"],"URL":"https:\/\/doi.org\/10.1007\/s10489-021-02247-z","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,3,9]]},"assertion":[{"value":"27 January 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 March 2021","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}