{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T22:43:52Z","timestamp":1760222632380,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":42,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030606350"},{"type":"electronic","value":"9783030606367"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-60636-7_26","type":"book-chapter","created":{"date-parts":[[2020,10,12]],"date-time":"2020-10-12T10:08:52Z","timestamp":1602497332000},"page":"310-319","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Soft-Root-Sign: A New Bounded Neural Activation Function"],"prefix":"10.1007","author":[{"given":"Dandan","family":"Li","sequence":"first","affiliation":[]},{"given":"Yuan","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,10,13]]},"reference":[{"issue":"1","key":"26_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1561\/2200000006","volume":"2","author":"Y Bengio","year":"2009","unstructured":"Bengio, Y., et al.: Learning deep architectures for AI. Found. Trends Mach. Learn. 2(1), 1\u2013127 (2009)","journal-title":"Found. Trends Mach. Learn."},{"issue":"1","key":"26_CR2","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1016\/j.neunet.2014.09.003","volume":"61","author":"J Schmidhuber","year":"2015","unstructured":"Schmidhuber, J.: Deep learning in neural networks: an overview. Neural Netw. 61(1), 85\u2013117 (2015)","journal-title":"Neural Netw."},{"issue":"7553","key":"26_CR3","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"26_CR4","unstructured":"Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 807\u2013814 (2010)"},{"key":"26_CR5","unstructured":"Clevert, D.A., Unterthiner, T., Hochreiter, S.: Fast and accurate deep network learning by exponential linear units. arXiv preprint arXiv:1511.07289 (2015)"},{"key":"26_CR6","unstructured":"Maas, A.L., Hannun, A.Y., Ng, A.Y.: Rectifier nonlinearities improve neural network acoustic models. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 3\u20139 (2013)"},{"key":"26_CR7","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 1026\u20131034 (2015)","DOI":"10.1109\/ICCV.2015.123"},{"key":"26_CR8","unstructured":"Xu, B., Wang, N., Chen, T., Li, M.: Empirical evaluation of rectified activations in convolutional network. arXiv preprint arXiv:1505.00853 (2015)"},{"key":"26_CR9","doi-asserted-by":"crossref","unstructured":"Misra, D.: Mish: a self regularized non-monotonic neural activation function. arXiv preprint arXiv:1908.08681 (2019)","DOI":"10.5244\/C.34.191"},{"key":"26_CR10","unstructured":"Gulcehre, C., Moczulski, M., Denil, M., Bengio, Y.: Noisy activation functions. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 3059\u20133068 (2016)"},{"key":"26_CR11","unstructured":"Basirat, M., Roth, P.M.: The quest for the golden activation function. arXiv preprint arXiv:1808.00783 (2018)"},{"key":"26_CR12","doi-asserted-by":"crossref","unstructured":"Jin, X., Xu, C., Feng, J., Wei, Y., Xiong, J., Yan, S.: Deep learning with S-shaped rectified linear activation units. In: 13th AAAI Conference on Artificial Intelligence (AAAI), pp. 1\u20138 (2016)","DOI":"10.1609\/aaai.v30i1.10287"},{"key":"26_CR13","unstructured":"Agostinelli, F., Hoffman, M., Sadowski, P., Baldi, P.: Learning activation functions to improve deep neural networks. arXiv preprint arXiv:1412.6830 (2014)"},{"issue":"1","key":"26_CR14","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.neunet.2017.12.012","volume":"107","author":"S Elfwing","year":"2018","unstructured":"Elfwing, S., Uchibe, E., Doya, K.: Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 107(1), 3\u201311 (2018)","journal-title":"Neural Netw."},{"key":"26_CR15","unstructured":"Shang, W., Sohn, K., Almeida, D., Lee, H.: Understanding and improving convolutional neural networks via concatenated rectified linear units. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 2217\u20132225 (2016)"},{"key":"26_CR16","unstructured":"Chen, T.Q., Behrmann, J., Duvenaud, D.K., Jacobsen, J.H.: Residual flows for invertible generative modeling. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 9913\u20139923 (2019)"},{"key":"26_CR17","doi-asserted-by":"crossref","unstructured":"Wuraola, A., Patel, N.: SQNL: a new computationally efficient activation function. In: International Joint Conference on Neural Networks (IJCNN), pp. 1\u20137 (2018)","DOI":"10.1109\/IJCNN.2018.8489043"},{"key":"26_CR18","unstructured":"Carlile, B., Delamarter, G., Kinney, P., Marti, A., Whitney, B.: Improving deep learning by inverse square root linear units (ISRLUs). arXiv preprint arXiv:1710.09967 (2017)"},{"key":"26_CR19","unstructured":"Nicolae, A.: PLU: the piecewise linear unit activation function. arXiv preprint arXiv:1809.09534 (2018)"},{"key":"26_CR20","unstructured":"Eidnes, L., N\u00f8kland, A.: Shifting mean activation towards zero with bipolar activation functions. arXiv preprint arXiv:1709.04054 (2017)"},{"key":"26_CR21","unstructured":"Hendrycks, D., Gimpel, K.: Gaussian error linear units (GELUs). arXiv preprint arXiv:1606.08415 (2016)"},{"key":"26_CR22","doi-asserted-by":"crossref","unstructured":"Godfrey, L.B., Gashler, M.S.: A continuum among logarithmic, linear, and exponential functions, and its potential to improve generalization in neural networks. In: 7th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management (IC3K), pp. 481\u2013486 (2015)","DOI":"10.5220\/0005635804810486"},{"key":"26_CR23","unstructured":"Alcaide, E.: E-swish: adjusting activations to different network depths. arXiv preprint arXiv:1801.07145 (2018)"},{"issue":"2","key":"26_CR24","doi-asserted-by":"publisher","first-page":"76","DOI":"10.26555\/ijain.v4i2.249","volume":"4","author":"HH Chieng","year":"2018","unstructured":"Chieng, H.H., Wahid, N., Pauline, O., Perla, S.R.K.: Flatten-T Swish: a thresholded RElU-swish-like activation function for deep learning. Int. J. Adv. Intell. Inform. 4(2), 76\u201386 (2018)","journal-title":"Int. J. Adv. Intell. Inform."},{"key":"26_CR25","doi-asserted-by":"crossref","unstructured":"Qiu, S., Xu, X., Cai, B.: FReLU: flexible rectified linear units for improving convolutional neural networks. In: 24th International Conference on Pattern Recognition (ICPR), pp. 1223\u20131228 (2018)","DOI":"10.1109\/ICPR.2018.8546022"},{"key":"26_CR26","doi-asserted-by":"crossref","unstructured":"Trottier, L., Gigu, P., Chaib-draa, B., et al.: Parametric exponential linear unit for deep convolutional neural networks. In: 16th IEEE International Conference on Machine Learning and Applications (ICMLA), pp. 207\u2013214 (2017)","DOI":"10.1109\/ICMLA.2017.00038"},{"issue":"1","key":"26_CR27","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.neucom.2018.01.084","volume":"301","author":"Y Li","year":"2018","unstructured":"Li, Y., Fan, C., Li, Y., Wu, Q., Ming, Y.: Improving deep neural network with multiple parametric exponential linear units. Neurocomputing 301(1), 11\u201324 (2018)","journal-title":"Neurocomputing"},{"issue":"1","key":"26_CR28","doi-asserted-by":"publisher","first-page":"281","DOI":"10.1016\/j.neunet.2020.02.012","volume":"125","author":"Q Cheng","year":"2020","unstructured":"Cheng, Q., Li, H., Wu, Q., Ma, L., King, N.N.: Parametric deformable exponential linear units for deep neural networks. Neural Netw. 125(1), 281\u2013289 (2020)","journal-title":"Neural Netw."},{"key":"26_CR29","doi-asserted-by":"crossref","unstructured":"Duggal, R., Gupta, A.: P-TELU: parametric tan hyperbolic linear unit activation for deep neural networks. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCV), pp. 974\u2013978 (2017)","DOI":"10.1109\/ICCVW.2017.119"},{"key":"26_CR30","unstructured":"Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 448\u2013456 (2015)"},{"key":"26_CR31","unstructured":"Klambauer, G., Unterthiner, T., Mayr, A., Hochreiter, S.: Self-normalizing neural networks. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 971\u2013980 (2017)"},{"issue":"1","key":"26_CR32","first-page":"9","volume":"1","author":"YA LeCun","year":"2012","unstructured":"LeCun, Y.A., Bottou, L., Orr, G.B., M\u00fcller, K.R.: Efficient backprop. Neural Netw.: Tricks Trade 1(1), 9\u201348 (2012)","journal-title":"Neural Netw.: Tricks Trade"},{"key":"26_CR33","unstructured":"Ramachandran, P., Zoph, B., Le, Q.V.: Searching for activation functions. arXiv preprint arXiv:1710.05941 (2017)"},{"key":"26_CR34","doi-asserted-by":"crossref","unstructured":"Liu, C., et al.: Progressive neural architecture search. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 19\u201334 (2018)","DOI":"10.1007\/978-3-030-01246-5_2"},{"issue":"1","key":"26_CR35","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1016\/S0893-6080(02)00228-9","volume":"16","author":"N Schweighofer","year":"2003","unstructured":"Schweighofer, N., Doya, K.: Meta-learning in reinforcement learning. Neural Netw. 16(1), 5\u20139 (2003)","journal-title":"Neural Netw."},{"key":"26_CR36","unstructured":"Glorot, X., Bordes, A., Bengio, Y.: Deep sparse rectifier neural networks. In: Proceedings of the 14th International Conference on Artificial Intelligence and Statistics, pp. 315\u2013323 (2011)"},{"key":"26_CR37","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"26_CR38","unstructured":"Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"26_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"630","DOI":"10.1007\/978-3-319-46493-0_38","volume-title":"Computer Vision \u2013 ECCV 2016","author":"K He","year":"2016","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Identity mappings in deep residual networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 630\u2013645. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_38"},{"key":"26_CR40","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"26_CR41","unstructured":"Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, pp. 249\u2013256 (2010)"},{"issue":"1","key":"26_CR42","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-60636-7_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T22:02:47Z","timestamp":1760220167000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-60636-7_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030606350","9783030606367"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-60636-7_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"13 October 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Nanjing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 October 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.prcv.cn\/index_en.html","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT system","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"402","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"158","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"39% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}