{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,2]],"date-time":"2025-10-02T06:10:48Z","timestamp":1759385448079,"version":"3.28.0"},"reference-count":25,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2018,10]]},"DOI":"10.1109\/acssc.2018.8645094","type":"proceedings-article","created":{"date-parts":[[2019,3,18]],"date-time":"2019-03-18T20:31:40Z","timestamp":1552941100000},"page":"66-70","source":"Crossref","is-referenced-by-count":4,"title":["Data Dropout in Arbitrary Basis for Deep Network Regularization"],"prefix":"10.1109","author":[{"given":"Mostafa","family":"Rahmani","sequence":"first","affiliation":[]},{"given":"George K.","family":"Atia","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","article-title":"Improving neural networks by preventing co-adaptation of feature detectors","author":"hinton","year":"2012","journal-title":"arXiv preprint arXiv 1207 0580"},{"key":"ref11","first-page":"545","article-title":"Rank, trace-norm and max-norm","volume":"5","author":"srebro","year":"2005","journal-title":"COLT"},{"key":"ref12","article-title":"Analyzing noise in autoencoders and deep networks","author":"poole","year":"2014","journal-title":"arXiv preprint arXiv 1406 1831"},{"key":"ref13","first-page":"1929","article-title":"Dropout: a simple way to prevent neural networks from overfitting","volume":"15","author":"srivastava","year":"2014","journal-title":"Journal of Machine Learning Research"},{"key":"ref14","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2014","journal-title":"arXiv preprint arXiv 1412 6980"},{"key":"ref15","first-page":"351","article-title":"Dropout training as adaptive regularization","author":"wager","year":"2013","journal-title":"Advances in neural information processing systems"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2014.7078567"},{"key":"ref17","first-page":"2523","article-title":"Improved dropout for shallow and deep learning","author":"li","year":"2016","journal-title":"Advances in neural information processing systems"},{"key":"ref18","article-title":"On fast dropout and its applicability to recurrent networks","author":"bayer","year":"2013","journal-title":"arXiv preprint arXiv 1311 0701"},{"key":"ref19","first-page":"118","article-title":"Fast dropout training","author":"wang","year":"2013","journal-title":"Proceedings of the 30th International Conference on Machine Learning (ICML-13)"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1179"},{"key":"ref3","article-title":"A neural conversational model","author":"vinyals","year":"2015","journal-title":"arXiv preprint arXiv 1506 04830"},{"key":"ref6","article-title":"OverFeat: Integrated recognition, localization and detection using convolutional networks","author":"sermanet","year":"2013","journal-title":"arXiv preprint arXiv 1312 6229"},{"key":"ref5","article-title":"Neural machine translation by jointly learning to align and translate","author":"bahdanau","year":"2014","journal-title":"arXiv preprint arXiv 1409 0473"},{"journal-title":"Deep Learning","year":"2016","author":"goodfellow","key":"ref8"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref2","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","author":"sutskever","year":"2014","journal-title":"Advances in neural information processing systems"},{"key":"ref9","article-title":"Wide residual networks","author":"zagoruyko","year":"2016","journal-title":"arXiv preprint arXiv 1605 01584"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2015.07.007"},{"key":"ref22","first-page":"3084","article-title":"Adaptive dropout for training deep neural networks","author":"ba","year":"2013","journal-title":"Advances in neural information processing systems"},{"key":"ref21","article-title":"Compacting neural network classifiers via dropout training","author":"kubo","year":"2016","journal-title":"arXiv preprint arXiv 1611 06148"},{"key":"ref24","first-page":"1139","article-title":"On the importance of initialization and momentum in deep learning","author":"sutskever","year":"2013","journal-title":"International Conference on Machine Learning"},{"key":"ref23","first-page":"2146","article-title":"What is the best multistage architecture for object recognition?","author":"jarrett","year":"2009","journal-title":"12th IEEE International Conference on Computer Vision"},{"key":"ref25","article-title":"Densely connected convolutional networks","author":"huang","year":"2016","journal-title":"arXiv preprint arXiv 1608 06993"}],"event":{"name":"2018 52nd Asilomar Conference on Signals, Systems, and Computers","start":{"date-parts":[[2018,10,28]]},"location":"Pacific Grove, CA, USA","end":{"date-parts":[[2018,10,31]]}},"container-title":["2018 52nd Asilomar Conference on Signals, Systems, and Computers"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8638648\/8645065\/08645094.pdf?arnumber=8645094","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,8,24]],"date-time":"2020-08-24T00:02:26Z","timestamp":1598227346000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8645094\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,10]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/acssc.2018.8645094","relation":{},"subject":[],"published":{"date-parts":[[2018,10]]}}}