{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,16]],"date-time":"2026-04-16T19:29:18Z","timestamp":1776367758686,"version":"3.51.2"},"reference-count":59,"publisher":"Association for Computing Machinery (ACM)","issue":"12","license":[{"start":{"date-parts":[[2020,11,17]],"date-time":"2020-11-17T00:00:00Z","timestamp":1605571200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Commun. ACM"],"published-print":{"date-parts":[[2020,11,17]]},"abstract":"<jats:p>Creating efficiency in AI research will decrease its carbon footprint and increase its inclusivity as deep learning study should not require the deepest pockets.<\/jats:p>","DOI":"10.1145\/3381831","type":"journal-article","created":{"date-parts":[[2020,11,18]],"date-time":"2020-11-18T23:58:12Z","timestamp":1605743892000},"page":"54-63","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1082,"title":["Green AI"],"prefix":"10.1145","volume":"63","author":[{"given":"Roy","family":"Schwartz","sequence":"first","affiliation":[{"name":"Hebrew University of Jerusalem, Israel"}]},{"given":"Jesse","family":"Dodge","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA"}]},{"given":"Noah A.","family":"Smith","sequence":"additional","affiliation":[{"name":"University of Washington"}]},{"given":"Oren","family":"Etzioni","sequence":"additional","affiliation":[{"name":"University of Washington, Seattle, WA"}]}],"member":"320","published-online":{"date-parts":[[2020,11,17]]},"reference":[{"key":"e_1_2_1_1_1","volume-title":"Proceedings of ICML Workshop on Climate Change","author":"Acharyya P.","year":"2019","unstructured":"Acharyya , P. , Rosario , S.D. , Flor , F. , Joshi , R. , Li , D. , Linares , R , and Zhang , H . Autopilot of cement plants for reduction of fuel consumption and emissions . In Proceedings of ICML Workshop on Climate Change , 2019 . Acharyya, P., Rosario, S.D., Flor, F., Joshi, R., Li, D., Linares, R, and Zhang, H. Autopilot of cement plants for reduction of fuel consumption and emissions. In Proceedings of ICML Workshop on Climate Change, 2019."},{"key":"e_1_2_1_2_1","volume-title":"AI and compute","author":"Amodei D.","year":"2018","unstructured":"Amodei , D. and Hernandez , D . AI and compute , 2018 . Blog post. Amodei, D. and Hernandez, D. AI and compute, 2018. Blog post."},{"key":"e_1_2_1_3_1","volume-title":"Proceedings of NeurIPS","author":"Bergstra J.S.","year":"2011","unstructured":"Bergstra , J.S. , Bardenet , R. , Bengio , Y. and K\u00e9gl , B . Algorithms for hyper-parameter optimization . In Proceedings of NeurIPS , 2011 . Bergstra, J.S., Bardenet, R., Bengio, Y. and K\u00e9gl, B. Algorithms for hyper-parameter optimization. In Proceedings of NeurIPS, 2011."},{"key":"e_1_2_1_4_1","volume-title":"et al. Language models are few-shot learners","author":"Brown T.B.","year":"2020","unstructured":"Brown , T.B. et al. Language models are few-shot learners , 2020 ; arXiv:2005.14165. Brown, T.B. et al. Language models are few-shot learners, 2020; arXiv:2005.14165."},{"key":"e_1_2_1_5_1","volume-title":"Proceedings of ISCAS","author":"Canziani A.","year":"2017","unstructured":"Canziani , A. , Paszke , A. and Culurciello , E . An analysis of deep neural network models for practical applications . In Proceedings of ISCAS , 2017 . Canziani, A., Paszke, A. and Culurciello, E. An analysis of deep neural network models for practical applications. In Proceedings of ISCAS, 2017."},{"key":"e_1_2_1_6_1","volume-title":"Proceedings of NeurIPS","author":"Chen Y.","year":"2017","unstructured":"Chen , Y. , Li , J. , Xiao , H. , Jin , X. , Yan , S. and Feng , J . Dual path networks . In Proceedings of NeurIPS , 2017 . Chen, Y., Li, J., Xiao, H., Jin, X., Yan, S. and Feng, J. Dual path networks. In Proceedings of NeurIPS, 2017."},{"key":"e_1_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_2_1_8_1","volume-title":"Proceedings of NAACL","author":"Devlin J.","year":"2019","unstructured":"Devlin , J. , Chang , M.W. , Lee , K. , and Toutanova , K . BERT: Pretraining of deep bidirectional transformers for language understanding . In Proceedings of NAACL , 2019 . Devlin, J., Chang, M.W., Lee, K., and Toutanova, K. BERT: Pretraining of deep bidirectional transformers for language understanding. In Proceedings of NAACL, 2019."},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1224"},{"key":"e_1_2_1_10_1","volume-title":"Fine-tuning pretrained language models: Weight initializations, data orders, and early stopping","author":"Dodge J.","year":"2020","unstructured":"Dodge , J. , Ilharco , G. , Schwartz , R. , Farhadi , A. , Hajishirzi , H. and Smith , N.A . Fine-tuning pretrained language models: Weight initializations, data orders, and early stopping , 2020 ; arXiv:2002.06305. Dodge, J., Ilharco, G., Schwartz, R., Farhadi, A., Hajishirzi, H. and Smith, N.A. Fine-tuning pretrained language models: Weight initializations, data orders, and early stopping, 2020; arXiv:2002.06305."},{"key":"e_1_2_1_11_1","volume-title":"Proceedings of AutoML","author":"Dodge J.","year":"2017","unstructured":"Dodge , J. , Jamieson , K. and Smith , N.A . Open loop hyperparameter optimization and determinantal point processes . In Proceedings of AutoML , 2017 . Dodge, J., Jamieson, K. and Smith, N.A. Open loop hyperparameter optimization and determinantal point processes. In Proceedings of AutoML, 2017."},{"key":"e_1_2_1_12_1","volume-title":"Proceedings of ICML Workshop on Climate Change","author":"Duhart C.","year":"2019","unstructured":"Duhart , C. , Dublon , G. , Mayton , B. , Davenport , G. and Paradiso , J.A . Deep learning for wildlife conservation and restoration efforts . In Proceedings of ICML Workshop on Climate Change , 2019 . Duhart, C., Dublon, G., Mayton, B., Davenport, G. and Paradiso, J.A. Deep learning for wildlife conservation and restoration efforts. In Proceedings of ICML Workshop on Climate Change, 2019."},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00171"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/MIS.2009.36"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_2_1_16_1","volume-title":"Towards the systematic reporting of the energy and carbon footprints of machine learning","author":"Henderson P.","year":"2020","unstructured":"Henderson , P. , Hu , J. , Romoff , J. , Brunskill , E. , Jurafsky , D. and Pineau , J . Towards the systematic reporting of the energy and carbon footprints of machine learning , 2020 ; arXiv:2002.05651. Henderson, P., Hu, J., Romoff, J., Brunskill, E., Jurafsky, D. and Pineau, J. Towards the systematic reporting of the energy and carbon footprints of machine learning, 2020; arXiv:2002.05651."},{"key":"e_1_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_2_1_18_1","volume-title":"et al. MobileNets: Efficient convolutional neural networks for mobile vision applications","author":"Howard A.G.","year":"2017","unstructured":"Howard , A.G. et al. MobileNets: Efficient convolutional neural networks for mobile vision applications , 2017 ; arXiv:1704.04861. Howard, A.G. et al. MobileNets: Efficient convolutional neural networks for mobile vision applications, 2017; arXiv:1704.04861."},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00745"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.351"},{"key":"e_1_2_1_21_1","volume-title":"Proceedings of NeurIPS","author":"Jeon Y.","year":"2018","unstructured":"Jeon , Y. and Kim , J . Constructing fast network through deconstruction of convolution . In Proceedings of NeurIPS , 2018 . Jeon, Y. and Kim, J. Constructing fast network through deconstruction of convolution. In Proceedings of NeurIPS, 2018."},{"key":"e_1_2_1_22_1","first-page":"1","volume-title":"Proceedings of ISCA 1","author":"Jouppi N.P.","year":"2017","unstructured":"Jouppi , N.P. et al. In-datacenter performance analysis of a tensor processing unit . In Proceedings of ISCA 1 , 1 ( 2017 ), Publ. date : June 2020. Jouppi, N.P. et al. In-datacenter performance analysis of a tensor processing unit. In Proceedings of ISCA 1, 1 (2017), Publ. date: June 2020."},{"key":"e_1_2_1_23_1","volume-title":"Proceedings of AISTATS","author":"Kamthe S.","year":"2018","unstructured":"Kamthe , S. and Deisenroth , M.P . Data-efficient reinforcement learning with probabilistic model predictive control . In Proceedings of AISTATS , 2018 . Kamthe, S. and Deisenroth, M.P. Data-efficient reinforcement learning with probabilistic model predictive control. In Proceedings of AISTATS, 2018."},{"key":"e_1_2_1_24_1","volume-title":"Proceedings of NeurIPS","author":"Krizhevsky A.","year":"2012","unstructured":"Krizhevsky , A. , Sutskever , I. and Hinton , G.E . Imagenet classification with deep convolutional neural networks . In Proceedings of NeurIPS , 2012 . Krizhevsky, A., Sutskever, I. and Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of NeurIPS, 2012."},{"key":"e_1_2_1_25_1","volume-title":"Proceedings of the Climate Change AI Workshop","author":"Lacoste A.","year":"2019","unstructured":"Lacoste , A. , Luccioni , A. , Schmidt , V. and Dandres , T . Quantifying the carbon emissions of machine learning . In Proceedings of the Climate Change AI Workshop , 2019 . Lacoste, A., Luccioni, A., Schmidt, V. and Dandres, T. Quantifying the carbon emissions of machine learning. In Proceedings of the Climate Change AI Workshop, 2019."},{"key":"e_1_2_1_26_1","volume-title":"Proceedings of ICLR","author":"Li L.","year":"2017","unstructured":"Li , L. , Jamieson , K. , DeSalvo , G. , Rostamizadeh , A. and Talwalkar , A . Hyperband: Bandit-based configuration evaluation for hyperparameter optimization . In Proceedings of ICLR , 2017 . Li, L., Jamieson, K., DeSalvo, G., Rostamizadeh, A. and Talwalkar, A. Hyperband: Bandit-based configuration evaluation for hyperparameter optimization. In Proceedings of ICLR, 2017."},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"e_1_2_1_28_1","volume-title":"et al. RoBERTa: A robustly optimized BERT pretraining approach","author":"Liu Y.","year":"2019","unstructured":"Liu , Y. et al. RoBERTa: A robustly optimized BERT pretraining approach , 2019 ; arXiv:1907.11692. Liu, Y. et al. RoBERTa: A robustly optimized BERT pretraining approach, 2019; arXiv:1907.11692."},{"key":"e_1_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"e_1_2_1_30_1","volume-title":"et al. Exploring the limits of weakly supervised pretraining","author":"Mahajan D.","year":"2018","unstructured":"Mahajan , D. et al. Exploring the limits of weakly supervised pretraining , 2018 ; arXiv:1805.00932. Mahajan, D. et al. Exploring the limits of weakly supervised pretraining, 2018; arXiv:1805.00932."},{"key":"e_1_2_1_31_1","volume-title":"Proceedings of EMNLP","author":"Melis G.","year":"2018","unstructured":"Melis , G. , Dyer , C. and Blunsom , P . On the state of the art of evaluation in neural language models . In Proceedings of EMNLP , 2018 . Melis, G., Dyer, C. and Blunsom, P. On the state of the art of evaluation in neural language models. In Proceedings of EMNLP, 2018."},{"key":"e_1_2_1_32_1","volume-title":"Proceedings of ICLR","author":"Molchanov P.","year":"2017","unstructured":"Molchanov , P. , Tyree , S. , Karras , T. , Aila , T. and Kautz , J . Pruning convolutional neural networks for resource efficient inference . In Proceedings of ICLR , 2017 . Molchanov, P., Tyree, S., Karras, T., Aila, T. and Kautz, J. Pruning convolutional neural networks for resource efficient inference. In Proceedings of ICLR, 2017."},{"key":"e_1_2_1_33_1","volume-title":"Cramming more components onto integrated circuits","author":"Moore G.E.","year":"1965","unstructured":"Moore , G.E. Cramming more components onto integrated circuits , 1965 . Moore, G.E. Cramming more components onto integrated circuits, 1965."},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"key":"e_1_2_1_35_1","volume-title":"OpenAI Blog","author":"Radford A.","year":"2019","unstructured":"Radford , A. , Wu , J. , Child , R. , Luan , D. , Amodei , D. and Sutskever , I . Language models are unsupervised multitask learners .. OpenAI Blog , 2019 . Radford, A., Wu, J., Child, R., Luan, D., Amodei, D. and Sutskever, I. Language models are unsupervised multitask learners.. OpenAI Blog, 2019."},{"key":"e_1_2_1_36_1","volume-title":"et al. Exploring the limits of transfer learning with a unified text-to-text transformer","author":"Raffel C.","year":"2019","unstructured":"Raffel , C. et al. Exploring the limits of transfer learning with a unified text-to-text transformer , 2019 ; arXiv:1910.10683. Raffel, C. et al. Exploring the limits of transfer learning with a unified text-to-text transformer, 2019; arXiv:1910.10683."},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.91"},{"key":"e_1_2_1_39_1","volume-title":"et al. Tackling climate change with machine learning","author":"Rolnick D.","year":"2019","unstructured":"Rolnick , D. et al. Tackling climate change with machine learning , 2019 ; arXiv:1905.12616. Rolnick, D. et al. Tackling climate change with machine learning, 2019; arXiv:1905.12616."},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"e_1_2_1_41_1","volume-title":"Proceedings of ACL","author":"Schwartz R.","year":"2018","unstructured":"Schwartz , R. , Thomson , S. and Smith , N.A . SoPa: Bridging CNNs, RNNs, and weighted finite-state machines . In Proceedings of ACL , 2018 . Schwartz, R., Thomson, S. and Smith, N.A. SoPa: Bridging CNNs, RNNs, and weighted finite-state machines. In Proceedings of ACL, 2018."},{"key":"e_1_2_1_42_1","volume-title":"Megatron-LM: Training multi-billion parameter language models using GPU model parallelism","author":"Shoeybi M.","year":"2019","unstructured":"Shoeybi , M. , Patwary , M. , Puri , R. , LeGresley , P. , Casper , J. , Catanzaro , B. Megatron-LM: Training multi-billion parameter language models using GPU model parallelism , 2019 ; arXiv:1909.08053. Shoeybi, M., Patwary, M., Puri, R., LeGresley, P., Casper, J., Catanzaro, B. Megatron-LM: Training multi-billion parameter language models using GPU model parallelism, 2019; arXiv:1909.08053."},{"key":"e_1_2_1_43_1","volume-title":"Human-Centered AI Initiative","author":"Shoham Y.","year":"2018","unstructured":"Shoham , Y. et al. The AI index 2018 annual report. AI Index Steering Committee , Human-Centered AI Initiative , Stanford University ; http:\/\/cdn.aiindex.org\/ 2018 \/AI%20Index%202018%20Annual%20Report.pdf. Shoham, Y. et al. The AI index 2018 annual report. AI Index Steering Committee, Human-Centered AI Initiative, Stanford University; http:\/\/cdn.aiindex.org\/2018\/AI%20Index%202018%20Annual%20Report.pdf."},{"key":"e_1_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature16961"},{"key":"e_1_2_1_45_1","volume-title":"et al. Mastering chess and shogi by self-play with a general reinforcement learning algorithm","author":"Silver D.","year":"2017","unstructured":"Silver , D. et al. Mastering chess and shogi by self-play with a general reinforcement learning algorithm , 2017 ; arXiv:1712.01815. Silver, D. et al. Mastering chess and shogi by self-play with a general reinforcement learning algorithm, 2017; arXiv:1712.01815."},{"key":"e_1_2_1_46_1","first-page":"7676","volume":"550","author":"Silver D.","year":"2017","unstructured":"Silver , D. Mastering the game of Go without human knowledge. Nature 550 , 7676 ( 2017 ), 354. Silver, D. et al. Mastering the game of Go without human knowledge. Nature 550, 7676 (2017), 354.","journal-title":"Mastering the game of Go without human knowledge. Nature"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1355"},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.97"},{"key":"e_1_2_1_49_1","volume-title":"Core vector machines: Fast SVM training on very large data sets. JMLR 6 (Apr","author":"Tsang I.","year":"2005","unstructured":"Tsang , I. , Kwok , J.T. and Cheung , P.M . Core vector machines: Fast SVM training on very large data sets. JMLR 6 (Apr . 2005 ), 363--392. Tsang, I., Kwok, J.T. and Cheung, P.M. Core vector machines: Fast SVM training on very large data sets. JMLR 6 (Apr. 2005), 363--392."},{"key":"e_1_2_1_50_1","volume-title":"Proceedings of NeurIPS","author":"Vaswani A.","year":"2017","unstructured":"Vaswani , A. , Shazeer , N. , Parmar , N. , Uszkoreit , J. , Jones , L. , Gomez , A.N. , Kaiser , L. and Polosukhin , I . Attention is all you need . In Proceedings of NeurIPS , 2017 . Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L. and Polosukhin, I. Attention is all you need. In Proceedings of NeurIPS, 2017."},{"key":"e_1_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00368"},{"key":"e_1_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8968165"},{"key":"e_1_2_1_53_1","volume-title":"SuperGLUE: A stickier benchmark for general-purpose language understanding systems","author":"Wang A.","year":"2019","unstructured":"Wang , A. Pruksachatkun , Y. , Nangia , N. , Singh , A. , Michael , J. , Hill , F. , Levy , O. and Bowman , S.R . SuperGLUE: A stickier benchmark for general-purpose language understanding systems , 2019 ; arXiv:1905.00537. Wang, A. Pruksachatkun, Y., Nangia, N., Singh, A., Michael, J., Hill, F., Levy, O. and Bowman, S.R. SuperGLUE: A stickier benchmark for general-purpose language understanding systems, 2019; arXiv:1905.00537."},{"key":"e_1_2_1_54_1","volume-title":"Proceedings of ICLR","author":"Wang A.","year":"2019","unstructured":"Wang , A. , Singh , A. , Michael , J. , Hill , F. , Levy , O. and Bowman , S.R . GLUE: A multi-task benchmark and analysis platform for natural language understanding . In Proceedings of ICLR , 2019 . Wang, A., Singh, A., Michael, J., Hill, F., Levy, O. and Bowman, S.R. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of ICLR, 2019."},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.634"},{"key":"e_1_2_1_56_1","volume-title":"XLNet: Generalized autoregressive pretraining for language understanding","author":"Yang Z.","year":"2019","unstructured":"Yang , Z. , Dai , Z. , Yang , Y. , Carbonell , J. , Salakhutdinov , R. and Le , Q.V . XLNet: Generalized autoregressive pretraining for language understanding , 2019 ; arXiv:1906.08237. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R. and Le, Q.V. XLNet: Generalized autoregressive pretraining for language understanding, 2019; arXiv:1906.08237."},{"key":"e_1_2_1_57_1","volume-title":"Defending against neural fake news","author":"Zellers R.","year":"2019","unstructured":"Zellers , R. , Holtzman , A. , Rashkin , H. , Bisk , Y. , Farhadi , A. , Roesner , F. and Choi , Y . Defending against neural fake news , 2019 ; arXiv:1905.12616. Zellers, R., Holtzman, A., Rashkin, H., Bisk, Y., Farhadi, A., Roesner, F. and Choi, Y. Defending against neural fake news, 2019; arXiv:1905.12616."},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00716"},{"key":"e_1_2_1_59_1","volume-title":"Proceedings of ICLR","author":"Zoph B.","year":"2017","unstructured":"Zoph , B. and Le , Q.V . Neural architecture search with reinforcement learning . In Proceedings of ICLR , 2017 . Zoph, B. and Le, Q.V. Neural architecture search with reinforcement learning. In Proceedings of ICLR, 2017."}],"container-title":["Communications of the ACM"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3381831","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3381831","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:33:07Z","timestamp":1750199587000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3381831"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,11,17]]},"references-count":59,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2020,11,17]]}},"alternative-id":["10.1145\/3381831"],"URL":"https:\/\/doi.org\/10.1145\/3381831","relation":{},"ISSN":["0001-0782","1557-7317"],"issn-type":[{"value":"0001-0782","type":"print"},{"value":"1557-7317","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,11,17]]},"assertion":[{"value":"2020-11-17","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}