{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:31:20Z","timestamp":1750221080768,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":62,"publisher":"ACM","license":[{"start":{"date-parts":[[2018,11,1]],"date-time":"2018-11-01T00:00:00Z","timestamp":1541030400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"the National Research Foundation of Korea","award":["NRF-2016M3C4A7952587"],"award-info":[{"award-number":["NRF-2016M3C4A7952587"]}]},{"name":"Samsung Advanced Institute of Technology","award":["0000"],"award-info":[{"award-number":["0000"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2018,11]]},"DOI":"10.1145\/3243176.3243180","type":"proceedings-article","created":{"date-parts":[[2018,10,10]],"date-time":"2018-10-10T13:32:32Z","timestamp":1539178352000},"page":"1-14","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":19,"title":["A portable, automatic data qantizer for deep neural networks"],"prefix":"10.1145","author":[{"given":"Young H.","family":"Oh","sequence":"first","affiliation":[{"name":"Sungkyunkwan University, Suwon, Korea"}]},{"given":"Quan","family":"Quan","sequence":"additional","affiliation":[{"name":"Seoul National University, Seoul, Korea"}]},{"given":"Daeyeon","family":"Kim","sequence":"additional","affiliation":[{"name":"Seoul National University, Seoul, Korea"}]},{"given":"Seonghak","family":"Kim","sequence":"additional","affiliation":[{"name":"Seoul National University, Seoul, Korea"}]},{"given":"Jun","family":"Heo","sequence":"additional","affiliation":[{"name":"Seoul National University, Seoul, Korea"}]},{"given":"Sungjun","family":"Jung","sequence":"additional","affiliation":[{"name":"Seoul National University, Seoul, Korea"}]},{"given":"Jaeyoung","family":"Jang","sequence":"additional","affiliation":[{"name":"Sungkyunkwan University, Suwon, Korea"}]},{"given":"Jae W.","family":"Lee","sequence":"additional","affiliation":[{"name":"Seoul National University, Seoul, Korea"}]}],"member":"320","published-online":{"date-parts":[[2018,11]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3123939.3123982"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2628071.2628092"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/1375581.1375595"},{"key":"e_1_3_2_1_4_1","volume-title":"Microsoft unveils Project Brainwave for real-time AI. https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-unveils-project-brainwave\/","author":"Burger Doug","year":"2017","unstructured":"Doug Burger . 2017. Microsoft unveils Project Brainwave for real-time AI. https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-unveils-project-brainwave\/ . 2017 . Doug Burger. 2017. Microsoft unveils Project Brainwave for real-time AI. https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-unveils-project-brainwave\/. 2017."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO.2014.58"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"E. Chung J. Fowers K. Ovtcharov M. Papamichael A. Caulfield T. Massengill M. Liu D. Lo S. Alkalay M. Haselman M. Abeydeera L. Adams H. Angepat C. Boehn D. Chiou O. Firestein A. Forin K. S. Gatlin M. Ghandi S. Heil K. Holohan A. El Husseini T. Juhasz K. Kagi R. Kovvuri S. Lanka F. van Megen D. Mukhortov P. Patel B. Perez A. Rapsang S. Reinhardt B. Rouhani A. Sapek R. Seera S. Shekar B. Sridharan G. Weisz L. Woods P. Yi Xiao D. Zhang R. Zhao and D. Burger. 2018. Serving DNNs in real time at datacenter scale with Project Brainwave. IEEE Micro 2018.  E. Chung J. Fowers K. Ovtcharov M. Papamichael A. Caulfield T. Massengill M. Liu D. Lo S. Alkalay M. Haselman M. Abeydeera L. Adams H. Angepat C. Boehn D. Chiou O. Firestein A. Forin K. S. Gatlin M. Ghandi S. Heil K. Holohan A. El Husseini T. Juhasz K. Kagi R. Kovvuri S. Lanka F. van Megen D. Mukhortov P. Patel B. Perez A. Rapsang S. Reinhardt B. Rouhani A. Sapek R. Seera S. Shekar B. Sridharan G. Weisz L. Woods P. Yi Xiao D. Zhang R. Zhao and D. Burger. 2018. Serving DNNs in real time at datacenter scale with Project Brainwave. IEEE Micro 2018.","DOI":"10.1109\/MM.2018.022071131"},{"key":"e_1_3_2_1_7_1","volume-title":"Training deep neural networks with low precision multiplications","author":"Courbariaux Matthieu","year":"2014","unstructured":"Matthieu Courbariaux , Yoshua Bengio , and Jean-Pierre David . 2014. Training deep neural networks with low precision multiplications . 2014 . arXiv:1412.7024. Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David. 2014. Training deep neural networks with low precision multiplications. 2014. arXiv:1412.7024."},{"key":"e_1_3_2_1_8_1","volume-title":"Binarized Neural Networks: Training deep neural networks with weights and activations constrained to +1 or -1","author":"Courbariaux Matthieu","year":"2016","unstructured":"Matthieu Courbariaux , Itay Hubara , Daniel Soudry , Ran El-Yaniv , and Yoshua Bengio . 2016. Binarized Neural Networks: Training deep neural networks with weights and activations constrained to +1 or -1 . 2016 . arXiv:1602.02830. Matthieu Courbariaux, Itay Hubara, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2016. Binarized Neural Networks: Training deep neural networks with weights and activations constrained to +1 or -1. 2016. arXiv:1602.02830."},{"volume-title":"2017 26th International Conference on Parallel Architectures and Compilation Techniques (PACT '17)","author":"Cummins C.","key":"e_1_3_2_1_9_1","unstructured":"C. Cummins , P. Petoumenos , Z. Wang , and H. Leather . 2017. End-to-end deep learning of optimization heuristics . In 2017 26th International Conference on Parallel Architectures and Compilation Techniques (PACT '17) . C. Cummins, P. Petoumenos, Z. Wang, and H. Leather. 2017. End-to-end deep learning of optimization heuristics. In 2017 26th International Conference on Parallel Architectures and Compilation Techniques (PACT '17)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_1_11_1","volume-title":"Coleman et al","author":"Cody","year":"2017","unstructured":"Cody A. Coleman et al . 2017 . DAWNBench : An end-to-end deep learning benchmark and competition. https:\/\/github.com\/stanford-futuredata\/dawn-bench-entries. 2017. Cody A. Coleman et al. 2017. DAWNBench: An end-to-end deep learning benchmark and competition. https:\/\/github.com\/stanford-futuredata\/dawn-bench-entries. 2017."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"G. Hinton et al. 2012. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal Processing Magazine 2012.  G. Hinton et al. 2012. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal Processing Magazine 2012.","DOI":"10.1109\/MSP.2012.2205597"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3079856.3080246"},{"key":"e_1_3_2_1_14_1","unstructured":"Paulius Micikevicius et al. 2017. Mixed precision training. 2017. arXiv:1710.03740.  Paulius Micikevicius et al. 2017. Mixed precision training. 2017. arXiv:1710.03740."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"e_1_3_2_1_16_1","volume-title":"Human-level control through deep reinforcement learning. Nature","author":"Volodymyr Mnih","year":"2015","unstructured":"Volodymyr Mnih et. al. 2015. Human-level control through deep reinforcement learning. Nature 2015 . Volodymyr Mnih et. al. 2015. Human-level control through deep reinforcement learning. Nature 2015."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-009-0275-4"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2018.00012"},{"key":"e_1_3_2_1_19_1","volume-title":"Hardware-oriented approximation of convolutional neural networks","author":"Gysel Philipp","year":"2016","unstructured":"Philipp Gysel , Mohammad Motamedi , and Soheil Ghiasi . 2016. Hardware-oriented approximation of convolutional neural networks . 2016 . arXiv:1604.03168. Philipp Gysel, Mohammad Motamedi, and Soheil Ghiasi. 2016. Hardware-oriented approximation of convolutional neural networks. 2016. arXiv:1604.03168."},{"key":"e_1_3_2_1_20_1","volume-title":"Ristretto: A framework for empirical study of resource-efficient inference in convolutional neural networks","author":"Gysel Philipp","year":"2018","unstructured":"Philipp Gysel , Jon Pimentel , Mohammad Motamedi , and Soheil Ghiasi . 2018 . Ristretto: A framework for empirical study of resource-efficient inference in convolutional neural networks . IEEE Transactions on Neural Networks and Learning Systems . 2018. Philipp Gysel, Jon Pimentel, Mohammad Motamedi, and Soheil Ghiasi. 2018. Ristretto: A framework for empirical study of resource-efficient inference in convolutional neural networks. IEEE Transactions on Neural Networks and Learning Systems. 2018."},{"key":"e_1_3_2_1_21_1","volume":"201","author":"Han Song","unstructured":"Song Han , Huizi Mao , and William J. Dally. 201 5. Deep Compression: Compressing deep neural network with pruning, trained quantization and Huffman coding. 2015. arXiv:1510.00149. Song Han, Huizi Mao, and William J. Dally. 2015. Deep Compression: Compressing deep neural network with pruning, trained quantization and Huffman coding. 2015. arXiv:1510.00149.","journal-title":"William J. Dally."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_23_1","volume-title":"Xilinx Kintex UltraSCALE FPGA Family https:\/\/www.xilinx.com\/products\/silicon-devices\/fpga\/kintex-ultrascale.html","author":"Xilinx INC.","year":"2018","unstructured":"Xilinx INC. 2018. Xilinx Kintex UltraSCALE FPGA Family https:\/\/www.xilinx.com\/products\/silicon-devices\/fpga\/kintex-ultrascale.html . 2018 . Xilinx INC. 2018. Xilinx Kintex UltraSCALE FPGA Family https:\/\/www.xilinx.com\/products\/silicon-devices\/fpga\/kintex-ultrascale.html. 2018."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/2647868.2654889"},{"key":"e_1_3_2_1_25_1","volume-title":"Raquel Urtasun, and Andreas Moshovos.","author":"Judd Patrick","year":"2015","unstructured":"Patrick Judd , Jorge Albericio , Tayler Hetherington , Tor Aamodt , Natalie Enright Jerger , Raquel Urtasun, and Andreas Moshovos. 2015 . Reduced-precision strategies for bounded memory in deep neural nets. 2015. arXiv:1511.05236. Patrick Judd, Jorge Albericio, Tayler Hetherington, Tor Aamodt, Natalie Enright Jerger, Raquel Urtasun, and Andreas Moshovos. 2015. Reduced-precision strategies for bounded memory in deep neural nets. 2015. arXiv:1511.05236."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/2925426.2926294"},{"volume-title":"2016 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO '16)","author":"Judd P.","key":"e_1_3_2_1_27_1","unstructured":"P. Judd , J. Albericio , T. Hetherington , T. M. Aamodt , and A. Moshovos . 2016. Stripes: Bit-serial deep neural network computing . In 2016 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO '16) . P. Judd, J. Albericio, T. Hetherington, T. M. Aamodt, and A. Moshovos. 2016. Stripes: Bit-serial deep neural network computing. In 2016 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO '16)."},{"key":"e_1_3_2_1_28_1","volume-title":"Compression of deep convolutional neural networks for fast and low power mobile applications","author":"Kim Yong-Deok","year":"2015","unstructured":"Yong-Deok Kim , Eunhyeok Park , Sungjoo Yoo , Taelim Choi , Lu Yang , and Dongjun Shin . 2015. Compression of deep convolutional neural networks for fast and low power mobile applications . 2015 . arXiv:1511.06530. Yong-Deok Kim, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu Yang, and Dongjun Shin. 2015. Compression of deep convolutional neural networks for fast and low power mobile applications. 2015. arXiv:1511.06530."},{"key":"e_1_3_2_1_29_1","volume-title":"cuda-convnet: High-performance c++\/cuda implementation of convolutional neural networks. https:\/\/code.google.com\/p\/cuda-convnet\/","author":"Krizhevsky Alex","year":"2012","unstructured":"Alex Krizhevsky . 2012. cuda-convnet: High-performance c++\/cuda implementation of convolutional neural networks. https:\/\/code.google.com\/p\/cuda-convnet\/ . 2012 . Alex Krizhevsky. 2012. cuda-convnet: High-performance c++\/cuda implementation of convolutional neural networks. https:\/\/code.google.com\/p\/cuda-convnet\/. 2012."},{"key":"e_1_3_2_1_30_1","volume-title":"Learning multiple layers of features from tiny images","author":"Krizhevsky Alex","year":"2009","unstructured":"Alex Krizhevsky and Geoffrey Hinton . 2009. Learning multiple layers of features from tiny images . 2009 . Alex Krizhevsky and Geoffrey Hinton. 2009. Learning multiple layers of features from tiny images. 2009."},{"key":"e_1_3_2_1_31_1","unstructured":"Alex Krizhevsky Ilya Sutskever and Geoffrey E Hinton. 2012. ImageNet classification with deep convolutional neural networks. In Advances in neural information processing systems.   Alex Krizhevsky Ilya Sutskever and Geoffrey E Hinton. 2012. ImageNet classification with deep convolutional neural networks. In Advances in neural information processing systems ."},{"key":"e_1_3_2_1_32_1","unstructured":"Yann LeCun. 1998. The MNIST database of handwritten digits. http:\/\/yann.lecun.com\/exdb\/mnist\/. 1998.  Yann LeCun. 1998. The MNIST database of handwritten digits. http:\/\/yann.lecun.com\/exdb\/mnist\/ . 1998."},{"key":"e_1_3_2_1_33_1","unstructured":"Yann LeCun etal 1998. LeNet-5 convolutional neural networks. 1998.  Yann LeCun et al. 1998. LeNet-5 convolutional neural networks. 1998."},{"key":"e_1_3_2_1_34_1","volume-title":"Proceedings of the 33rd International Conference on Machine Learning (ICML '16)","author":"Lin Darryl","year":"2016","unstructured":"Darryl Lin , Sachin Talathi , and Sreekanth Annapureddy . 2016 . Fixed point quantization of deep convolutional networks . In Proceedings of the 33rd International Conference on Machine Learning (ICML '16) . Darryl Lin, Sachin Talathi, and Sreekanth Annapureddy. 2016. Fixed point quantization of deep convolutional networks. In Proceedings of the 33rd International Conference on Machine Learning (ICML '16)."},{"key":"e_1_3_2_1_35_1","volume-title":"Network in network","author":"Lin Min","year":"2013","unstructured":"Min Lin , Qiang Chen , and Shuicheng Yan . 2013. Network in network . 2013 . arXiv:1312.4400. Min Lin, Qiang Chen, and Shuicheng Yan. 2013. Network in network. 2013. arXiv:1312.4400."},{"key":"e_1_3_2_1_36_1","volume-title":"Neural networks with few multiplications","author":"Lin Zhouhan","year":"2015","unstructured":"Zhouhan Lin , Matthieu Courbariaux , Roland Memisevic , and Yoshua Bengio . 2015. Neural networks with few multiplications . 2015 . arXiv:1510.03009. Zhouhan Lin, Matthieu Courbariaux, Roland Memisevic, and Yoshua Bengio. 2015. Neural networks with few multiplications. 2015. arXiv:1510.03009."},{"volume-title":"2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '11)","author":"Mikolov T.","key":"e_1_3_2_1_37_1","unstructured":"T. Mikolov , S. Kombrink , L. Burget , J. Cernocky , and S. Khudanpur . 2011. Extensions of recurrent neural network language model . In 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '11) . T. Mikolov, S. Kombrink, L. Burget, J. Cernocky, and S. Khudanpur. 2011. Extensions of recurrent neural network language model. In 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '11)."},{"key":"e_1_3_2_1_38_1","volume-title":"Convolutional neural networks using logarithmic data representation","author":"Miyashita Daisuke","year":"2016","unstructured":"Daisuke Miyashita , Edward H. Lee , and Boris Murmann . 2016. Convolutional neural networks using logarithmic data representation . 2016 . arXiv:1603.01025. Daisuke Miyashita, Edward H. Lee, and Boris Murmann. 2016. Convolutional neural networks using logarithmic data representation. 2016. arXiv:1603.01025."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.681"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/2847263.2847265"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/2491956.2462176"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/1993498.1993502"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/2254064.2254082"},{"key":"e_1_3_2_1_44_1","volume-title":"XNOR-Net: ImageNet classification using binary convolutional neural networks","author":"Rastegari Mohammad","year":"2016","unstructured":"Mohammad Rastegari , Vicente Ordonez , Joseph Redmon , and Ali Farhadi . 2016. XNOR-Net: ImageNet classification using binary convolutional neural networks . 2016 . arXiv:1603.05279. Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, and Ali Farhadi. 2016. XNOR-Net: ImageNet classification using binary convolutional neural networks. 2016. arXiv:1603.05279."},{"key":"e_1_3_2_1_45_1","unstructured":"Joseph Redmon. 2013-2016. Darknet: Open source neural networks in c. 2013-2016. http:\/\/pjreddie.com\/darknet\/  Joseph Redmon. 2013-2016. Darknet: Open source neural networks in c. 2013-2016. http:\/\/pjreddie.com\/darknet\/"},{"key":"e_1_3_2_1_46_1","volume-title":"YOLO9000: better, faster, stronger","author":"Redmon Joseph","year":"2016","unstructured":"Joseph Redmon and Ali Farhadi . 2016. YOLO9000: better, faster, stronger . 2016 . arXiv:1612.08242. Joseph Redmon and Ali Farhadi. 2016. YOLO9000: better, faster, stronger. 2016. arXiv:1612.08242."},{"key":"e_1_3_2_1_47_1","unstructured":"Mengye Ren Ryan Kiros and Richard Zemel. 2015. Exploring models and data for image question answering. In Advances in neural information processing systems.   Mengye Ren Ryan Kiros and Richard Zemel. 2015. Exploring models and data for image question answering. In Advances in neural information processing systems ."},{"key":"e_1_3_2_1_48_1","volume":"199","author":"Hochreiter","unstructured":"Hochreiter S. and Schmidhuberm J. 199 7. Long short-term memory. Neural Computation 1997. Hochreiter S. and Schmidhuberm J. 1997. Long short-term memory. Neural Computation 1997.","journal-title":"Schmidhuberm J."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"e_1_3_2_1_50_1","volume-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and Andrew Zisserman . 2014. Very deep convolutional networks for large-scale image recognition . 2014 . arXiv:1409.1556. Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. 2014. arXiv:1409.1556."},{"key":"e_1_3_2_1_51_1","volume-title":"Jason Weston, and Rob Fergus.","author":"Sukhbaatar Sainbayar","year":"2015","unstructured":"Sainbayar Sukhbaatar , arthur szlam , Jason Weston, and Rob Fergus. 2015 . End-to-end memory networks. In Advances in neural information processing systems. Sainbayar Sukhbaatar, arthur szlam, Jason Weston, and Rob Fergus. 2015. End-to-end memory networks. In Advances in neural information processing systems."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3061639.3062259"},{"volume-title":"TensorFlow mechanics 101. https:\/\/github.com\/tensorflow\/tensorflow\/tree\/r1.2\/tensorflow\/examples\/tutorials\/mnist","year":"2017","key":"e_1_3_2_1_53_1","unstructured":"TensorFlow\u2122. 2017. TensorFlow mechanics 101. https:\/\/github.com\/tensorflow\/tensorflow\/tree\/r1.2\/tensorflow\/examples\/tutorials\/mnist . 2017 . TensorFlow\u2122. 2017. TensorFlow mechanics 101. https:\/\/github.com\/tensorflow\/tensorflow\/tree\/r1.2\/tensorflow\/examples\/tutorials\/mnist. 2017."},{"volume-title":"TensorFlow: How to quantize neural networks with TensorFlow. https:\/\/www.tensorflow.org\/performance\/quantization","year":"2018","key":"e_1_3_2_1_54_1","unstructured":"TensorFlow\u2122. 2018. TensorFlow: How to quantize neural networks with TensorFlow. https:\/\/www.tensorflow.org\/performance\/quantization . 2018 . TensorFlow\u2122. 2018. TensorFlow: How to quantize neural networks with TensorFlow. https:\/\/www.tensorflow.org\/performance\/quantization. 2018."},{"key":"e_1_3_2_1_55_1","volume-title":"CIFAR-10 classifier. https:\/\/github.com\/vrakesh\/CIFAR-10-Classifier","author":"Vasudevan Rakesh","year":"2017","unstructured":"Rakesh Vasudevan . 2017. CIFAR-10 classifier. https:\/\/github.com\/vrakesh\/CIFAR-10-Classifier . 2017 . Rakesh Vasudevan. 2017. CIFAR-10 classifier. https:\/\/github.com\/vrakesh\/CIFAR-10-Classifier. 2017."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/2677036"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.257"},{"key":"e_1_3_2_1_58_1","volume-title":"Proceedings of the International Conference on Learning Representations (ICLR '15)","author":"Weston Jason","year":"2015","unstructured":"Jason Weston , Sumit Chopra , and Antoine Bordes . 2015 . Memory networks . In Proceedings of the International Conference on Learning Representations (ICLR '15) . Jason Weston, Sumit Chopra, and Antoine Bordes. 2015. Memory networks. In Proceedings of the International Conference on Learning Representations (ICLR '15)."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/2684746.2689060"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3020078.3021741"},{"key":"e_1_3_2_1_61_1","volume-title":"Incremental network quantization: Towards lossless CNNs with low-precision weights","author":"Zhou Aojun","year":"2017","unstructured":"Aojun Zhou , Anbang Yao , Yiwen Guo , Lin Xu , and Yurong Chen . 2017. Incremental network quantization: Towards lossless CNNs with low-precision weights . 2017 . arXiv:1702.03044. Aojun Zhou, Anbang Yao, Yiwen Guo, Lin Xu, and Yurong Chen. 2017. Incremental network quantization: Towards lossless CNNs with low-precision weights. 2017. arXiv:1702.03044."},{"key":"e_1_3_2_1_62_1","volume":"201","author":"Zhu Chenzhuo","unstructured":"Chenzhuo Zhu , Song Han , Huizi Mao , and William J. Dally. 201 6. Trained ternary quantization. 2016. arXiv:1612.01064. Chenzhuo Zhu, Song Han, Huizi Mao, and William J. Dally. 2016. Trained ternary quantization. 2016. arXiv:1612.01064.","journal-title":"William J. Dally."}],"event":{"name":"PACT '18: International conference on Parallel Architectures and Compilation Techniques","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture","IFIP WG 10.3 IFIP WG 10.3","IEEE CS"],"location":"Limassol Cyprus","acronym":"PACT '18"},"container-title":["Proceedings of the 27th International Conference on Parallel Architectures and Compilation Techniques"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3243176.3243180","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3243176.3243180","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T00:57:39Z","timestamp":1750208259000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3243176.3243180"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,11]]},"references-count":62,"alternative-id":["10.1145\/3243176.3243180","10.1145\/3243176"],"URL":"https:\/\/doi.org\/10.1145\/3243176.3243180","relation":{},"subject":[],"published":{"date-parts":[[2018,11]]},"assertion":[{"value":"2018-11-01","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}