{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T08:27:48Z","timestamp":1760171268523,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":19,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,7,8]],"date-time":"2022-07-08T00:00:00Z","timestamp":1657238400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,7,8]]},"DOI":"10.1145\/3491418.3530761","type":"proceedings-article","created":{"date-parts":[[2022,7,8]],"date-time":"2022-07-08T16:36:23Z","timestamp":1657298183000},"page":"1-7","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["A Fully Automated Scratch Storage Cleanup Tool for Heterogeneous Parallel Filesystems"],"prefix":"10.1145","author":[{"given":"Fang","family":"Liu","sequence":"first","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Dan","family":"Zhou","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Ken","family":"Suda","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Michael D.","family":"Weiner","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Mehmet","family":"Belgin","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Ruben","family":"Lara","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Pam","family":"Buffington","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,7,8]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2022. The Bourne Again SHell - GPN Project\u2019s shell. https:\/\/www.gnu.org\/software\/bash\/  2022. The Bourne Again SHell - GPN Project\u2019s shell. https:\/\/www.gnu.org\/software\/bash\/"},{"key":"e_1_3_2_1_2_1","unstructured":"2022. DDN Exascaler - Advanced Data Solutions for Modern Workloads. https:\/\/www.ddn.com\/products\/lustre-file-system-exascaler\/  2022. DDN Exascaler - Advanced Data Solutions for Modern Workloads. https:\/\/www.ddn.com\/products\/lustre-file-system-exascaler\/"},{"key":"e_1_3_2_1_3_1","unstructured":"2022. IBM Spectrum Scale - General Parallel File System (GPFS). https:\/\/en.wikipedia.org\/wiki\/GPFS  2022. IBM Spectrum Scale - General Parallel File System (GPFS). https:\/\/en.wikipedia.org\/wiki\/GPFS"},{"key":"e_1_3_2_1_4_1","unstructured":"2022. Lustre : A Scalable High-Performance File System Cluster. https:\/\/cse.buffalo.edu\/faculty\/tkosar\/cse710\/papers\/lustre-whitepaper.pdf  2022. Lustre : A Scalable High-Performance File System Cluster. https:\/\/cse.buffalo.edu\/faculty\/tkosar\/cse710\/papers\/lustre-whitepaper.pdf"},{"key":"e_1_3_2_1_5_1","unstructured":"2022. Quota-Info. https:\/\/docs.google.com\/spreadsheets\/d\/1O1heNM84wT4WUm4NzKzXAVM52G56b2k8RMNMcdt_76Q\/edit#gid=0  2022. Quota-Info. https:\/\/docs.google.com\/spreadsheets\/d\/1O1heNM84wT4WUm4NzKzXAVM52G56b2k8RMNMcdt_76Q\/edit#gid=0"},{"key":"e_1_3_2_1_6_1","unstructured":"2022. Robinhood Policy Engine. https:\/\/wiki.lustre.org\/Robinhood_Policy_Engine  2022. Robinhood Policy Engine. https:\/\/wiki.lustre.org\/Robinhood_Policy_Engine"},{"key":"e_1_3_2_1_7_1","unstructured":"National Energy Research Scientific\u00a0Computing Center. 2022. NERSC Scratch Purge policy. https:\/\/docs.nersc.gov\/filesystems\/quotas\/  National Energy Research Scientific\u00a0Computing Center. 2022. NERSC Scratch Purge policy. https:\/\/docs.nersc.gov\/filesystems\/quotas\/"},{"key":"e_1_3_2_1_8_1","volume-title":"Using Robinhood to Purge Data from Lustre File Systems","author":"Declerck M.","year":"2014","unstructured":"Tina\u00a0 M. Declerck . 2014. Using Robinhood to Purge Data from Lustre File Systems . In Cray User Group . https:\/\/cug.org\/proceedings\/cug 2014 _proceedings\/includes\/files\/pap157-file2.pdf Tina\u00a0M. Declerck. 2014. Using Robinhood to Purge Data from Lustre File Systems. In Cray User Group. https:\/\/cug.org\/proceedings\/cug2014_proceedings\/includes\/files\/pap157-file2.pdf"},{"key":"e_1_3_2_1_9_1","unstructured":"Oak Ridge\u00a0National Laboratory. 2022. Oak Ridge Leadership Computing Facility - Data retention policy. https:\/\/docs.olcf.ornl.gov\/accounts\/olcf_policy_guide.html#retention-policy  Oak Ridge\u00a0National Laboratory. 2022. Oak Ridge Leadership Computing Facility - Data retention policy. https:\/\/docs.olcf.ornl.gov\/accounts\/olcf_policy_guide.html#retention-policy"},{"key":"e_1_3_2_1_10_1","volume-title":"An Integrated Scratch Management Service for HPC Centers. In 2011 IEEE International Symposium on Parallel and Distributed Processing Workshops and Phd Forum. 2081\u20132084","author":"Monti M.","year":"2011","unstructured":"Henry\u00a0 M. Monti . 2011 . An Integrated Scratch Management Service for HPC Centers. In 2011 IEEE International Symposium on Parallel and Distributed Processing Workshops and Phd Forum. 2081\u20132084 . https:\/\/doi.org\/10.1109\/IPDPS.2011.382 10.1109\/IPDPS.2011.382 Henry\u00a0M. Monti. 2011. An Integrated Scratch Management Service for HPC Centers. In 2011 IEEE International Symposium on Parallel and Distributed Processing Workshops and Phd Forum. 2081\u20132084. https:\/\/doi.org\/10.1109\/IPDPS.2011.382"},{"key":"e_1_3_2_1_11_1","unstructured":"NCAR. 2022. National Center For Atmospheric Research. https:\/\/ncar.ucar.edu  NCAR. 2022. National Center For Atmospheric Research. https:\/\/ncar.ucar.edu"},{"key":"e_1_3_2_1_12_1","unstructured":"NCAR. 2022. NCAR Advance Research Computing Scratch file space purge policy. https:\/\/arc.ucar.edu\/knowledge_base\/68878466#GLADEfilespaces-scratchScratchfilespace  NCAR. 2022. NCAR Advance Research Computing Scratch file space purge policy. https:\/\/arc.ucar.edu\/knowledge_base\/68878466#GLADEfilespaces-scratchScratchfilespace"},{"key":"e_1_3_2_1_13_1","unstructured":"The\u00a0University of Utah. 2022. CHPC - Research Computing Support for the University Scratch disk space policy. https:\/\/www.chpc.utah.edu\/documentation\/policies\/3.1FileStoragePolicies.php  The\u00a0University of Utah. 2022. CHPC - Research Computing Support for the University Scratch disk space policy. https:\/\/www.chpc.utah.edu\/documentation\/policies\/3.1FileStoragePolicies.php"},{"key":"e_1_3_2_1_14_1","unstructured":"PACE. 2022. PACE Hive Gateway Portal. https:\/\/gateway.hive.pace.gatech.edu\/  PACE. 2022. PACE Hive Gateway Portal. https:\/\/gateway.hive.pace.gatech.edu\/"},{"key":"e_1_3_2_1_15_1","unstructured":"TACC. 2022. TACC Stampede2 Scratch file system purge policy. https:\/\/portal.tacc.utexas.edu\/user-guides\/stampede2#scratch-file-system-purge-policy  TACC. 2022. TACC Stampede2 Scratch file system purge policy. https:\/\/portal.tacc.utexas.edu\/user-guides\/stampede2#scratch-file-system-purge-policy"},{"key":"e_1_3_2_1_16_1","unstructured":"TACC. 2022. Texas Advanced Computing Center. https:\/\/www.tacc.utexas.edu\/  TACC. 2022. Texas Advanced Computing Center. https:\/\/www.tacc.utexas.edu\/"},{"key":"e_1_3_2_1_17_1","unstructured":"Harvard University. 2022. Harvard Faculty of Arts & Sciences Research Computing Scratch file system retention policy. https:\/\/docs.rc.fas.harvard.edu\/kb\/running-jobs\/  Harvard University. 2022. Harvard Faculty of Arts & Sciences Research Computing Scratch file system retention policy. https:\/\/docs.rc.fas.harvard.edu\/kb\/running-jobs\/"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2567656"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458817.3476201"}],"event":{"name":"PEARC '22: Practice and Experience in Advanced Research Computing","sponsor":["SIGAPP ACM Special Interest Group on Applied Computing","SIGHPC ACM Special Interest Group on High Performance Computing, Special Interest Group on High Performance Computing"],"location":"Boston MA USA","acronym":"PEARC '22"},"container-title":["Practice and Experience in Advanced Research Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3491418.3530761","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3491418.3530761","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:48Z","timestamp":1750188648000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3491418.3530761"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,7,8]]},"references-count":19,"alternative-id":["10.1145\/3491418.3530761","10.1145\/3491418"],"URL":"https:\/\/doi.org\/10.1145\/3491418.3530761","relation":{},"subject":[],"published":{"date-parts":[[2022,7,8]]},"assertion":[{"value":"2022-07-08","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}