I deleted/moved 351 GB in mc_production. The following things are owned by other users though, and ~6 GB more can be freed up if no longer needed: Luca: 762M: /work/hallb/hps/mc_production/Luca/lhe/Vegas_10_10_2016/ Holly: 3.1G: /work/hallb/hps/mc_production/MG5/dst/ 2G: /work/hallb/hps/mc_production/postTriSummitFixes/tritrig/1pt05/NOSUMCUT/ 25M: /work/hallb/hps/mc_production/logs/slic/ap/ 16M: /work/hallb/hps/mc_production/logs/readout/beam-tri/1pt92/ Matt G: 35M: /work/hallb/hps/mc_production/dst/ 11M: /work/hallb/hps/mc_production/logs/dqm/ There is also 2 GB of old 2.2 GeV A' MC, which should no longer be relevant, but I didn't want to do anything with it since it had mock data stuff in there: /work/hallb/hps/mc_production/lhe/ap/2pt2/ ________________________________ From: [log in to unmask] <[log in to unmask]> on behalf of Nathan Baltzell <[log in to unmask]> Sent: Thursday, June 1, 2017 1:23:19 PM To: HPS-SOFTWARE Cc: [log in to unmask] Subject: Re: [Hps-analysis] Fwd: [Clas_offline] Fwd: ENP consumption of disk space under /work Here’s the most relevant usage 649G mrsolt/ 570G sebouh/ 459G mc_production/ 228G holly 159G mccaky/ 78G rafopar/ 45G omoreno/ 44G spaul 39G fxgirod 34G jeremym data/engrun2015: 3.2T tweakpass6 50G tweakpass6fail 64G tpass7 2.4G tpass7b 39G tpass7c 6.5G t_tweakpass_a 373G pass6/skim 201G pass6/dst data/physrun2016: 3.5T pass0 690G feeiter4 94M feeiter0 327M feeiter1 339M feeiter2 338M feeiter3 15G noPass 24G pass0_allign 52G pass0fail 4.5G tmp_test 281G tpass1 11G upass0 On Jun 1, 2017, at 11:05, Stepan Stepanyan <[log in to unmask]> wrote: > FYI, we need to move files. > > Stepan > >> Begin forwarded message: >> >> From: Harut Avakian <[log in to unmask]> >> Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work >> Date: June 1, 2017 at 5:01:24 PM GMT+2 >> To: "[log in to unmask]" <[log in to unmask]> >> >> >> >> >> Dear All, >> >> As you can see from the e-mail below, keeping all our work disk space requires some additional funding. >> Option 3 will inevitably impact on farm operations, removing of ~20% space from Lustre. >> >> We can also choose something between options 1) and 3). >> Please revise the content and move at least 75% of what is in /work/clas to either /cache or /volatile? >> The current Hall-B usage includes: >> 550G hallb/bonus >> 1.5T hallb/clase1 >> 3.6T hallb/clase1-6 >> 3.3T hallb/clase1dvcs >> 2.8T hallb/clase1dvcs2 >> 987G hallb/clase1f >> 1.8T hallb/clase2 >> 1.6G hallb/clase5 >> 413G hallb/clase6 >> 2.2T hallb/claseg1 >> 3.9T hallb/claseg1dvcs >> 1.2T hallb/claseg3 >> 4.1T hallb/claseg4 >> 2.7T hallb/claseg5 >> 1.7T hallb/claseg6 >> 367G hallb/clas-farm-output >> 734G hallb/clasg10 >> 601G hallb/clasg11 >> 8.1T hallb/clasg12 >> 2.4T hallb/clasg13 >> 2.4T hallb/clasg14 >> 28G hallb/clasg3 >> 5.8G hallb/clasg7 >> 269G hallb/clasg8 >> 1.2T hallb/clasg9 >> 1.3T hallb/clashps >> 1.8T hallb/clas-production >> 5.6T hallb/clas-production2 >> 1.4T hallb/clas-production3 >> 12T hallb/hps >> 13T hallb/prad >> >> Regards, >> >> Harut >> >> P.S. Few times we had crashes and they may also happen in future, so keeping important files in /work is not recommended. >> You can see the list of lost files in /site/scicomp/lostfiles.txt and /site/scicomp/lostfiles-jan-2017.txt >> >> >> >> -------- Forwarded Message -------- >> Subject: ENP consumption of disk space under /work >> Date: Wed, 31 May 2017 10:35:51 -0400 >> From: Chip Watson <[log in to unmask]> >> To: Sandy Philpott <[log in to unmask]>, Graham Heyes <[log in to unmask]>, Ole Hansen <[log in to unmask]>, Harut Avakian <[log in to unmask]>, Brad Sawatzky <[log in to unmask]>, Mark M. Ito <[log in to unmask]> >> >> All, >> >> As I have started on the procurement of the new /work file server, I >> have discovered that Physics' use of /work has grown unrestrained over >> the last year or two. >> >> "Unrestrained" because there is no way under Lustre to restrain it >> except via a very unfriendly Lustre quota system. As we leave some >> quota headroom to accommodate large swings in usage for each hall for >> cache and volatile, then /work continues to grow. >> >> Total /work has now reached 260 TB, several times larger than I was >> anticipating. This constitutes more than 25% of Physics' share of >> Lustre, compared to LQCD which uses less than 5% of its disk space on >> the un-managed /work. >> >> It would cost Physics an extra $25K (total $35K - $40K) to treat the 260 >> TB as a requirement. >> >> There are 3 paths forward: >> >> (1) Physics cuts its use of /work by a factor of 4-5. >> (2) Physics increases funding to $40K >> (3) We pull a server out of Lustre, decreasing Physics' share of the >> system, and use that as half of the new active-active pair, beefing it >> up with SSDs and perhaps additional memory; this would actually shrink >> Physics near term costs, but puts higher pressure on the file system for >> the farm >> >> The decision is clearly Physics', but I do need a VERY FAST response to >> this question, as I need to move quickly now for LQCD's needs. >> >> Hall D + GlueX, 96 TB >> CLAS + CLAS12, 98 TB >> Hall C, 35 TB >> Hall A <unknown, still scanning> >> >> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion. >> >> thanks, >> Chip >> >> >> _______________________________________________ >> Clas_offline mailing list >> [log in to unmask] >> https://mailman.jlab.org/mailman/listinfo/clas_offline > > _______________________________________________ > Hps-analysis mailing list > [log in to unmask] > https://mailman.jlab.org/mailman/listinfo/hps-analysis ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the HPS-SOFTWARE list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the HPS-SOFTWARE list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1