I have reduced my /work/ usage to 16 MB. On Thu, Jun 1, 2017 at 3:05 PM, Bradley T Yale <[log in to unmask]> wrote: > I deleted/moved 351 GB in mc_production. > > > The following things are owned by other users though, and ~6 GB more can > be freed up if no longer needed: > > > Luca: > > 762M: /work/hallb/hps/mc_production/Luca/lhe/Vegas_10_10_2016/ > > > Holly: > 3.1G: /work/hallb/hps/mc_production/MG5/dst/ > > 2G: /work/hallb/hps/mc_production/postTriSummitFixes/tritrig/ > 1pt05/NOSUMCUT/ > > 25M: /work/hallb/hps/mc_production/logs/slic/ap/ > > 16M: /work/hallb/hps/mc_production/logs/readout/beam-tri/1pt92/ > > Matt G: > > 35M: /work/hallb/hps/mc_production/dst/ > > 11M: /work/hallb/hps/mc_production/logs/dqm/ > > > There is also 2 GB of old 2.2 GeV A' MC, which should no longer be > relevant, but I didn't want to do anything with it since it had mock data > stuff in there: > > /work/hallb/hps/mc_production/lhe/ap/2pt2/ > > > ------------------------------ > *From:* [log in to unmask] <[log in to unmask]> > on behalf of Nathan Baltzell <[log in to unmask]> > *Sent:* Thursday, June 1, 2017 1:23:19 PM > *To:* HPS-SOFTWARE > *Cc:* [log in to unmask] > *Subject:* Re: [Hps-analysis] Fwd: [Clas_offline] Fwd: ENP consumption of > disk space under /work > > Here’s the most relevant usage > > 649G mrsolt/ > 570G sebouh/ > 459G mc_production/ > 228G holly > 159G mccaky/ > 78G rafopar/ > 45G omoreno/ > 44G spaul > 39G fxgirod > 34G jeremym > > data/engrun2015: > 3.2T tweakpass6 > 50G tweakpass6fail > 64G tpass7 > 2.4G tpass7b > 39G tpass7c > 6.5G t_tweakpass_a > 373G pass6/skim > 201G pass6/dst > > data/physrun2016: > 3.5T pass0 > 690G feeiter4 > 94M feeiter0 > 327M feeiter1 > 339M feeiter2 > 338M feeiter3 > 15G noPass > 24G pass0_allign > 52G pass0fail > 4.5G tmp_test > 281G tpass1 > 11G upass0 > > > > > On Jun 1, 2017, at 11:05, Stepan Stepanyan <[log in to unmask]> wrote: > > > FYI, we need to move files. > > > > Stepan > > > >> Begin forwarded message: > >> > >> From: Harut Avakian <[log in to unmask]> > >> Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work > >> Date: June 1, 2017 at 5:01:24 PM GMT+2 > >> To: "[log in to unmask]" <[log in to unmask]> > >> > >> > >> > >> > >> Dear All, > >> > >> As you can see from the e-mail below, keeping all our work disk space > requires some additional funding. > >> Option 3 will inevitably impact on farm operations, removing of ~20% > space from Lustre. > >> > >> We can also choose something between options 1) and 3). > >> Please revise the content and move at least 75% of what is in > /work/clas to either /cache or /volatile? > >> The current Hall-B usage includes: > >> 550G hallb/bonus > >> 1.5T hallb/clase1 > >> 3.6T hallb/clase1-6 > >> 3.3T hallb/clase1dvcs > >> 2.8T hallb/clase1dvcs2 > >> 987G hallb/clase1f > >> 1.8T hallb/clase2 > >> 1.6G hallb/clase5 > >> 413G hallb/clase6 > >> 2.2T hallb/claseg1 > >> 3.9T hallb/claseg1dvcs > >> 1.2T hallb/claseg3 > >> 4.1T hallb/claseg4 > >> 2.7T hallb/claseg5 > >> 1.7T hallb/claseg6 > >> 367G hallb/clas-farm-output > >> 734G hallb/clasg10 > >> 601G hallb/clasg11 > >> 8.1T hallb/clasg12 > >> 2.4T hallb/clasg13 > >> 2.4T hallb/clasg14 > >> 28G hallb/clasg3 > >> 5.8G hallb/clasg7 > >> 269G hallb/clasg8 > >> 1.2T hallb/clasg9 > >> 1.3T hallb/clashps > >> 1.8T hallb/clas-production > >> 5.6T hallb/clas-production2 > >> 1.4T hallb/clas-production3 > >> 12T hallb/hps > >> 13T hallb/prad > >> > >> Regards, > >> > >> Harut > >> > >> P.S. Few times we had crashes and they may also happen in future, so > keeping important files in /work is not recommended. > >> You can see the list of lost files in /site/scicomp/lostfiles.txt and > /site/scicomp/lostfiles-jan-2017.txt > >> > >> > >> > >> -------- Forwarded Message -------- > >> Subject: ENP consumption of disk space under /work > >> Date: Wed, 31 May 2017 10:35:51 -0400 > >> From: Chip Watson <[log in to unmask]> > >> To: Sandy Philpott <[log in to unmask]>, Graham Heyes <[log in to unmask]>, > Ole Hansen <[log in to unmask]>, Harut Avakian <[log in to unmask]>, Brad > Sawatzky <[log in to unmask]>, Mark M. Ito <[log in to unmask]> > >> > >> All, > >> > >> As I have started on the procurement of the new /work file server, I > >> have discovered that Physics' use of /work has grown unrestrained over > >> the last year or two. > >> > >> "Unrestrained" because there is no way under Lustre to restrain it > >> except via a very unfriendly Lustre quota system. As we leave some > >> quota headroom to accommodate large swings in usage for each hall for > >> cache and volatile, then /work continues to grow. > >> > >> Total /work has now reached 260 TB, several times larger than I was > >> anticipating. This constitutes more than 25% of Physics' share of > >> Lustre, compared to LQCD which uses less than 5% of its disk space on > >> the un-managed /work. > >> > >> It would cost Physics an extra $25K (total $35K - $40K) to treat the > 260 > >> TB as a requirement. > >> > >> There are 3 paths forward: > >> > >> (1) Physics cuts its use of /work by a factor of 4-5. > >> (2) Physics increases funding to $40K > >> (3) We pull a server out of Lustre, decreasing Physics' share of the > >> system, and use that as half of the new active-active pair, beefing it > >> up with SSDs and perhaps additional memory; this would actually shrink > >> Physics near term costs, but puts higher pressure on the file system > for > >> the farm > >> > >> The decision is clearly Physics', but I do need a VERY FAST response to > >> this question, as I need to move quickly now for LQCD's needs. > >> > >> Hall D + GlueX, 96 TB > >> CLAS + CLAS12, 98 TB > >> Hall C, 35 TB > >> Hall A <unknown, still scanning> > >> > >> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion. > >> > >> thanks, > >> Chip > >> > >> > >> _______________________________________________ > >> Clas_offline mailing list > >> [log in to unmask] > >> https://mailman.jlab.org/mailman/listinfo/clas_offline > > > > _______________________________________________ > > Hps-analysis mailing list > > [log in to unmask] > > https://mailman.jlab.org/mailman/listinfo/hps-analysis > > ######################################################################## > Use REPLY-ALL to reply to list > > To unsubscribe from the HPS-SOFTWARE list, click the following link: > https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 > > ------------------------------ > > Use REPLY-ALL to reply to list > > To unsubscribe from the HPS-SOFTWARE list, click the following link: > https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 > ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the HPS-SOFTWARE list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1