FYI, we need to move files. Stepan > Begin forwarded message: > > From: Harut Avakian <[log in to unmask]> > Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work > Date: June 1, 2017 at 5:01:24 PM GMT+2 > To: "[log in to unmask]" <[log in to unmask]> > > > > > Dear All, > > As you can see from the e-mail below, keeping all our work disk space requires some additional funding. > Option 3 will inevitably impact on farm operations, removing of ~20% space from Lustre. > > We can also choose something between options 1) and 3). > Please revise the content and move at least 75% of what is in /work/clas to either /cache or /volatile? > The current Hall-B usage includes: > 550G hallb/bonus > 1.5T hallb/clase1 > 3.6T hallb/clase1-6 > 3.3T hallb/clase1dvcs > 2.8T hallb/clase1dvcs2 > 987G hallb/clase1f > 1.8T hallb/clase2 > 1.6G hallb/clase5 > 413G hallb/clase6 > 2.2T hallb/claseg1 > 3.9T hallb/claseg1dvcs > 1.2T hallb/claseg3 > 4.1T hallb/claseg4 > 2.7T hallb/claseg5 > 1.7T hallb/claseg6 > 367G hallb/clas-farm-output > 734G hallb/clasg10 > 601G hallb/clasg11 > 8.1T hallb/clasg12 > 2.4T hallb/clasg13 > 2.4T hallb/clasg14 > 28G hallb/clasg3 > 5.8G hallb/clasg7 > 269G hallb/clasg8 > 1.2T hallb/clasg9 > 1.3T hallb/clashps > 1.8T hallb/clas-production > 5.6T hallb/clas-production2 > 1.4T hallb/clas-production3 > 12T hallb/hps > 13T hallb/prad > > Regards, > > Harut > > P.S. Few times we had crashes and they may also happen in future, so keeping important files in /work is not recommended. > You can see the list of lost files in /site/scicomp/lostfiles.txt and /site/scicomp/lostfiles-jan-2017.txt > > > > -------- Forwarded Message -------- > Subject: ENP consumption of disk space under /work > Date: Wed, 31 May 2017 10:35:51 -0400 > From: Chip Watson <[log in to unmask]> <mailto:[log in to unmask]> > To: Sandy Philpott <[log in to unmask]> <mailto:[log in to unmask]>, Graham Heyes <[log in to unmask]> <mailto:[log in to unmask]>, Ole Hansen <[log in to unmask]> <mailto:[log in to unmask]>, Harut Avakian <[log in to unmask]> <mailto:[log in to unmask]>, Brad Sawatzky <[log in to unmask]> <mailto:[log in to unmask]>, Mark M. Ito <[log in to unmask]> <mailto:[log in to unmask]> > > All, > > As I have started on the procurement of the new /work file server, I > have discovered that Physics' use of /work has grown unrestrained over > the last year or two. > > "Unrestrained" because there is no way under Lustre to restrain it > except via a very unfriendly Lustre quota system. As we leave some > quota headroom to accommodate large swings in usage for each hall for > cache and volatile, then /work continues to grow. > > Total /work has now reached 260 TB, several times larger than I was > anticipating. This constitutes more than 25% of Physics' share of > Lustre, compared to LQCD which uses less than 5% of its disk space on > the un-managed /work. > > It would cost Physics an extra $25K (total $35K - $40K) to treat the 260 > TB as a requirement. > > There are 3 paths forward: > > (1) Physics cuts its use of /work by a factor of 4-5. > (2) Physics increases funding to $40K > (3) We pull a server out of Lustre, decreasing Physics' share of the > system, and use that as half of the new active-active pair, beefing it > up with SSDs and perhaps additional memory; this would actually shrink > Physics near term costs, but puts higher pressure on the file system for > the farm > > The decision is clearly Physics', but I do need a VERY FAST response to > this question, as I need to move quickly now for LQCD's needs. > > Hall D + GlueX, 96 TB > CLAS + CLAS12, 98 TB > Hall C, 35 TB > Hall A <unknown, still scanning> > > Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion. > > thanks, > Chip > > _______________________________________________ > Clas_offline mailing list > [log in to unmask] > https://mailman.jlab.org/mailman/listinfo/clas_offline ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the HPS-SOFTWARE list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1