Print

Print


I have reduced my /work/ usage to 16 MB.

On Thu, Jun 1, 2017 at 3:05 PM, Bradley T Yale <[log in to unmask]>
wrote:

> I deleted/moved 351 GB in mc_production.
>
>
> The following things are owned by other users though, and ~6 GB more can
> be freed up if no longer needed:
>
>
> Luca:
>
> 762M: /work/hallb/hps/mc_production/Luca/lhe/Vegas_10_10_2016/
>
>
> Holly:
> 3.1G: /work/hallb/hps/mc_production/MG5/dst/
>
> 2G: /work/hallb/hps/mc_production/postTriSummitFixes/tritrig/
> 1pt05/NOSUMCUT/
>
> 25M:  /work/hallb/hps/mc_production/logs/slic/ap/
>
> 16M: /work/hallb/hps/mc_production/logs/readout/beam-tri/1pt92/
>
> Matt G:
>
> 35M: /work/hallb/hps/mc_production/dst/
>
> 11M: /work/hallb/hps/mc_production/logs/dqm/
>
>
> There is also 2 GB of old 2.2 GeV A' MC, which should no longer be
> relevant, but I didn't want to do anything with it since it had mock data
> stuff in there:
>
> /work/hallb/hps/mc_production/lhe/ap/2pt2/
>
>
> ------------------------------
> *From:* [log in to unmask] <[log in to unmask]>
> on behalf of Nathan Baltzell <[log in to unmask]>
> *Sent:* Thursday, June 1, 2017 1:23:19 PM
> *To:* HPS-SOFTWARE
> *Cc:* [log in to unmask]
> *Subject:* Re: [Hps-analysis] Fwd: [Clas_offline] Fwd: ENP consumption of
> disk space under /work
>
> Here’s the most relevant usage
>
> 649G    mrsolt/
> 570G    sebouh/
> 459G    mc_production/
> 228G    holly
> 159G    mccaky/
> 78G     rafopar/
> 45G     omoreno/
> 44G     spaul
> 39G     fxgirod
> 34G     jeremym
>
> data/engrun2015:
> 3.2T    tweakpass6
> 50G     tweakpass6fail
> 64G     tpass7
> 2.4G    tpass7b
> 39G     tpass7c
> 6.5G    t_tweakpass_a
> 373G    pass6/skim
> 201G    pass6/dst
>
> data/physrun2016:
> 3.5T    pass0
> 690G    feeiter4
> 94M     feeiter0
> 327M    feeiter1
> 339M    feeiter2
> 338M    feeiter3
> 15G     noPass
> 24G     pass0_allign
> 52G     pass0fail
> 4.5G    tmp_test
> 281G    tpass1
> 11G     upass0
>
>
>
>
> On Jun 1, 2017, at 11:05, Stepan Stepanyan <[log in to unmask]> wrote:
>
> > FYI, we need to move files.
> >
> > Stepan
> >
> >> Begin forwarded message:
> >>
> >> From: Harut Avakian <[log in to unmask]>
> >> Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work
> >> Date: June 1, 2017 at 5:01:24 PM GMT+2
> >> To: "[log in to unmask]" <[log in to unmask]>
> >>
> >>
> >>
> >>
> >> Dear All,
> >>
> >> As you can see from the e-mail below,  keeping all our work disk space
> requires some additional funding.
> >> Option 3 will inevitably impact on farm operations, removing of ~20%
> space from Lustre.
> >>
> >> We can also choose something between options 1) and 3).
> >> Please revise the content and move at least 75% of what is in
> /work/clas  to either /cache or /volatile?
> >> The current Hall-B usage includes:
> >> 550G    hallb/bonus
> >> 1.5T    hallb/clase1
> >> 3.6T    hallb/clase1-6
> >> 3.3T    hallb/clase1dvcs
> >> 2.8T    hallb/clase1dvcs2
> >> 987G    hallb/clase1f
> >> 1.8T    hallb/clase2
> >> 1.6G    hallb/clase5
> >> 413G    hallb/clase6
> >> 2.2T    hallb/claseg1
> >> 3.9T    hallb/claseg1dvcs
> >> 1.2T    hallb/claseg3
> >> 4.1T    hallb/claseg4
> >> 2.7T    hallb/claseg5
> >> 1.7T    hallb/claseg6
> >> 367G    hallb/clas-farm-output
> >> 734G    hallb/clasg10
> >> 601G    hallb/clasg11
> >> 8.1T    hallb/clasg12
> >> 2.4T    hallb/clasg13
> >> 2.4T    hallb/clasg14
> >> 28G    hallb/clasg3
> >> 5.8G    hallb/clasg7
> >> 269G    hallb/clasg8
> >> 1.2T    hallb/clasg9
> >> 1.3T    hallb/clashps
> >> 1.8T    hallb/clas-production
> >> 5.6T    hallb/clas-production2
> >> 1.4T    hallb/clas-production3
> >> 12T    hallb/hps
> >> 13T    hallb/prad
> >>
> >> Regards,
> >>
> >> Harut
> >>
> >> P.S. Few times we had crashes and they may also happen in future, so
> keeping important files in /work is not recommended.
> >> You can see the list of lost files in /site/scicomp/lostfiles.txt  and
> /site/scicomp/lostfiles-jan-2017.txt
> >>
> >>
> >>
> >> -------- Forwarded Message --------
> >> Subject:     ENP consumption of disk space under /work
> >> Date:        Wed, 31 May 2017 10:35:51 -0400
> >> From:        Chip Watson <[log in to unmask]>
> >> To:  Sandy Philpott <[log in to unmask]>, Graham Heyes <[log in to unmask]>,
> Ole Hansen <[log in to unmask]>, Harut Avakian <[log in to unmask]>, Brad
> Sawatzky <[log in to unmask]>, Mark M. Ito <[log in to unmask]>
> >>
> >> All,
> >>
> >> As I have started on the procurement of the new /work file server, I
> >> have discovered that Physics' use of /work has grown unrestrained over
> >> the last year or two.
> >>
> >> "Unrestrained" because there is no way under Lustre to restrain it
> >> except via a very unfriendly Lustre quota system.  As we leave some
> >> quota headroom to accommodate large swings in usage for each hall for
> >> cache and volatile, then /work continues to grow.
> >>
> >> Total /work has now reached 260 TB, several times larger than I was
> >> anticipating.  This constitutes more than 25% of Physics' share of
> >> Lustre, compared to LQCD which uses less than 5% of its disk space on
> >> the un-managed /work.
> >>
> >> It would cost Physics an extra $25K (total $35K - $40K) to treat the
> 260
> >> TB as a requirement.
> >>
> >> There are 3 paths forward:
> >>
> >> (1) Physics cuts its use of /work by a factor of 4-5.
> >> (2) Physics increases funding to $40K
> >> (3) We pull a server out of Lustre, decreasing Physics' share of the
> >> system, and use that as half of the new active-active pair, beefing it
> >> up with SSDs and perhaps additional memory; this would actually shrink
> >> Physics near term costs, but puts higher pressure on the file system
> for
> >> the farm
> >>
> >> The decision is clearly Physics', but I do need a VERY FAST response to
> >> this question, as I need to move quickly now for LQCD's needs.
> >>
> >> Hall D + GlueX,  96 TB
> >> CLAS + CLAS12, 98 TB
> >> Hall C,                35 TB
> >> Hall A <unknown, still scanning>
> >>
> >> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion.
> >>
> >> thanks,
> >> Chip
> >>
> >>
> >> _______________________________________________
> >> Clas_offline mailing list
> >> [log in to unmask]
> >> https://mailman.jlab.org/mailman/listinfo/clas_offline
> >
> > _______________________________________________
> > Hps-analysis mailing list
> > [log in to unmask]
> > https://mailman.jlab.org/mailman/listinfo/hps-analysis
>
> ########################################################################
> Use REPLY-ALL to reply to list
>
> To unsubscribe from the HPS-SOFTWARE list, click the following link:
> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1
>
> ------------------------------
>
> Use REPLY-ALL to reply to list
>
> To unsubscribe from the HPS-SOFTWARE list, click the following link:
> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1
>

########################################################################
Use REPLY-ALL to reply to list

To unsubscribe from the HPS-SOFTWARE list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1