Print

Print


Here’s the most relevant usage

649G	mrsolt/
570G	sebouh/
459G	mc_production/
228G	holly
159G	mccaky/
78G	rafopar/
45G	omoreno/
44G	spaul
39G	fxgirod
34G	jeremym

data/engrun2015:
3.2T	tweakpass6
50G	tweakpass6fail
64G	tpass7
2.4G	tpass7b
39G	tpass7c
6.5G	t_tweakpass_a
373G    pass6/skim
201G    pass6/dst

data/physrun2016:
3.5T	pass0
690G	feeiter4
94M	feeiter0
327M	feeiter1
339M	feeiter2
338M	feeiter3
15G	noPass
24G	pass0_allign
52G	pass0fail
4.5G	tmp_test
281G	tpass1
11G	upass0




On Jun 1, 2017, at 11:05, Stepan Stepanyan <[log in to unmask]> wrote:

> FYI, we need to move files.
> 
> Stepan
> 
>> Begin forwarded message:
>> 
>> From: Harut Avakian <[log in to unmask]>
>> Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work
>> Date: June 1, 2017 at 5:01:24 PM GMT+2
>> To: "[log in to unmask]" <[log in to unmask]>
>> 
>> 
>> 
>> 
>> Dear All,
>> 
>> As you can see from the e-mail below,  keeping all our work disk space requires some additional funding.
>> Option 3 will inevitably impact on farm operations, removing of ~20% space from Lustre.
>> 
>> We can also choose something between options 1) and 3).
>> Please revise the content and move at least 75% of what is in /work/clas  to either /cache or /volatile?  
>> The current Hall-B usage includes:
>> 550G    hallb/bonus
>> 1.5T    hallb/clase1
>> 3.6T    hallb/clase1-6
>> 3.3T    hallb/clase1dvcs
>> 2.8T    hallb/clase1dvcs2
>> 987G    hallb/clase1f
>> 1.8T    hallb/clase2
>> 1.6G    hallb/clase5
>> 413G    hallb/clase6
>> 2.2T    hallb/claseg1
>> 3.9T    hallb/claseg1dvcs
>> 1.2T    hallb/claseg3
>> 4.1T    hallb/claseg4
>> 2.7T    hallb/claseg5
>> 1.7T    hallb/claseg6
>> 367G    hallb/clas-farm-output
>> 734G    hallb/clasg10
>> 601G    hallb/clasg11
>> 8.1T    hallb/clasg12
>> 2.4T    hallb/clasg13
>> 2.4T    hallb/clasg14
>> 28G    hallb/clasg3
>> 5.8G    hallb/clasg7
>> 269G    hallb/clasg8
>> 1.2T    hallb/clasg9
>> 1.3T    hallb/clashps
>> 1.8T    hallb/clas-production
>> 5.6T    hallb/clas-production2
>> 1.4T    hallb/clas-production3
>> 12T    hallb/hps
>> 13T    hallb/prad
>> 
>> Regards,
>> 
>> Harut
>> 
>> P.S. Few times we had crashes and they may also happen in future, so keeping important files in /work is not recommended.
>> You can see the list of lost files in /site/scicomp/lostfiles.txt  and  /site/scicomp/lostfiles-jan-2017.txt
>> 
>> 
>> 
>> -------- Forwarded Message --------
>> Subject:	ENP consumption of disk space under /work
>> Date:	Wed, 31 May 2017 10:35:51 -0400
>> From:	Chip Watson <[log in to unmask]>
>> To:	Sandy Philpott <[log in to unmask]>, Graham Heyes <[log in to unmask]>, Ole Hansen <[log in to unmask]>, Harut Avakian <[log in to unmask]>, Brad Sawatzky <[log in to unmask]>, Mark M. Ito <[log in to unmask]>
>> 
>> All,
>> 
>> As I have started on the procurement of the new /work file server, I 
>> have discovered that Physics' use of /work has grown unrestrained over 
>> the last year or two.
>> 
>> "Unrestrained" because there is no way under Lustre to restrain it 
>> except via a very unfriendly Lustre quota system.  As we leave some 
>> quota headroom to accommodate large swings in usage for each hall for 
>> cache and volatile, then /work continues to grow.
>> 
>> Total /work has now reached 260 TB, several times larger than I was 
>> anticipating.  This constitutes more than 25% of Physics' share of 
>> Lustre, compared to LQCD which uses less than 5% of its disk space on 
>> the un-managed /work.
>> 
>> It would cost Physics an extra $25K (total $35K - $40K) to treat the 260 
>> TB as a requirement.
>> 
>> There are 3 paths forward:
>> 
>> (1) Physics cuts its use of /work by a factor of 4-5.
>> (2) Physics increases funding to $40K
>> (3) We pull a server out of Lustre, decreasing Physics' share of the 
>> system, and use that as half of the new active-active pair, beefing it 
>> up with SSDs and perhaps additional memory; this would actually shrink 
>> Physics near term costs, but puts higher pressure on the file system for 
>> the farm
>> 
>> The decision is clearly Physics', but I do need a VERY FAST response to 
>> this question, as I need to move quickly now for LQCD's needs.
>> 
>> Hall D + GlueX,  96 TB
>> CLAS + CLAS12, 98 TB
>> Hall C,                35 TB
>> Hall A <unknown, still scanning>
>> 
>> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion.
>> 
>> thanks,
>> Chip
>> 
>> 
>> _______________________________________________
>> Clas_offline mailing list
>> [log in to unmask]
>> https://mailman.jlab.org/mailman/listinfo/clas_offline
> 
> _______________________________________________
> Hps-analysis mailing list
> [log in to unmask]
> https://mailman.jlab.org/mailman/listinfo/hps-analysis

########################################################################
Use REPLY-ALL to reply to list

To unsubscribe from the HPS-SOFTWARE list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1