[Hps-analysis] Fwd: [Clas_offline] Fwd: ENP consumption of disk space under /work
Nathan Baltzell
baltzell at jlab.org
Thu Jun 1 13:23:19 EDT 2017
Here’s the most relevant usage
649G mrsolt/
570G sebouh/
459G mc_production/
228G holly
159G mccaky/
78G rafopar/
45G omoreno/
44G spaul
39G fxgirod
34G jeremym
data/engrun2015:
3.2T tweakpass6
50G tweakpass6fail
64G tpass7
2.4G tpass7b
39G tpass7c
6.5G t_tweakpass_a
373G pass6/skim
201G pass6/dst
data/physrun2016:
3.5T pass0
690G feeiter4
94M feeiter0
327M feeiter1
339M feeiter2
338M feeiter3
15G noPass
24G pass0_allign
52G pass0fail
4.5G tmp_test
281G tpass1
11G upass0
On Jun 1, 2017, at 11:05, Stepan Stepanyan <stepanya at jlab.org> wrote:
> FYI, we need to move files.
>
> Stepan
>
>> Begin forwarded message:
>>
>> From: Harut Avakian <avakian at jlab.org>
>> Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work
>> Date: June 1, 2017 at 5:01:24 PM GMT+2
>> To: "clas_offline at jlab.org" <clas_offline at jlab.org>
>>
>>
>>
>>
>> Dear All,
>>
>> As you can see from the e-mail below, keeping all our work disk space requires some additional funding.
>> Option 3 will inevitably impact on farm operations, removing of ~20% space from Lustre.
>>
>> We can also choose something between options 1) and 3).
>> Please revise the content and move at least 75% of what is in /work/clas to either /cache or /volatile?
>> The current Hall-B usage includes:
>> 550G hallb/bonus
>> 1.5T hallb/clase1
>> 3.6T hallb/clase1-6
>> 3.3T hallb/clase1dvcs
>> 2.8T hallb/clase1dvcs2
>> 987G hallb/clase1f
>> 1.8T hallb/clase2
>> 1.6G hallb/clase5
>> 413G hallb/clase6
>> 2.2T hallb/claseg1
>> 3.9T hallb/claseg1dvcs
>> 1.2T hallb/claseg3
>> 4.1T hallb/claseg4
>> 2.7T hallb/claseg5
>> 1.7T hallb/claseg6
>> 367G hallb/clas-farm-output
>> 734G hallb/clasg10
>> 601G hallb/clasg11
>> 8.1T hallb/clasg12
>> 2.4T hallb/clasg13
>> 2.4T hallb/clasg14
>> 28G hallb/clasg3
>> 5.8G hallb/clasg7
>> 269G hallb/clasg8
>> 1.2T hallb/clasg9
>> 1.3T hallb/clashps
>> 1.8T hallb/clas-production
>> 5.6T hallb/clas-production2
>> 1.4T hallb/clas-production3
>> 12T hallb/hps
>> 13T hallb/prad
>>
>> Regards,
>>
>> Harut
>>
>> P.S. Few times we had crashes and they may also happen in future, so keeping important files in /work is not recommended.
>> You can see the list of lost files in /site/scicomp/lostfiles.txt and /site/scicomp/lostfiles-jan-2017.txt
>>
>>
>>
>> -------- Forwarded Message --------
>> Subject: ENP consumption of disk space under /work
>> Date: Wed, 31 May 2017 10:35:51 -0400
>> From: Chip Watson <watson at jlab.org>
>> To: Sandy Philpott <philpott at jlab.org>, Graham Heyes <heyes at jlab.org>, Ole Hansen <ole at jlab.org>, Harut Avakian <avakian at jlab.org>, Brad Sawatzky <brads at jlab.org>, Mark M. Ito <marki at jlab.org>
>>
>> All,
>>
>> As I have started on the procurement of the new /work file server, I
>> have discovered that Physics' use of /work has grown unrestrained over
>> the last year or two.
>>
>> "Unrestrained" because there is no way under Lustre to restrain it
>> except via a very unfriendly Lustre quota system. As we leave some
>> quota headroom to accommodate large swings in usage for each hall for
>> cache and volatile, then /work continues to grow.
>>
>> Total /work has now reached 260 TB, several times larger than I was
>> anticipating. This constitutes more than 25% of Physics' share of
>> Lustre, compared to LQCD which uses less than 5% of its disk space on
>> the un-managed /work.
>>
>> It would cost Physics an extra $25K (total $35K - $40K) to treat the 260
>> TB as a requirement.
>>
>> There are 3 paths forward:
>>
>> (1) Physics cuts its use of /work by a factor of 4-5.
>> (2) Physics increases funding to $40K
>> (3) We pull a server out of Lustre, decreasing Physics' share of the
>> system, and use that as half of the new active-active pair, beefing it
>> up with SSDs and perhaps additional memory; this would actually shrink
>> Physics near term costs, but puts higher pressure on the file system for
>> the farm
>>
>> The decision is clearly Physics', but I do need a VERY FAST response to
>> this question, as I need to move quickly now for LQCD's needs.
>>
>> Hall D + GlueX, 96 TB
>> CLAS + CLAS12, 98 TB
>> Hall C, 35 TB
>> Hall A <unknown, still scanning>
>>
>> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion.
>>
>> thanks,
>> Chip
>>
>>
>> _______________________________________________
>> Clas_offline mailing list
>> Clas_offline at jlab.org
>> https://mailman.jlab.org/mailman/listinfo/clas_offline
>
> _______________________________________________
> Hps-analysis mailing list
> Hps-analysis at jlab.org
> https://mailman.jlab.org/mailman/listinfo/hps-analysis
More information about the Hps-analysis
mailing list