[G12] Fwd: [Clas_offline] Fwd: ENP consumption of disk space under /work
Michael Kunkel
mkunkel at jlab.org
Fri Jun 2 02:27:53 EDT 2017
Greetings G12,
I am not sure how many received this email from Harut, so please read below.
I think we, g12, can help out drastically on this because we all know the amount of active analyses in g12 should not be on the order of 8.1 TB of data.
I have previewed some directories from g12 members long departed.
Please review your workspace for ONLY g12 stuff. CLAS12 data/analyses belong on the CLAS12 directory, same for GLUEX.
Let's discuss this topic at our Tuesday meeting. I rather not be the experiment that had a hand in having the disk availability cut for halb.
BR
MK
----------------------------------------
Michael C. Kunkel, USMC, PhD
Forschungszentrum Jülich
Nuclear Physics Institute and Juelich Center for Hadron Physics
Experimental Hadron Structure (IKP-1)
www.fz-juelich.de/ikp
Begin forwarded message:
> From: Harut Avakian <avakian at jlab.org>
> Date: June 1, 2017 at 17:01:24 GMT+2
> To: "clas_offline at jlab.org" <clas_offline at jlab.org>
> Subject: [Clas_offline] Fwd: ENP consumption of disk space under /work
>
>
>
>
> Dear All,
>
> As you can see from the e-mail below, keeping all our work disk space requires some additional funding.
> Option 3 will inevitably impact on farm operations, removing of ~20% space from Lustre.
>
> We can also choose something between options 1) and 3).
> Please revise the content and move at least 75% of what is in /work/clas to either /cache or /volatile?
> The current Hall-B usage includes:
> 550G hallb/bonus
> 1.5T hallb/clase1
> 3.6T hallb/clase1-6
> 3.3T hallb/clase1dvcs
> 2.8T hallb/clase1dvcs2
> 987G hallb/clase1f
> 1.8T hallb/clase2
> 1.6G hallb/clase5
> 413G hallb/clase6
> 2.2T hallb/claseg1
> 3.9T hallb/claseg1dvcs
> 1.2T hallb/claseg3
> 4.1T hallb/claseg4
> 2.7T hallb/claseg5
> 1.7T hallb/claseg6
> 367G hallb/clas-farm-output
> 734G hallb/clasg10
> 601G hallb/clasg11
> 8.1T hallb/clasg12
> 2.4T hallb/clasg13
> 2.4T hallb/clasg14
> 28G hallb/clasg3
> 5.8G hallb/clasg7
> 269G hallb/clasg8
> 1.2T hallb/clasg9
> 1.3T hallb/clashps
> 1.8T hallb/clas-production
> 5.6T hallb/clas-production2
> 1.4T hallb/clas-production3
> 12T hallb/hps
> 13T hallb/prad
>
> Regards,
>
> Harut
>
> P.S. Few times we had crashes and they may also happen in future, so keeping important files in /work is not recommended.
> You can see the list of lost files in /site/scicomp/lostfiles.txt and /site/scicomp/lostfiles-jan-2017.txt
>
>
>
> -------- Forwarded Message --------
> Subject: ENP consumption of disk space under /work
> Date: Wed, 31 May 2017 10:35:51 -0400
> From: Chip Watson <watson at jlab.org>
> To: Sandy Philpott <philpott at jlab.org>, Graham Heyes <heyes at jlab.org>, Ole Hansen <ole at jlab.org>, Harut Avakian <avakian at jlab.org>, Brad Sawatzky <brads at jlab.org>, Mark M. Ito <marki at jlab.org>
>
> All,
>
> As I have started on the procurement of the new /work file server, I
> have discovered that Physics' use of /work has grown unrestrained over
> the last year or two.
>
> "Unrestrained" because there is no way under Lustre to restrain it
> except via a very unfriendly Lustre quota system. As we leave some
> quota headroom to accommodate large swings in usage for each hall for
> cache and volatile, then /work continues to grow.
>
> Total /work has now reached 260 TB, several times larger than I was
> anticipating. This constitutes more than 25% of Physics' share of
> Lustre, compared to LQCD which uses less than 5% of its disk space on
> the un-managed /work.
>
> It would cost Physics an extra $25K (total $35K - $40K) to treat the 260
> TB as a requirement.
>
> There are 3 paths forward:
>
> (1) Physics cuts its use of /work by a factor of 4-5.
> (2) Physics increases funding to $40K
> (3) We pull a server out of Lustre, decreasing Physics' share of the
> system, and use that as half of the new active-active pair, beefing it
> up with SSDs and perhaps additional memory; this would actually shrink
> Physics near term costs, but puts higher pressure on the file system for
> the farm
>
> The decision is clearly Physics', but I do need a VERY FAST response to
> this question, as I need to move quickly now for LQCD's needs.
>
> Hall D + GlueX, 96 TB
> CLAS + CLAS12, 98 TB
> Hall C, 35 TB
> Hall A <unknown, still scanning>
>
> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion.
>
> thanks,
> Chip
>
> _______________________________________________
> Clas_offline mailing list
> Clas_offline at jlab.org
> https://mailman.jlab.org/mailman/listinfo/clas_offline
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.jlab.org/pipermail/g12/attachments/20170602/f487a721/attachment.html>
More information about the G12
mailing list