[Halld-offline] ENP consumption of disk space under /work
Mark Ito
marki at jlab.org
Thu Jun 8 09:40:19 EDT 2017
That was just an observation. I would not interpret it as a vote for
anything... :-)
On 06/08/2017 09:33 AM, Chip Watson wrote:
>
> I'll take this as a vote by GlueX to have more work and reduce cache.
>
> Do A,B,C concur?
>
>
> On 6/8/17 9:28 AM, Mark Ito wrote:
>>
>> In my previous estimate, of the cache portion, 278 TB, only 105 TB of
>> that is pinned. The unpinned part is presumeably old files that
>> should be gone, but have not been deleted since there happens to be
>> no demand for the space. If we use 105 TB as our cache usage then
>> re-doing your estimate gives 555 TB, which means in 9 months we will
>> have 270 TB of unused space. Which would mean that we have room to
>> increase our usage without buying anything!
>>
>>
>> On 06/07/2017 05:54 PM, Chip Watson wrote:
>>>
>>> Mark,
>>>
>>> I still need you to answer the question of how to further reduce
>>> usage and how to configure. Your usage as you report it is about
>>> 370 TB. Assuming that Hall B needs the same within 9 months, and
>>> that A+C need half as much, then that leads to a total of 925TB
>>> which is more than Physics owns, by 100 TB (NOT CURRENT USAGE, JUST
>>> PROJECTION BASED ON GLUEX USAGE).
>>>
>>> There is also the question of how to split the storage budget. In
>>> budget, you can have new half a JBOD: 21 disks configured as 3 RAID
>>> z2 stripes of 5+2 disks, 8TB, thus 120 raw data, 108 in a file
>>> system, and 86 TB at 80% -- for all of GlueX, CLAS-12, A and C. If
>>> GlueX is 40% of the total, that makes 35TB, and you are still high
>>> by 70%.
>>>
>>> The other low cost option is to re-purpose a 2016 Lustre node so
>>> that /work is twice this size (one full JBOD), and GlueX can use
>>> 70TB as /work. But then you must reduce /cache + /volatile by a
>>> comparable amount since we have to pull a node out of production.
>>> And this still isn't free since we'll need a total of 4 RAID cards
>>> instead of 2 to provide correct performance, and we'll need to add
>>> SSD's to the mix.
>>>
>>> So, in the absence of money (which clearly seems to be the case), do
>>> you choose (a) reduce your use of work by 1.7x, or (b) reduce your
>>> use of cache + volatile by 25%. There is no middle case.
>>>
>>> thanks,
>>>
>>> Chip
>>>
>>>
>>> On 6/7/17 5:30 PM, Mark Ito wrote:
>>>>
>>>> Summarizing Hall D work disk usage (/work/halld only):
>>>>
>>>> o using du, today 2017-06-06, 59 TB
>>>>
>>>> o from our disk-management database, a couple of days ago,
>>>> 2017-06-04, 86 TB
>>>>
>>>> I also know that one of our students got rid of about 20 TB of
>>>> unneeded files yesterday. That accounts for part of the drop.
>>>>
>>>> We produce a report from that database
>>>> <https://halldweb.jlab.org/disk_management/work_report.html> that
>>>> is updated every few days.
>>>>
>>>> From the SciComp pages, Hall D is using 287 TB on cache and 21 TB
>>>> on volatile.
>>>>
>>>> My view is that this level of work disk usage is more or less as
>>>> expected, consistent with our previous estimates, and not
>>>> particularly abusive. That having been said, I am sure there is a
>>>> lot that can be cleaned up. But as Ole pointed out, disk usage
>>>> grows naturally and we were not aware that this was a problem. I
>>>> seem to recall that we agreed to respond to emails that would be
>>>> sent when we reached 90% of too much, no? Was the email sent out?
>>>>
>>>> One mystery: when I ask Lustre what we are using I get:
>>>>
>>>> ifarm1402:marki:marki> lfs quota -gh halld /lustre
>>>> Disk quotas for group halld (gid 267):
>>>> Filesystem used quota limit grace files quota limit grace
>>>> /lustre 290T 470T 500T - 15106047 0 0 -
>>>> which is less than cache + volatile, not to mention work. I thought
>>>> that to a good approximation this 290 TB should be the sum of all
>>>> three. What am I missing?
>>>>
>>>> On 05/31/2017 10:35 AM, Chip Watson wrote:
>>>>> All,
>>>>>
>>>>> As I have started on the procurement of the new /work file server,
>>>>> I have discovered that Physics' use of /work has grown
>>>>> unrestrained over the last year or two.
>>>>>
>>>>> "Unrestrained" because there is no way under Lustre to restrain it
>>>>> except via a very unfriendly Lustre quota system. As we leave
>>>>> some quota headroom to accommodate large swings in usage for each
>>>>> hall for cache and volatile, then /work continues to grow.
>>>>>
>>>>> Total /work has now reached 260 TB, several times larger than I
>>>>> was anticipating. This constitutes more than 25% of Physics'
>>>>> share of Lustre, compared to LQCD which uses less than 5% of its
>>>>> disk space on the un-managed /work.
>>>>>
>>>>> It would cost Physics an extra $25K (total $35K - $40K) to treat
>>>>> the 260 TB as a requirement.
>>>>>
>>>>> There are 3 paths forward:
>>>>>
>>>>> (1) Physics cuts its use of /work by a factor of 4-5.
>>>>> (2) Physics increases funding to $40K
>>>>> (3) We pull a server out of Lustre, decreasing Physics' share of
>>>>> the system, and use that as half of the new active-active pair,
>>>>> beefing it up with SSDs and perhaps additional memory; this would
>>>>> actually shrink Physics near term costs, but puts higher pressure
>>>>> on the file system for the farm
>>>>>
>>>>> The decision is clearly Physics', but I do need a VERY FAST
>>>>> response to this question, as I need to move quickly now for
>>>>> LQCD's needs.
>>>>>
>>>>> Hall D + GlueX, 96 TB
>>>>> CLAS + CLAS12, 98 TB
>>>>> Hall C, 35 TB
>>>>> Hall A <unknown, still scanning>
>>>>>
>>>>> Email, call (x7101), or drop by today 1:30-3:00 p.m. for discussion.
>>>>>
>>>>> thanks,
>>>>> Chip
>>>>>
>>>>
>>>> --
>>>> Mark Ito,marki at jlab.org, (757)269-5295
>>>
>>
>> --
>> Mark Ito,marki at jlab.org, (757)269-5295
>
--
Mark Ito, marki at jlab.org, (757)269-5295
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.jlab.org/pipermail/halld-offline/attachments/20170608/054c5c09/attachment-0002.html>
More information about the Halld-offline
mailing list