[Halld-offline] Fwd: FYI disk fileservers
Mark Ito
marki at jlab.org
Fri Nov 3 14:55:22 EDT 2017
from Graham Heyes, explaining the recent history of disk use and disk
procurements...
-------- Forwarded Message --------
Subject: FYI disk fileservers
Date: Thu, 2 Nov 2017 12:53:04 -0400
From: Graham Heyes <heyes at jlab.org>
To: Mark Ito <marki at jlab.org>, Ole Hansen <ole at jlab.org>, Brad Sawatzky
<brads at jlab.org>, Harut Avakian <avakian at jlab.org>
Someone outside yesterday's meeting had a question about disk space so I
wrote down the story so far. I thought it was a useful enough summary
that I would send it to you in case you would like to share it with your
people too.
Last year we identified the “work” filesystem as an area in need of
improvement. By it’s very nature a “work” filesystem can contain a large
number of small files that churn as people run jobs, compile code, etc.
Unfortunately the Luster based system is not very efficient in this
mode. What users see as the “work” filesystem is a virtual space carved
out of a larger filesystem that also provides /cache and /volitile for
the farm and space paid for by LQCD. The non-work areas are machine
managed using algorithms that automatically free space by migrating old
or little used files to tape. Conversely /work is “human managed” using
quotas. A problem with this is that growth in the work area is at the
expense of /cache and /volatile which gradually shrink. Also accessing
many small files simultaneously in /work impacts performance for /cache,
/volatile and the LQCD users.
The solution to the problem was to buy a new file server specifically
designed for use as “work”. Due to funding constraints in FY17 we bought
a system that was not fully populated with drives and controllers with
the plan to expand at a later date. This minimal system was designed to
meet the needs of FY18 as understood at the time. Halls A/C offered to
add funds of their own to aid the procurement in return for a “larger
slice of the pie”. Normally I buy disk out of my budget but I was glad
of the one time help. I spoke with Rolf and we agreed that, for many
reasons, we do not want the halls buying their own disk to add to the
farm. The model is that the halls provide requirements and Rolf pays,
via me, for IT to procure something to meet the requirements. One reason
for this is that typically we add disk space in large chunks to keep the
cost per terabyte down so we don’t want to buy piecemeal.
Here is a summary of where we are:
Currently on Luster: work=170TB, cache=400 TB, Volatile=165 TB, for
a total of 735 TB
Note that ENP only bought 690 TB and the 45 TB difference is “on
loan" from LQCD.
The FY17 procurement is a high performance server optimized to be
/work with 144 TB useable.
We asked all the halls to temporarily reduce their use of /work to
fit in the new server. After some pushback we are allowing some
rarely used data to stay on Luster and are commencing the move to
the new server.
The FY18 procurement which will be installed in January adds 216 TB.
So, after January we will have 360 TB of high performance /work on
the new server compared with 170 TB of /work now under Luster. So
double the space and higher performance hardware.
The space freed up on Luster will be added to /cache and /volatile. As
data processing ramps up, whether on or off site, we will need more
/cache to stage the data being processed. Based on requirement
projections I expect this to exceed what we own so I have asked Chip to
prepare another REQ to add 250 TB to the /cache and /volatile which will
expand it to almost 1PB.
I am looking closely at the disk requirements that we had in the last
computing review and am asking everyone to update them in light of
additional experience in recent months. In particular to be sure that we
do need to add the 250 TB of cache and if we do that will it be enough
for the long term or will we need more.
I hope that this is useful!
Regards,
Graham
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.jlab.org/pipermail/halld-offline/attachments/20171103/01483d8a/attachment-0001.html>
More information about the Halld-offline
mailing list