[Halld-offline] [EXTERNAL] first step toward a global filesystem for Gluex

Mark Ito marki at jlab.org
Tue Nov 16 21:03:46 EST 2021


Richard,

Is this new?

My immediate thought is the possibility of having our REST files 
available via this mechanism. For each run period, we could have the 
latest generation of REST files out there. The space at JLab has not 
turned out to be sufficient to support such a scheme. If we want to 
analyze REST files at JLab, the savings in avoiding tape latency might 
overcome the overhead in getting our REST files over the network.

Even if the scheme described above has some fatal flaw, it sounds to me 
that this resource is available now, at least at some level. If that is 
the case it would be a shame to not come up with a use for it. Random 
trigger files maybe?

   -- Mark

On 11/16/21 10:13 AM, Richard Jones wrote:
> Hello all,
>
> A few weeks ago, the possibility was raised of a shared global 
> filesystem to provide easy access to shared Gluex data (eg. REST, 
> analysis data sets, skims of various kinds) from anywhere on or off 
> site without having to wait to stage data from tape. As a first step, 
> I have created a namespace for these files under the osgstorage file 
> catalog, managed by the osg ops.
>
>   * /cvmfs/gluex.osgstorage.org/gluex
>     <https://urldefense.proofpoint.com/v2/url?u=http-3A__gluex.osgstorage.org_gluex&d=DwMFaQ&c=CJqEzB1piLOyyvZjb8YUQw&r=Te_hCR4EUlJ6iCDYLJ8Viv2aDOR7D9ZZMoBAvf2H0M4&m=2oEolNTMkADUFZV21KX3t875hunlJB8Xll9JSnHrs6U8hK2lILwL9Y7klqBibkxJ&s=OVDdt1Bt4yRmZhJL7NtZE68NN5hAsTLoSsSgp6a3d5k&e=>
>
> The purpose of the second /gluex is to allow various physics working 
> groups (eg. cpp, primex) to have their own separate branch under 
> /cvmfs/gluex.osgstorage.org 
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__gluex.osgstorage.org&d=DwMFaQ&c=CJqEzB1piLOyyvZjb8YUQw&r=Te_hCR4EUlJ6iCDYLJ8Viv2aDOR7D9ZZMoBAvf2H0M4&m=2oEolNTMkADUFZV21KX3t875hunlJB8Xll9JSnHrs6U8hK2lILwL9Y7klqBibkxJ&s=qdZEQVKOakAlhVROu5FIQAdPuIh1qRFquFmpc_yMazg&e=>. 
> This osgstorage.org 
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__osgstorage.org&d=DwMFaQ&c=CJqEzB1piLOyyvZjb8YUQw&r=Te_hCR4EUlJ6iCDYLJ8Viv2aDOR7D9ZZMoBAvf2H0M4&m=2oEolNTMkADUFZV21KX3t875hunlJB8Xll9JSnHrs6U8hK2lILwL9Y7klqBibkxJ&s=HLeKQCfxRbyNBlNqlPDf-EA-_PUpSs_OFEL8rCe1qXI&e=> 
> is built around a network of shared caches across North America that 
> automatically finds and serves you the nearest copy of any file that 
> is registered in the catalog. The data are also locally cached on your 
> local machine through the cvmfs caching mechanism, for repeated access 
> to the same files.
>
> Right now UConn is contributing the "origin" service for the gluex 
> namespace, but hopefully JLab will also contribute to this in the near 
> future. To provide an osgstorage origin service, all you need is to 
> export your files using a standard xrootd server. Just email 
> support.opensciencegrid.org 
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__support.opensciencegrid.org&d=DwMFaQ&c=CJqEzB1piLOyyvZjb8YUQw&r=Te_hCR4EUlJ6iCDYLJ8Viv2aDOR7D9ZZMoBAvf2H0M4&m=2oEolNTMkADUFZV21KX3t875hunlJB8Xll9JSnHrs6U8hK2lILwL9Y7klqBibkxJ&s=w8Z8-6cXZfKrrnTBAdsrQB_VKWzJdWCuV2afHPIlijk&e=> 
> and tell them what portion of the /gluex namespace you want to occupy, 
> and they will start automatically indexing your files and adding them 
> to the catalog.
>
> If you don't have any storage to contribute, but you would like to 
> take advantage of this shared Gluex storage, write to Mark Ito or to 
> me and tell us what datasets you would like to see published through 
> the system. If you don't have /cvmfs installed, you can still access 
> any file in the namespace using the stashcp command. If you log onto 
> an ifarm machine at the lab, you can poke around on 
> /cvmfs/gluex.osgstorage.org/gluex 
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__gluex.osgstorage.org_gluex&d=DwMFaQ&c=CJqEzB1piLOyyvZjb8YUQw&r=Te_hCR4EUlJ6iCDYLJ8Viv2aDOR7D9ZZMoBAvf2H0M4&m=2oEolNTMkADUFZV21KX3t875hunlJB8Xll9JSnHrs6U8hK2lILwL9Y7klqBibkxJ&s=OVDdt1Bt4yRmZhJL7NtZE68NN5hAsTLoSsSgp6a3d5k&e=> 
> and see what is currently stored there, maybe 120 TB of various bits 
> and pieces. There is about 600 TB of additional space available at 
> present, so there is plenty of room for anything you would like to add.
>
> -Richard Jones
>
> _______________________________________________
> Halld-offline mailing list
> Halld-offline at jlab.org
> https://mailman.jlab.org/mailman/listinfo/halld-offline
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.jlab.org/pipermail/halld-offline/attachments/20211116/00266df7/attachment.html>


More information about the Halld-offline mailing list