...
Jira Legacy | ||||||
---|---|---|---|---|---|---|
|
DRAFT!
The Data access for the LHC experiments and increasing numbers of other HEP and astronomy communities is provided at UK Tier-1 facility at RAL provides data access to the LHC and HEP communities via through its ECHO storage.
The storage ECHO - currently in excess of 40PB of usable space - is deployed as a Ceph-backed erasure-coded object store, with frontend access to data is provided via XRootD - using the XrdCeph plugin - or gridFTP, via the libradosstriper library within of Ceph.
The storage must service the needs of: high-throughput compute, with staged and direct file access passing through an XCache on each workernode; data access to compute running at storageless sites, increasing utilising XCaches; and, managed inter-site data transfers using the recently adopted HTTP HTTPs protocol (using WebDav), including multihop data transfers to and from RAL’s recently newly commissioned CTA tape endpoint to external sites.
A review of the experiences of running an Object Store within these HEP data workflows, is presented, including the details of the improvements necessary improvements for the transition to WebDav from GridFTP for most inter-site data movements, and enhancements for direct-IO access.
For , where the development and optimisation of buffering and range coalescence strategies is explored.
In addition to serving the requirements of LHC Run-3, preparations for Run-4, and for large astronomy experiments is underway. One example is for ROOT-based data formats, the evolution from a TTree to RNTuple data structure provides opportunities an opportunity for storage providers to optimise and benchmark against this new format. A comparison of the current performance between data formats within ECHO is made presented and the details of potential improvements explored.
...