Abstract

The increasingly larger data volumes that the LHC experiments will accumulate in the coming years, especially in the High-Luminosity LHC era, call for a paradigm shift in the way experimental datasets are accessed and analyzed. The current model, based on data reduction on the Grid infrastructure, followed by interactive data analysis of manageable size samples on the physicists’ individual computers, will be superseded by the adoption of Analysis Facilities. This rapidly evolving concept is converging to include dedicated hardware infrastructures and computing services optimized for the effective analysis of large HEP data samples. This paper describes the actual implementation of this new analysis facility model at the CIEMAT institute, in Spain, to support the local CMS experiment community. Our work details the deployment of dedicated highly performant hardware, the operation of data staging and caching services ensuring prompt and efficient access to CMS physics analysis datasets, and the integration and optimization of a custom analysis framework based on ROOT’s RDataFrame and CMS NanoAOD format. Finally, performance results obtained by benchmarking the deployed infrastructure and software against a CMS analysis workflow are summarized.

Details

Title
The Spanish CMS Analysis Facility at CIEMAT
Author
Cárdenas-Montes, M; A. Delgado Peris; Flix, J; Hernández, J M; J. León Holgado; C. Morcillo Pérez; Pérez-Calero Yzquierdo, A; F.J. Rodríguez Calonge
Section
Facilities and Virtualization
Publication year
2024
Publication date
2024
Publisher
EDP Sciences
ISSN
21016275
e-ISSN
2100014X
Source type
Conference Paper
Language of publication
English
ProQuest document ID
3057079745
Copyright
© 2024. This work is licensed under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and conditions, you may use this content in accordance with the terms of the License.