Distributed data management on Belle II

The Belle II experiment started taking physics data in April 2018 with an estimated total volume of all files including raw events, Monte-Carlo and skim statistics of 340 petabytes expected by the end of operations in the late-2020s. Originally designed as a fully integrated component of the BelleDI...

Full description

Saved in:
Bibliographic Details
Published in:EPJ Web of Conferences Vol. 245; p. 4007
Main Authors: Padolski, Siarhei, Ito, Hironori, Laycock, Paul, Mashinistov, Ruslan, Miyake, Hideki, Ueda, Ikuo
Format: Journal Article Conference Proceeding
Language:English
Published: Les Ulis EDP Sciences 2020
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The Belle II experiment started taking physics data in April 2018 with an estimated total volume of all files including raw events, Monte-Carlo and skim statistics of 340 petabytes expected by the end of operations in the late-2020s. Originally designed as a fully integrated component of the BelleDIRAC production system, the Belle II distributed data management (DDM) software needs to manage data across about 29 storage elements worldwide for a collaboration of nearly 1000 physicists. By late 2018, this software required significant performance improvements to meet the requirements of physics data taking and was seriously lacking in automation. Rucio, the DDM solution created by ATLAS, was an obvious alternative but required tight integration with BelleDIRAC and a seamless yet non-trivial migration. This contribution describes the work done on both DDM options, the current status of the software running successfully in production and the problems associated with trying to balance long-term operations cost against short term risk.
ISSN:2100-014X
2101-6275
2100-014X
DOI:10.1051/epjconf/202024504007