Please use this identifier to cite or link to this item:
Title: Development of DKB ETL module in case of data conversion
Authors: Kaida, A. Y.
Golosova, M. V.
Grigorieva, M. A.
Gubin, Maksim Yurjevich
Keywords: модули; большие данные; обработка данных; хранение; методанные; агрегация; программные модули
Issue Date: 2018
Publisher: IOP Publishing
Citation: Development of DKB ETL module in case of data conversion / A. Y. Kaida [et al.] // Journal of Physics: Conference Series. — Bristol : IOP Publishing, 2018. — Vol. 1015 : Information Technologies in Business and Industry (ITBI2018) : International Conference, January 17-20, 2018, Tomsk, Russian Federation : [proceedings]. — [032055, 5 p.].
Abstract: Modern scientific experiments involve the producing of huge volumes of data that requires new approaches in data processing and storage. These data themselves, as well as their processing and storage, are accompanied by a valuable amount of additional information, called metadata, distributed over multiple informational systems and repositories, and having a complicated, heterogeneous structure. Gathering these metadata for experiments in the field of high energy nuclear physics (HENP) is a complex issue, requiring the quest for solutions outside the box. One of the tasks is to integrate metadata from different repositories into some kind of a central storage. During the integration process, metadata taken from original source repositories go through several processing steps: metadata aggregation, transformation according to the current data model and loading it to the general storage in a standardized form. The R&D project of ATLAS experiment on LHC, Data Knowledge Base, is aimed to provide fast and easy access to significant information about LHC experiments for the scientific community. The data integration subsystem, being developed for the DKB project, can be represented as a number of particular pipelines, arranging data flow from data sources to the main DKB storage. The data transformation process, represented by a single pipeline, can be considered as a number of successive data transformation steps, where each step is implemented as an individual program module. This article outlines the specifics of program modules, used in the dataflow, and describes one of the modules developed and integrated into the data integration subsystem of DKB.
Appears in Collections:Материалы конференций

Files in This Item:
File SizeFormat,13 kBAdobe PDFView/Open

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.