Common DFOS tools:
|dfos = Data Flow Operations System, the common tool set for DFO|
|phoenix: the process|
1. Internal data products (IDP)
Until September 2011, the QC group maintained a process for supervised, quality-checked, PI-package oriented processing of science data. When this process was terminated in October 2011, it was replaced by a service to associate science data to quality-checked and certified master calibrations. In its first step, these associations to master calibrations were expanded to their parent raw files, and the complete set of raw files (calib and science) was delivered to the archive user for further processing in their home institutes, e.g. with REFLEX workflow tools and the ESO pipelines.
At the same time (in 2011), a scheme was built to accomodate the external data products delivered by the survey PIs and offer their data products (images, mosaics, catalogues) to the community. These so-called EDPs were supported by a new ESO infrastructure called the Phase 3 infrastructure. While initially these EDPs were images created from the public surveys at the VISTA and VST survey telescopes, since 2013 also spectral EDPs are returned, from the first spectroscopic surveys executed with FLAMES.
An unsupervised (automatic) processing of science data to the science-grade level was proposed by R. Hanuschik in 2012. It was implemented in May 2013 on muc08, for the UVES ECHELLE mode. These science-grade data products are called Internal Data Products (IDPs). Due to its success and popularity among the community, the list of supported instruments is growing and meanwhile includes XSHOOTER ECHELLE, GIRAFFE MEDUSA, MUSE IFU etc. The IDP processes currently focus on spectroscopic modes.
With its version 2.0, phoenix can also be used to create huge batches of master calibrations. This can be necessary e.g. for historical periods of an instruments when the pipeline was not yet mature, or when no historical master calibrations were stored in the archive, or when the existing master calibrations are not compatible with the current pipeline. The main goal remains - creation of science-grade data products. Hence a master calibration phoenix project is always followed by an IDP project. As an example, Figure 1 shows how phoenix has been used to close the gap of missing (not archived, or never created) master calibrations, and then extend an existing IDP project to the full history.
The version 3.x of phoenix implements the functionality to combine several OB-based products in a DEEP product ("DEEP" mode). This is the most challenging mode of phoenix since it requires multi-step combination of many IDPs into a final, deep one.
The version 4.x of phoenix implements the COMPLETE_MODE as an combination of the MCALIB mode and the IDP mode, by processing in combined batches all master calibrations required for the science IDPs, and the IDPs themselves. This new mode has implemented also the possibility to create CALIB and SCIENCE ABs by using calSelector associations. The version 4.1 of phoenix has implemented the option for creating SCIENCE ABs, for the upcoming separation of the QC1 process and the IDP process.
2. Data product standards
In order to extract the proper information from the headers, and to store and offer these metadata, the phase3 process needs data standards. The spectroscopic data standard is defined in this document.
3. How to process IDPs: concepts of PHOENIX
All IDPs are created by a dedicated workflow tool called phoenix, on a PHOENIX account (on muc08, muc09, muc10, muc11). An IDP process always needs the following main components:
Each PHOENIX account requires a configuration of the supported instrument mode (in config.phoenix) and other details of the processing. The tool is then called for a specific date. It downloads the stored ABs, the raw data and the mcalibs, processes them, creates QC reports and scores (if available), and finally ingests the products with the phase3 ingestion tool.
The PHOENIX process is extracting QC parameters in order to monitor the process stability and quality, and in some cases also the quality of individual products, as long as they can be assessed automatically (saturation).
If used for master calibrations, the tool is called for a configurable pool of data, e.g. a month. It creates all ABs from the configured OCA rules (which might be fine-tuned versions of the DFOS_OPS rules and may even be versioned). Further steps include the autoCertifier which implements some concepts of the dfos certifyProducts tool.
The PHOENIX process for a given instrument mode has two stages:
The processing history is stored on a specific website in the histoMonitor style (examples for UVES, other IDP projects linked from there).
The IDPs are ingested into the phase3 infrastructure of the archive. This means that their properties, as defined by specific metadata, can be queried on standardized web interfaces, and that external users can download them directly. Technically, this requires a set of ingestion tools which are part of each PHOENIX installation. The main entry point for Phase 3 spectral products is here.
Each IDP data stream has a release description which is written by the QC scientist and describes, for the external user, all details of the data processing. This is the same release description as required for EDPs. Find examples for UVES and XSHOOTER here.
If historical master calibrations have been created with phoenix, they are ingested in the standard dfos way, after a careful check for previous versions which need to be deleted before the new version can be ingested.
5. How to query for IDPs
Go to the Phase 3 spectral product interface and select by target, coordinates, etc. The data download is provided by the request handler and requires authentication on the User Portal.
|Last update: April 26, 2021 by rhanusch|