Data Processing Overview¶
The imaging data and catalogs included in DP0.2 are based on the wide-fast-deep simulated survey generated as part of the LSST Dark Energy Science Collaboration (DESC) Data Challenge 2 (DC2). This page provides a brief overview of the production and processing of those data. A much more comprehensive description of DC2 can be found in LSST Dark Energy Science Collaboration (2020). This page draws heavily from those sources.
Image simulation¶
The DESC DC2 image simulations are described in Section 6 of the DC2 Paper.
These data consist of simulated LSST images generated for 5 years of LSST wide-fast-deep observations covering ~300 deg2 of the sky centered at (RA, DEC)
= 61.863, -35.790
degrees.
Images are simulated by the imSim software package, a modular Python code that calls the GalSim software library (Rowe et al. 2015) for astronomical object rendering
and is run in the LSST Science Pipelines and LSST Simulation Framework software environment (Connolly et al. 2014).
The LSST software libraries provide the telescope and hardware-specific information necessary to simulate an LSST exposure, such as pixel coordinates on the focal plane, telescope filter characteristics, and the brightness of the sky.
Using that description of LSSTCam, imSim produces output files that simulate the pixel data after readout.
While imSim includes many realistic aspects of the LSST imaging (e.g., CCD geometry, electronic readout, cosmic rays, and bleed trails), there are some aspects such as scattered and reflected light from bright objects that are not included.
The input catalogs of astronomical sources are described in Section 5 of the DC2 paper.
There are several astronomical inputs into the DC2 image simulations. The extra-galactic catalog is described in detail in Korytov et al. (2019) and is based on the Outer Rim dark-matter-only cosmological simulations (Heitmann et al. 2019). Supernovae are also included in the DC2 data, modeled with a volumetric rate and time evolution that matches observed data (Dilday et al. 2010, Guy et al. 2007). Stars are drawn from the Galfast model of Juric et al. (2008) with approximately 10% of the stellar sources assumed to be variable. Solar system objects are not included. A flat sky background is assumed in each filter.
Single-image processing¶
For DP0.2, the images simulated by DESC were processed by Rubin staff with the Data Release Production (DRP) pipeline included as part of the v23.0.1 release of the LSST Science Pipelines. More detailed descriptions of the LSST Science Pipelines can be found in Bosch et al. (2018) and Bosch et al. (2019). Briefly, image processing with the DRP pipeline involves four major steps – single-frame processing, calibration, image coaddition, and coadd processing. For single-frame processing, individual visits are processed on a per-CCD basis. This step starts with instrument signature removal (ISR) that consists of bias subtraction, crosstalk correction, non-linearity correction, flat fielding, brighter-fatter correction, and masking of bad and saturated pixels (see Appendix A of the DC2 paper for details). ISR is followed by an image characterization step that performs background estimation and subtraction, Point-Spread Function (PSF) modeling, cosmic-ray detection and removal, measuring and applying aperture corrections, and source detection, deblending, and measurement. Various measurement algorithms are applied including centroiding, aperture photometry, PSF photometry, model photometry, and shape fitting. The image catalogs are compared to a reference catalog to generate photometric and astrometric calibrations for the images and associated catalogs. For DC2, the photometric and astrometric calibrations are based on a simulated reference catalog. The resulting calibrated images are known formally as “Processed Visit Images” (PVIs) and informally as “calibrated exposures” (calexps).
Coadded-image processing¶
Calibrated images are resampled onto a common pixel grid on the sky and combined to generate deeper coadded images. The coadd image grid is defined in terms of “tracts” and “patches”, where each tract is composed of 7 × 7 patches, and each patch is 4,100 × 4,100 pixels with a pixel scale of 0.2 arcsec (note that these choices are specific to DP0.2; the choice of tract and patch sizes is configurable in the Science Pipelines). Tracts have dimensions of 1.6 degrees on a side, while patches are ~13.7 arcmin on a side (roughly the size of a CCD). Patches overlap by 100 pixels along each edge so that objects lying on the edge of one patch are typically fully contained on the neighboring overlapping patch. Similarly, tracts overlap their neighbors by 1 arcmin. When producing the coadded images variable sources and artifacts are removed using resampled PSF-matched images to produce a static image of the sky. As described in Aihara et al. (2019), each image is resampled, PSF-matched, and stacked into a 2.5-sigma-clipped mean coadd that serves as a model of the static scene. A difference image was created for each image with respect to this model to identify regions associated with transient detections that only appear in a small number of epochs. With these regions identified, the final coadded image is created as a weighted mean stack of images where the transient detections are ignored. The PSF at a given location in the coadded image is calculated by taking a weighted sum of the PSFs from individual visits that have been resampled and weighted in the same way as the coadds. Regions that have clipped areas will not have the correct PSF, and these are flagged for individual objects. Before individual images are combined to form the coadd, an empirical background model is fit to the entire focal plane to control the extent to which extended features are included in the background model.
Coadded catalog production¶
The coadd catalog creation consists of five main steps: (1) above-threshold detection in each band, (2) merging the detections across bands, (3) deblending the merged detections to generate “objects” and measuring object properties in each band, (4) identifying a reference band for each object and merging the per-band catalogs into a single object catalog to use for forced photometry, and (5) performing forced measurements in each band using the reference band positions and shapes. This last step produces a catalog of independent per-band object measurements that are provided to science users.
Difference Image Analysis (DIA) production¶
The version (23.0.1) of the LSST Science Pipelines that was used for DP0.2 processing now includes portions of the Prompt Processing pipelines, which centers on template image subtraction and transient detection on the resulting difference images. A template coadd image is subtracted from each science observation (in LSST survey operations, templates will be created from the previous year of observations). The template images are resampled to the coordinate system of the science image, then convolved with a kernel to produce an image whose PSF matches that of the new science image. Once the image subtraction has been performed, similar algorithms to those used in Data Release Processing are used to detect and measure sources on the resulting difference image. These source detections and measurements make up the _DIASource_ catalog, and the _DIAObject_ table is made up of _DIASources_ associated by sky coordinate. The _DIAObject_ table includes statistical summary parameters of the associated _DIASources_ (i.e., lightcurve properties).
Matching the Object and Truth Tables¶
The full details of how the matching was done are embedded in the code matcher_probabilistic.py, and the configurable settings are documented in the matcher’s configuration class.
To summarize, the algorithm starts with the brightest true objects and only attempts to match true objects with “total magnitudes” (the AB magnitude from the summed flux in all six filters, ugrizy) brighter than 27th magnitude. This was done to avoid spurious matches with undetectable true objects, and is why the “match_candidate” column is false for some true objects in the “MatchesTruth” table.
The maximum match radius was not changed from the default used for DP0.1, 0.5 arcseconds, and the best match is the measured object with the lowest reduced chi-squared within the maximum match radius that has not already been matched to a brighter reference object. The “MatchesTruth” table contains the “match_chisq” column, and the matching considers both coordinates and cModel fluxes – although in practice, Rubin staff found that matching to photometry only made a difference for <1% of objects, because the astrometry was much more precise, and because most true objects only had one match candidate within the maximum match radius anyway. The “match_chisq” column is only relevant if there are multiple measured objects considered in the matching process for the true object (i.e., if “match_count” is greater than 1). As a side note, the matching was actually done in pixel coordinates due to the current lack of uncertainties for sky coordinates (but in the future, RA and Dec will have uncertainties).
As a final note, the matcher can only match on coordinate and flux columns that are finite for a given measured object (i.e., not “NaN”). There is a default configuration setting for the matching algorithm that requires at least three finite columns to compute the (reduced) chi-squared. This basically requires at least one finite flux, because the two centroid columns must be finite or no match is possible. Therefore, any objects that had a “NaN” CModel flux in every band could not be matched, even if there was a reference object within the match radius. The column “match_n_chisq_finite” contains how many columns were finite.