LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions
waived without prior approval.
1
Large Synoptic Survey Telescope (LSST)
Data Products Definition Document
LSE-163
Latest Revision Date: October 7, 2013
This LSST document has been approved as a Content-Controlled Document. Its contents are subject to
configuration control and may not be changed, altered, or their provisions waived without prior
approval. If this document is changed or superseded, the new document will retain the Handle
designation shown above. The control is on the most recent digital document with this Handle in the
LSST digital archive and not printed versions.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions
waived without prior approval.
i
Change Record
Version
Date
Description
Owner name
1
10/7/2013
Initial Release
Mario Juric
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
Large Synoptic Survey Telescope
Data Products De?nition Document
LSST Document LSE-163
Mario Juri?c
?
, R.H. Lupton, T. Axelrod, J.F. Bosch,
G.P. Dubois-Felsmann, Z.? Ivezi?c, A.C. Becker, J. Becla,
A.J. Connolly, M. Freemon, J. Kantor, K-T Lim, D. Shaw,
M. Strauss, and J.A. Tyson
for the LSST Project
October 7, 2013
Abstract
This document describes the data products and processing services
to be delivered by the Large Synoptic Survey Telescope (LSST).
The LSST will deliver three levels of data products and services.
Level 1 (nightly) data products will include images, di?erence im-
ages, catalogs of sources and objects detected in di?erence images,
and catalogs of Solar System objects. Their primary purpose is to
enable rapid follow-up of time-domain events. Level 2 (annual) data
products will include well calibrated single-epoch images, deep coadds,
and catalogs of objects, sources, and forced sources, enabling static
sky and precision time-domain science. Level 3 (user-created) data
product services will enable science cases that greatly bene?t from
co-location of user processing and/or data within the LSST Archive
Center. LSST will also devote 10% of observing time to programs
with special cadence. Their data products will be created using the
same software and hardware as Levels 1 and 2. All data products will
be made available using user-friendly databases and web services.
?
Please direct comments to
<mjuric@lsst.org>.
1
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
CONTENTS
2
Contents
1 Preface
4
2 Introduction
5
2.1 The Large Synoptic Survey Telescope . . . . . . . . . . . . . .
5
2.2 Classes of LSST Data Products . . . . . . . . . . . . . . . . .
6
3 General Considerations
9
3.1 Estimator and Naming Conventions . . . . . . . . . . . . . . .
9
3.2 Image Characterization Data . . . . . . . . . . . . . . . . . . . 10
3.3 Fluxes and Magnitudes . . . . . . . . . . . . . . . . . . . . . . 10
3.4 Uniqueness of IDs across database versions . . . . . . . . . . . 11
3.5 Repeatability of Queries . . . . . . . . . . . . . . . . . . . . . 11
4 Level 1 Data Products
13
4.1 Overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
4.2 Level 1 Data Processing . . . . . . . . . . . . . . . . . . . . . 14
4.2.1 Di?erence Image Analysis . . . . . . . . . . . . . . . . 14
4.2.2 Solar System Object Processing . . . . . . . . . . . . . 16
4.3 Level1Catalogs . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.3.1 DIASource Table . . . . . . . . . . . . . . . . . . . . . 18
4.3.2 DIAObject Table . . . . . . . . . . . . . . . . . . . . . 23
4.3.3 SSObject Table . . . . . . . . . . . . . . . . . . . . . . 24
4.3.4 Precovery Measurements . . . . . . . . . . . . . . . . . 26
4.3.5 Reprocessing the Level 1 Data Set . . . . . . . . . . . . 26
4.4 Level 1 Image Products. . . . . . . . . . . . . . . . . . . . . . 28
4.4.1 VisitImages . . . . . . . . . . . . . . . . . . . . . . . . 28
4.4.2 Di?erence Images . . . . . . . . . . . . . . . . . . . . . 28
4.4.3 Image Di?erencing Templates . . . . . . . . . . . . . . 28
4.5 Alertsto DIASources . . . . . . . . . . . . . . . . . . . . . . . 29
4.5.1 Information Contained in Each Alert . . . . . . . . . . 29
4.5.2 Receiving and Filtering the Alerts . . . . . . . . . . . . 30
5 Level 2 Data Products
32
5.1 Overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
5.2 Level 2 Data Processing . . . . . . . . . . . . . . . . . . . . . 33
5.2.1 Object Characterization Measures . . . . . . . . . . . . 36
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
CONTENTS
3
5.2.2 Supporting Science Cases Requiring Full Posteriors . . 38
5.2.3 Source Characterization . . . . . . . . . . . . . . . . . 39
5.2.4 Forced Photometry . . . . . . . . . . . . . . . . . . . . 40
5.2.5 Crowded Field Photometry . . . . . . . . . . . . . . . 40
5.3 TheLevel2Catalogs . . . . . . . . . . . . . . . . . . . . . . . 41
5.3.1 The Object Table. . . . . . . . . . . . . . . . . . . . . 41
5.3.2 Source Table . . . . . . . . . . . . . . . . . . . . . . . 47
5.3.3 ForcedSource Table . . . . . . . . . . . . . . . . . . . 49
5.4 Level 2 Image Products. . . . . . . . . . . . . . . . . . . . . . 49
5.4.1 VisitImages . . . . . . . . . . . . . . . . . . . . . . . . 49
5.4.2 Calibration Data . . . . . . . . . . . . . . . . . . . . . 50
5.4.3 CoaddedImages . . . . . . . . . . . . . . . . . . . . . 50
5.5 Data Release Availability and Retention Policies . . . . . . . . 52
6 Level 3 Data Products and Capabilities
54
6.1 Level 3 Data Products and Associated Storage Resources . . . 54
6.2 Level 3 Processing Resources . . . . . . . . . . . . . . . . . . . 55
6.3 Level 3 Programming Environment and Framework . . . . . . 56
6.4 Migration of Level 3 data products to Level 2 . . . . . . . . . 58
7 Data Products for Special Programs
59
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
1 PREFACE
4
1 Preface
The purpose of this document is to describe the data products produced by
the Large Synoptic Survey Telescope (LSST).
To a future LSST user, it should clarify what catalogs, image data, soft-
ware, and services they can expect from LSST. To LSST builders, it provides
direction on how to ?ow down the LSST System Requirements Document
to system design, sizing, budget and schedule as they pertain to the data
products.
Though under strict change control
1
, this is a living document . LSST
will undergo a period of construction and commissioning lasting no less than
seven years, followed by a decade of survey operations. To ensure their
continued scienti?c adequacy, the designs and plans for LSST Data Products
will be periodically reviewed and updated.
1
LSST Docushare handle for this document is LSE-163 .
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
2 INTRODUCTION
5
2 Introduction
2.1 The Large Synoptic Survey Telescope
LSST will be a large, wide-?eld ground-based optical telescope system de-
signed to obtain multiple images covering the sky that is visible from Cerro
Pach?on in Northern Chile. The current baseline design, with an 8.4m (6.7m
e?ective) primary mirror, a 9.6 deg
2
?eld of view, and a 3.2 Gigapixel cam-
era, will allow about 10,000 square degrees of sky to be covered every night
using pairs of 15-second exposures, with typical 5˙ depth for point sources
of r ˘ 24:5 (AB). The system is designed to yield high image quality as well
as superb astrometric and photometric accuracy. The total survey area will
include ˘30,000 deg
2
with ? < +34:5
?
, and will be imaged multiple times in
six bands, ugrizy, covering the wavelength range 320{1050 nm.
The project is scheduled to begin the regular survey operations at the
start of next decade. About 90% of the observing time will be devoted to
a deep-wide-fast survey mode which will uniformly observe a 18,000 deg
2
region about 1000 times (summed over all six bands) during the anticipated
10 years of operations, and yield a coadded map to r ˘ 27:5. These data
will result in catalogs including over 38 billion stars and galaxies, that will
serve the majority of the primary science programs. The remaining 10% of
the observing time will be allocated to special projects such as a Very Deep
and Fast time domain survey
2
.
The LSST will be operated in fully automated survey mode. The images
acquired by the LSST Camera will be processed by LSST Data Management
software to a) detect and characterize imaged astrophysical sources and b) de-
tect and characterize temporal changes in the LSST-observed universe. The
results of that processing will be reduced images, catalogs of detected objects
and the measurements of their properties, and prompt alerts to \events" {
changes in astrophysical scenery discovered by di?erencing incoming images
against older, deeper, images of the sky in the same direction (templates, see
x4.4.3). Measurements will be internally and absolutely calibrated.
The broad, high-level, requirements for LSST Data Products are given by
the LSST Science Requirements Document
3
(SRD). This document lays out
the speci?cs of what the data products will comprise of, how those data will
2
Informally known as \Deep Drilling Fields".
3
LSST Document Handle LPM-17, available at
http://ls.st/srd
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
2 INTRODUCTION
6
be generated, and when. It serves to inform the ?ow-down from the LSST
SRD through the LSST System Requirements Document (the LSR; LSE-29)
and the LSST Observatory System Speci?cations (OSS; LSE-30), to the LSST
Data Management System Requirements (DMSR; LSE-61), the UML model,
and the database schema.
2.2 Classes of LSST Data Products
LSST Data Management will perform two, somewhat overlapping in scienti?c
intent, types of image analyses:
1. Analysis of di?erence images, with the goal of detecting and charac-
terizing astrophysical phenomena revealed by their time-dependent na-
ture. The detection of supernovae superimposed on bright extended
galaxies is an example of this analysis. The processing will be done
on nightly or daily basis and result in Level 1 data products. Level 1
products will include di?erence images, catalogs of sources detected in
di?erence images ( DIASources ), astrophysical objects
4
these are asso-
ciated to ( DIAObjects ), and Solar System objects ( SSObjects
5
). The
catalogs will entered into the Level 1 database and made available
in near real time. Noti?cations (\alerts") about new DIASources will
be issued using community-accepted standards within 60 seconds of
observation. Level 1 data products are discussed in x 4.
2. Analysis of direct images, with the goal of detecting and characteriz-
ing astrophysical objects. Detection of faint galaxies on deep coadds
and their subsequent characterization is an example of this analysis.
The results are Level 2 data products. These products, generated and
released annually
6
, will include the single-epoch images, deep coadds,
catalogs of characterized Objects (detected on deep coadds as well as
4
The LSST has adopted the nomenclature by which single-epoch detections of astro-
physical objects are called sources. The reader is cautioned that this nomenclature is not
universal: some surveys call detections what LSST calls sources, and use the term sources
for what LSST calls objects.
5
SSObjects used to be called call \Moving Objects" in previous versions of the LSST
Data Products baseline. The name is potentially confusing as high-proper motion stars
are moving objects as well. A more accurate distinction is the one between objects inside
and outside of the Solar System.
6
Except for the ?rst two data releases, which will be created six months apart.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
2 INTRODUCTION
7
individual visits
7
), Sources
8
(detections and measurements on individ-
ual visits), and ForcedSources (constrained measurement of ?ux on
individual visits). It will also include fully reprocessed Level 1 data
products (see x4.3.5). In contrast to the Level 1 database, which is up-
dated in real-time, the Level 2 databases are static and will not change
after release. Level 2 data products are discussed in x 5.
The two types of analyses have di?erent requirements on timeliness.
Changes in ?ux or position of objects may need to be immediately followed
up, lest interesting information be lost. Thus the primary results of analysis
of di?erence images { discovered and characterized DIASources { generally
need to be broadcast as event alerts within 60 seconds of end of visit acqui-
sition. The analysis of science (direct) images is less time sensitive, and will
be done as a part of annual data release process.
Recognizing the diversity of astronomical community needs, and the need
for specialized processing not part of the automatically generated Level 1
and 2 products, LSST plans to devote 10% of its data management system
capabilities to enabling the creation, use, and federation of Level 3 (user-
created) data products. Level 3 capabilities will enable science cases that
greatly bene?t from co-location of user processing and/or data within the
LSST Archive Center. The high-level requirement for Level 3 is established
in x 3.5 of the LSST SRD. Their details are discussed in x 6 of this document.
Finally, LSST Survey Speci?cations (x 3.4 of LSST SRD) prescribe that
90% of LSST observing time be spent in the so-called \universal cadence"
mode of surveying the sky. These observations will result in Level 1 and 2
data products discussed above. The remaining 10% of observing time will
be devoted to special programs , designed to obtain improved coverage
of interesting regions of observational parameter space. Examples include
very deep (r ˘ 26, per exposure) observations, observations with very short
revisit times (˘1 minute), and observations of \special" regions such as the
Ecliptic, Galactic plane, and the Large and Small Magellanic Clouds. The
data products for these programs will be generated using the same processing
7
The LSST takes two exposures per pointing, nominally 15 seconds in duration each,
called snaps. For the purpose of data processing, that pair of exposures will typically be
coadded and treated as a single exposure, called a visit.
8
When written in bold monospace type (i.e., n tt ), Objects and Sources refer to objects
and sources detected and measured as a part of Level 2 processing.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
2 INTRODUCTION
8
software and hardware and possess the general characteristics of Level 1 and
2 data products, but may be performed on a somewhat di?erent cadence.
They will be discussed in x 7.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
3 GENERAL CONSIDERATIONS
9
3 General Considerations
Most LSST data products will consist of images and/or catalogs. The cata-
logs will be stored and o?ered to the users as relational databases which they
will be able to query. This approach was shown to work well by prior large
surveys, for example the Sloan Digital Sky Survey (SDSS).
Di?erent data products will generally be stored in di?erent databases. For
example, Level 1 data products will be stored in a Level 1 database. Level 2
\universal cadence" products will be stored in a Level 2 database database.
The products for special programs may be stored in many di?erent databases,
depending on the nature of the program.
Nevertheless, all these databases will follow certain naming and other
conventions. We discuss these in the subsections to follow.
3.1 Estimator and Naming Conventions
For all catalogs data, we will employ a convention where estimates of stan-
dard errors have the su?x Err , while the estimates of inherent widths of
distribution (or functions in general) have the su?x Sigma
9
. The latter are
de?ned as the square roots of the second moment about the quoted value of
the quantity at hand.
Unless noted otherwise, maximum likelihood values will be quoted for
all ?tted parameters (measurements). Together with covariances, these let
the end-user apply whatever prior they deem appropriate when computing
posteriors
10
. Where appropriate, multiple independent samples from the
likelihood may be provided to characterize departures from Gaussianity.
We will provide values of log likelihood, the ˜
2
for the ?tted parameters,
and the number of data points used in the ?t. Database functions (or precom-
puted columns) will be provided for frequently used combinations of these
quantities (e.g., ˜
2
=dof). These can be used to assess the model ?t quality.
Note that, if the errors of measured quantities are normally distributed, the
likelihood is related to the ˜
2
as:
L=
Y
k
1
˙
k
p
2ˇ
!
exp
?
?
˜
2
2
?
(1)
9
Given N measurements, standard errors scale as N
? 1 = 2
, while widths remain constant.
10
There's a tacit assumption that a Gaussian is a reasonably good description of the
likelihood surface around the ML peak.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
3 GENERAL CONSIDERATIONS
10
where the index k runs over all data points included in the ?t.
For ?uxes, we recognize that a substantial fraction of astronomers will just
want the posteriors marginalized over all other parameters, trusting the LSST
experts to select an appropriate prior
11
. For example, this is nearly always
the case when constructing color-color or color-magnitude diagrams. We
will support these use cases by providing additional pre-computed columns,
taking care to name them appropriately so as to minimize accidental incorrect
usage. For example, a column named gFlux may be the expectation value
of the g-band ?ux, while gFluxML may represent the maximum likelihood
value.
3.2 Image Characterization Data
Raw images will be processed to remove instrumental signature and char-
acterize their properties, including backgrounds (both due to night sky and
astrophysical), the point spread function and its variation, photometric zero-
point model, and the world coordinate system (WCS).
That characterization is crucial for deriving LSST catalogs and under-
standing the images. It will be kept and made available to the users. The
exact format used to store these (meta)data will depend on the ?nal adopted
algorithm in consultation with the scienti?c community to ensure the formats
in which these data are served are maximally useful.
Each processed image
12
, including the coadds, will record information on
pixel variance (the \variance plane"), as well as per-pixel masks (the \mask
plane"). These will allow the users to determine the validity and usefullness
of each pixel in estimating the ?ux density recorded in that area of the sky.
This information will be per-pixel, and potentially unwieldy to use for
certain science cases. We plan to investigate approximate schemes for storing
masks based on geometry (e.g., similar to Mangle or STOMP), in addition
to storing them on a per pixel basis.
3.3 Fluxes and Magnitudes
Because ?ux measurements on di?erence images (Level 1 data products; x 4)
are performed against a template, the measured ?ux of a source on the dif-
11
It's likely that most cases will require just the expectation value alone.
12
It is also frequently referred to as calibrated exposure
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
3 GENERAL CONSIDERATIONS
11
ference image can be negative. The ?ux can also go negative for faint sources
in the presence of noise. Negative ?uxes cannot be stored as (Pogson) mag-
nitudes; log of a negative number is unde?ned. We therefore prefer to store
?uxes, rather than magnitudes, in database tables
13
.
We quote ?uxes in units of \maggie". A maggie, as introduced by SDSS,
is a linear measure of ?ux. It is de?ned so that an object having a ?ux of
one maggie (integrated over the bandpass) has an AB magnitude of zero:
m
AB
= ?2:5log
10
(f=maggie)
(2)
We chose to use maggies (as opposed to, say, Jansky) to allow the user
to di?erentiate between two distinct sources of photometric calibration er-
ror: the error in relative (internal) calibration of the survey, and the error
in absolute calibration that depends on the knowledge of absolute ?ux of
photometric standards.
Nevertheless, we acknowledge that the large majority of users will want
to work with magnitudes. For convenience, we plan to provide columns
with (Pogson) magnitudes
14
, where values with negative ?ux will evaluate
to NULL . Similarly, we will provide columns with ?ux expressed in Jy (and
its error estimate that includes the relative and absolute calibration error
contributions).
3.4 Uniqueness of IDs across database versions
To reduce the likelihood for confusion, all IDs shall be unique across databases
and database versions, other than those corresponding to uniquely identi?-
able entities (i.e., IDs of exposures).
For example, DR4 and DR5 (or any other) release will share no iden-
tical Object , Source , DIAObject or DIASource IDs (see x 4 and 5 for the
de?nitions of Objects , DIAObjects , etc.).
3.5 Repeatability of Queries
We require that queries executed at a known point in time against any LSST-
delivered database be repeatable at a later date. This promotes the repro-
13
This is a good idea in general. Eg. given multi-epoch observations, one should always
be averaging ?uxes, rather than magnitudes.
14
These will most likely be implemented as \virtual" or \computed" columns
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
3 GENERAL CONSIDERATIONS
12
ducibility of science derived from LSST data. It is of special importance for
Level 1 catalogs (x 4) that will change on a nightly basis as new time domain
data is being processed and added to the catalogs.
The exact implementation of this requirement is left to the LSST Data
Management database team. One possibility may be to make the key tables
(nearly) append-only, with each row having two timestamps { createdTai
and deletedTai , so that queries may be limited by a WHERE clause:
SELECT * FROM DIASource WHERE 'YYYY-MM-DD-HH-mm-SS' BETWEEN
createdTAI and deletedTAI
or, more generally:
SELECT * FROM DIASource WHERE "data is valid as of YYYY-MM-DD"
A perhaps less error-prone alternative, if technically feasible, may be to
provide multiple virtual databases that the user would access as:
CONNECT lsst-dr5-yyyy-mm-dd
SELECT * FROM DIASource
The latter method would probably be limited to nightly granularity, unless
there's a mechanism to create virtual databases/views on-demand.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
13
4 Level 1 Data Products
4.1 Overview
Level 1 data products are a result of di?erence image analysis (DIA; x4.2.1).
They include the sources detected in di?erence images ( DIASources ), astro-
physical objects that these are associated to ( DIAObjects ), identi?ed Solar
System objects
15
( SSObject ), and related, broadly de?ned, metadata (in-
cluding eg., cut-outs
16
).
DIASources are sources detected on di?erence images (those above S=N =
transSNR after correlation with the PSF pro?le, with transSNR de?ned in
the SRD and presently set to 5). They represent changes in ?ux with respect
to a deep template. Physically, a DIASource may be an observation of new
astrophysical object that was not present at that position in the template
image (for example, an asteroid), or an observation of ?ux change in an ex-
isting source (for example, a variable star). Their ?ux can be negative (eg.,
if a source present in the template image reduced its brightness, or moved
away). Their shape can be complex (eg., trailed, for a source with proper
motion approaching ˘ deg=day, or \dipole-like", if an object's observed po-
sition exhibits an o?set { true or apparent { compared to its position on the
template).
Clusters of DIASources detected on visits taken at di?erent times are
associated with either a DIAObject or an SSObject , to represent the un-
derlying astrophysical phenomenon. The association can be made in two
di?erent ways: by assuming the underlying phenomenon is an object within
the Solar System moving on an orbit around the Sun
17
, or by assuming it
to be distant enough to only exhibit small parallactic and proper motion
18
.
The latter type of association is performed during di?erence image analysis
right after the image has been acquired. The former is done at daytime by
15
The SRD considers Solar System object orbit catalog to be a Level 2 data product
(LSST SRD, Sec 3.5). Nevertheless, to successfully di?erentiate between apparitions of
known Solar System objects and other types DIASources we consider it functionally a
part of Level 1.
16
Small, 30?30, sub-images at the position of a detected source. Also known as postage
stamps.
17
We don't plan to ?t for motion around other Solar System bodies; eg., identifying new
satellites of Jupiter is left to the community.
18
Where 'small' is small enough to unambiguously positionally associate together indi-
vidual apparitions of the object.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
14
the Moving Objects Processing Software ( MOPS ), unless the DIASource is an
apparition of an already known SSObject . In that case, it will be ?agged as
such during di?erence image analysis.
At the end of the di?erence image analysis of each visit, we will issue time
domain event alerts for all newly discovered DIASources
19
.
4.2 Level 1 Data Processing
4.2.1 Di?erence Image Analysis
The following is a high-level description of steps which will occur during
regular di?erence image analysis:
1. A visit is acquired and reduced to a single visit image (cosmic ray
rejection, instrumental signature removal
20
, combining of snaps, etc.).
2. The visit image is di?erenced against the appropriate template and
DIASources are detected. If necessary, deblending will be performed
at this stage. Both the parent blend and the deblended children will be
measured and stored as DIASources (see next item), but only the chil-
dren will be matched against DIAObjects and alerted on. Deblended
objects will be ?agged as such.
3. The ?ux and shape
21
of the DIASource are measured on the di?erence
image. PSF photometry is performed on the visit image at the position
of the DIASource to obtain a measure of the absolute ?ux.
4. The Level 1 database (see x4.3) is searched for a DIAObject or an
SSObject with which to positionally associate the newly discovered
DIASource
22
. If no match is found, a new DIAObject is created and
the observed DIASource is associated to it.
19
For observations on the Ecliptic near the opposition Solar System objects will dominate
the DIASource counts and (until they're recognized as such) overwhelm the explosive
transient signal. It will therefore be advantageous to quickly identify the majority of Solar
System objects early in the survey.
20
Eg., subtraction of bias and dark frames, ?at ?elding, bad pixel/column interpolation,
etc.
21
The \shape" in this context consists of weighted 2
nd
moments, as well as a ?t to a
trailed source model.
22
The association algorithm will guarantee that a DIASource is associated with not
more than one existing DIAObject or SSObject . The algorithm will take into account the
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
15
5. If the DIASource has been associated with an SSObject (a known Solar
System object), it will be ?agged as such and an alert will be issued.
Further processing will occur in daytime (see section 4.2.2).
6. Otherwise, the associated DIAObject measurements will be updated
with new data. All a?ected columns will be recomputed, including
proper motions, centroids, light curves, etc.
7. The Level 2 database
23
is searched for one or more Objects positionally
close to the DIAObject , out to some maximum radius
24
. The IDs of
these Objects are recorded in the DIAObject record and provided in
the issued event alert (see below).
8. An alert is issued that includes: the name of the Level 1 database,
the timestamp of when this database has been queried to issue this
alert, the DIASource ID, the DIAObject ID
25
, name of the Level 2
database and the IDs of nearby Objects , and the associated science
content (centroid, ?uxes, low-order lightcurve moments, periods, etc.),
including the full light curves. See Section 4.5 for a more complete
enumeration.
9. For all DIAObjects overlapping the ?eld of view to which a DIASource
from this visit has not been associated, forced photometry will be per-
formed (point source photometry only). Those measurements will be
stored as appropriately ?agged DIASources
26
. No alerts will be issued
for these DIASources .
10. Within 24 hours of discovery, precovery PSF forced photometry will
be performed on any di?erence image overlapping the position of new
parallax and proper (or Keplerian) motions, as well as the errors in estimated positions
of DIAObject , SSObject , and DIASource , to ?nd the maximally likely match. Multiple
DIASources in the same visit will not be matched to the same DIAObject .
23
Level 2 database is a database resulting from annual data release processing. See x 5
for details.
24
Eg., a few arcseconds.
25
We guarantee that a receiver will always be able to regenerate the alert contents at
any later date using the included timestamps and metadata (IDs and database names).
26
For the purposes of this document, we're treating the DIASources generated by forced
photometry or precovery measurements to be the same as DIASources detected in di?er-
ence images (but ?agged appropriately). In the logical schema, these may be divided into
two separate tables.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
16
DIAObjects taken within the past 30 days, and added to the database.
Alerts will not be issued with precovery photometry information.
In addition to the processing described above, a smaller sample of sources
detected on di?erence images below the nominal S=N = 5 threshold will
be measured and stored, in order to enable monitoring of di?erence image
analysis quality.
Also, the system will have the ability to measure and alert on a limited
27
number of sources detected below the nominal threshold for which additional
criteria are satis?ed. For example, a S=N = 3 source detection near a grav-
itational keyhole may be highly signi?cant in assessing the danger posed by
a potentially hazardous asteroid. The project will de?ne the initial set of
criteria by the start of Operations.
4.2.2 Solar System Object Processing
The following will occur during regular Solar System object processing (in
daytime
28
, after a night of observing):
1. The orbits and physical properties of all SSObjects re-observed on the
previous night are recomputed. External orbit catalogs (or observa-
tions) are used to improve orbit estimates. Updated data are entered
to the SSObjects table.
2. All DIASources detected on the previous night, that have not been
matched at a high con?dence level to a known Object , SSObject , or
an artifact, are analyzed for potential pairs, forming tracklets.
3. The collection of tracklets collected over the past 30 days is searched
for subsets forming tracks consistent with being on the same Keplerian
orbit around the Sun.
27
It will be sized for no less than ˘ 10% of average DIASource per visit rate.
28
Note that there is no strict bound on when daytime Solar System processing must
?nish, just that, averaged over some reasonable timescale (eg., a month), a night's worth
of observing is processed within 24 hours. Nights rich in moving objects may take longer
to process, while nights with less will ?nish more quickly. In other words, the requirement
is on throughput, not latency.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
17
4. For those that are, an orbit is ?tted and a new SSObject table entry
created. DIASource records are updated to point to the new SSObject
record. DIAObjects \orphaned" by this unlinking are deleted.
29
.
5. Precovery linking is attempted for all SSObjects whose orbits were
updated in this process. Where successful, SSObjects (orbits) are re-
computed as needed.
4.3 Level 1 Catalogs
The described alert processing design relies on the Level 1 database that
contains the objects and sources detected on di?erence images. At the
very least
30
, this database will have tables of DIASources , DIAObjects , and
SSObjects , populated in the course of di?erence image and Solar System
object processing
31
. As these get updated and added to, their updated con-
tents becomes visible (query-able) immediately
32
.
This database is only loosely coupled to the Level 2 database. All of the
coupling is through positional matches between the DIAObjects entries in the
Level 1 database and the Objects in the Level 2 database database. There is
no direct DIASource -to- Object match. The adopted data model emphasizes
that having a DIASource be positionally coincident with an Object does
not imply it is physically related to it. Absent other information, the least
presumptuous data model relationship is one of positional association, not
physical identity.
This may seem odd at ?rst: for example, in a simple case of a variable star,
matching individual DIASources to Objects is exactly what an astronomer
would want. That approach, however, fails in the following scenarios:
? A supernova in a galaxy. The matched object in the Object table will
be the galaxy, which is a distinct astrophysical object. We want to keep
29
Some DIAObjects may only be left with forced photometry measurements at their
location (since all DIAObjects are force-photometered on previous and subsequent visits);
these will be kept but ?agged as such.
30
It will also contain exposure and visit metadata, MOPS-speci?c tables, etc. These are
either standard/uncontroversial, implementation-dependent, or less directly relevant for
science and therefore not discussed in this document.
31
The latter is also colloquially known as DayMOPS.
32
No later than the moment of issuance of any event alert that may refer to it.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
18
the information related to the supernova (eg., colors, the light curve)
separate from those measurements for the galaxy.
? An asteroid occulting a star. If associated with the star on ?rst ap-
parition, the association would need to be dissolved when the source is
recognized as an asteroid (perhaps even as early as a day later).
? A supernova on top of a pair of blended galaxies. It is not clear in
general to which galaxy this DIASource would \belong". That in itself
is a research question.
DIASource -to- Object matches can still be emulated via a three-step re-
lation ( DIASource - DIAObject - Object ). For ease of use, views or pre-built
table with these will be o?ered to the end-users.
In the sections to follow, we present the conceptual schemas for the most
important Level 1 database tables. These convey what data will be recorded
in each table, rather than the details of how. For example, columns whose
type is an array (eg., radec ) may be expanded to one table column per el-
ement of the array (eg., ra , decl ) once this schema is translated to SQL
33
.
Secondly, the tables to be presented are largely normalized (i.e., contain no
redundant information). For example, since the band of observation can
be found by joining a DIASource table to the table with exposure meta-
data, there's no column named band in the DIASource table. In the as-built
database, the views presented to the users will be appropriately denormalized
for ease of use.
4.3.1 DIASource Table
This is a table of sources detected at SNR ? 5 on di?erence images
34
( DIASources ). On average, the LSST SRD expects ˘ 2000 DIASources
per visit (˘ 2M per night; 20,000 per deg
2
of the sky per hour).
Some SNR ? 5 sources will not be caused by observed astrophysical
phenomena, but by artifacts (bad columns, di?raction spikes, etc.). The
di?erence image analysis software will attempt to identify and ?ag these as
such.
33
The SQL realization of this schema can be browsed at
http://ls.st/8g4
34
The SNR=5 requirements is speci?ed in the LSST SRD as transSNR.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
19
Unless noted otherwise, all DIASource quantities (?uxes, centroids, etc.)
are measured on the di?erence image.
Table 1: DIASource Table
Name
Type
Unit
Description
diaSourceId
uint64
Unique source identi?er
ccdVisitId
uint64
ID of CCD and visit where
this source was measured
diaObjectId
uint64
ID of the DIAObject this
source was associated with,
if any.
ssOb jectId
uint64
ID of the SSObject this
source has been linked to, if
any.
parentSourceId
uint64
ID of the parent Source this
object has been deblended
from, if any.
midPointTai
double
time
Time of mid-exposure for
this DIASource.
radec
double[2]
degrees
(?; ?)
35
radecCov
?oat[3]
various
radec covariance matrix
xy
?oat[2]
pixels
Column and row of the cen-
troid.
xyCov
?oat[3]
various
Centroid covariance matrix
SNR
?oat
The signal-to-noise ratio at
which this source was de-
tected in the di?erence im-
age.
36
psFlux
?oat
nmgy
37
Calibrated ?ux for point
source model.
Note this
actually measures the ?ux
di?erence between the tem-
plate and the visit image.
Continued on next page
35
The astrometric reference frame will be chosen closer to start of operations.
36
This is not necessarily the same as psFlux/psFluxSigma, as the ?ux measurement
algorithm may be more accurate than the detection algorithm.
37
A \maggie", as introduced by SDSS, is a linear measure of ?ux; one maggie has an
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
20
Table 1: DIASource Table
Name
Type
Unit
Description
psFluxSigma
?oat
nmgy
Estimated uncertainty of
psFlux .
psLnL
?oat
Natural log likelihood of
the observed data given the
point source model.
psChi2
?oat
˜
2
statistic of the model ?t.
psN
int
The number of data points
(pixels) used to ?t the
model.
trailFlux
?oat
nmgy
Calibrated ?ux for a trailed
source model
38 ; 39
. Note this
actually measures the ?ux
di?erence between the tem-
plate and the visit image.
trailLength
?oat
arcsec
Maximum likelihood ?t of
trail length
40 ; 41
.
Continued on next page
AB magnitude of 0. \nmgy" is short for a nanomaggie. Flux of 0:063 nmgy corresponds
to a 24:5
th
magnitude star. See x3.3 for details.
38
A Trailed Source Model attempts to ?t a (PSF-convolved) model of a point source that
was trailed by a certain amount in some direction (taking into account the two-snap nature
of the visit, which may lead to a dip in ?ux around the mid-point of the trail). Roughly,
it's a ?t to a PSF-convolved line. The primary use case is to characterize fast-moving
Solar System objects.
39
This model does not ?t for the direction of motion; to recover it, we would need to
?t the model to separately to individual snaps of a visit. This adds to system complexity,
and is not clearly justi?ed by increased MOPS performance given the added information.
40
Note that we'll likely measure trailRow and trailCol, and transform to trail-
Length/trailAngle (or trailRa/trailDec) for storage in the database. A stretch goal is
to retain both.
41
TBD: Do we need a separate trailCentroid? It's unlikely that we do, but one may
wish to prove it.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
21
Table 1: DIASource Table
Name
Type
Unit
Description
trailAngle
?oat
degrees
Maximum likelihood ?t
of the angle between the
meridian
through
the
centroid and the trail direc-
tion (bearing, direction of
motion).
trailCov
?oat[6]
various
Covariance matrix of trailed
source model parameters.
trailLnL
?oat
Natural log likelihood of
the observed data given the
trailed source model.
trailChi2
?oat
˜
2
statistic of the model ?t.
trailN
int
The number of data points
(pixels) used to ?t the
model.
fpFlux
?oat
nmgy
Calibrated ?ux for point
source model measured on
the visit image centered at
the centroid measured on
the di?erence image (forced
photometry ?ux)
fpFluxSigma
?oat
nmgy
Estimated uncertainty of
fpFlux .
di?Flux
?oat
nmgy
Calibrated ?ux for point
source model centered on
radec but measured on the
di?erence of snaps compris-
ing this visit
42
.
di?FluxSigma
?oat
nmgy
Estimated uncertainty of
diffFlux .
Continued on next page
42
This ?ux can be used to detect sources changing on timescales comparable to snap
exposure length (˘ 15 sec).
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
22
Table 1: DIASource Table
Name
Type
Unit
Description
fpSky
?oat
nmgy/asec
2
Estimated sky background
at the position (centroid) of
the object.
fpSkySigma
?oat
nmgy/asec
2
Estimated uncertainty of
fpSky .
E1
?oat
Adaptive e
1
shape measure
of the source as measured on
the di?erence image
43
.
E2
?oat
Adaptive e
2
shape measure.
E1E2cov
?oat[3]
E1 , E2 covariance matrix.
mSum
?oat
Sum of second adaptive mo-
ments.
mSumSigma
?oat
Uncertainty in mSum
extendedness
?oat
A measure of extendedness,
computed using a com-
bination of available mo-
ments and model ?uxes or
from a likelihood ratio of
point/trailed source mod-
els (exact algorithm TBD).
extendedness = 1 implies
a high degree of con?dence
that the source is extended.
extendedness = 0 implies
a high degree of con?dence
that the source is point-like.
?ags
bit[64]
bit
Flags
Some fast-moving, trailed, sources may be due to passages of nearby
asteroids. Their trails may exhibit signi?cant curvature. For trails longer
than a certain treshold (to be determined by start of Operations), we will ?t
43
See Bernstein & Jarvis (2002) for detailed discussion of all adaptive-moment related
http://ls.st/5f4
for a brief summary.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
23
a curved trail model which will be included with the alert (but likely stored
in a separate table).
4.3.2 DIAObject Table
Table 2: DIAObject Table
Name
Type
Unit
Description
diaObjectId
uint64
Unique identi?er
radec
double[2]
degrees
(?; ?) position of the object
at time radecTai
radecCov
?oat[3]
various
radec covariance matrix
radecTai
double
time
Time at which the object
was at a position radec .
pm
?oat[2]
mas/yr
Proper motion vector
44
parallax
?oat
mas
Parallax
pmParallaxCov
?oat[6]
various
Proper motion - parallax co-
variances.
pmParallaxLnL
?oat
Natural log of the likelihood
of the linear proper motion-
paralax ?t
45
.
pmParallaxChi2 ?oat
˜
2
statistic of the model ?t.
pmParallaxN
int
The number of data points
used to ?t the model.
psFlux
?oat[ugrizy] nmgy
Weighted
mean
point-
source model magnitude.
psFluxErr
?oat[ugrizy] nmgy
Standard error of psFlux
psFluxSigma
?oat[ugrizy] nmgy
Standard deviation of the
distribution of psFlux .
fpFlux
?oat[ugruzy] nmgy
Weighted mean forced pho-
tometry ?ux.
fpFluxErr
?oat[ugrizy] nmgy
Standard error of fpFlux
Continued on next page
44
High proper-motion or parallax objects will appear as \dipoles" in di?erence images.
Great care will have to be taken not to misidentify these as subtraction artifacts.
45
radec , pm , and parllax will all be simultaneously ?tted for. positionMotionLnL is
the logarithm of the likelihod of their maximum-likelihood estimates.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
24
Table 2: DIAObject Table
Name
Type
Unit
Description
fpFluxSigma
?oat[ugrizy] nmgy
Standard deviation of the
distribution of fpFlux .
lcPeriodic
?oat[6 ? 32]
Periodic features extracted
from light-curves using gen-
eralized Lomb-Scargle peri-
odogram (Table 4, Richards
et al. 2011)
46
lcNonPeriodic
?oat[6 ? 20]
Non-periodic features ex-
tracted from light-curves
(Table 5, Richards et al.
2011)
nearbyObj
uint64[3]
Closest Objects in Level 2
database.
nearbyObjDist
?oat[3]
arcsec
Distances to nearbyObj .
nearbyObjLnP
?oat[3]
Natural log of the prob-
ability that the observed
DIAObject is the same as
the nearby Object
47
.
?ags
bit[64]
bit
Flags
4.3.3 SSObject Table
Table 3: SSObject Table
Name
Type
Unit
Description
ssOb jectId
uint64
Unique identi?er
Continued on next page
46
The exact features in use when LSST begins operations are likely to be di?erent
compared to the baseline described here. This is to be expected given the rapid pace of
research in time domain astronomy. However, the number of computed features is unlikely
to grow beyond the present estimate.
47
This quantity will be computed by marginalizing over the product of position and
proper motion error ellipses of the Object and DIAObject , assuming a ?at prior.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
25
Table 3: SSObject Table
Name
Type
Unit
Description
oe
double[7]
various
Osculating orbital elements
at epoch (q, e, i, ?, !, M
0
,
epoch)
oeCov
double[21]
various
Covariance matrix for oe
arc
?oat
days
Arc of observation.
orbFitLnL
?oat
Natural log of the likelihood
of the orbital elements ?t.
orbFitChi2
?oat
˜
2
statistic of the orbital el-
ements ?t.
orbFitN
int
The number of data points
(observations) used to ?t
the orbital elements.
MOID
?oat[2]
AU
Minimum orbit intersection
distances
48
moidLon
double[2]
degrees
MOID longitudes.
H
?oat[6]
mag
Mean absolute magnitude,
per band (Muinonen et al.
2010 magnitude-phase sys-
tem).
G
1
?oat[6]
mag
G
1
slope parameter, per
band (Muinonen et al. 2010
magnitude-phase system).
G
2
?oat[6]
mag
G
2
slope parameter, per
band (Muinonen et al. 2010
magnitude-phase system).
hErr
?oat[6]
mag
Uncertainty of H estimate.
g1Err
?oat[6]
mag
Uncertainty of G
1
estimate.
g2Err
?oat[6]
mag
Uncertainty of G
2
estimate.
?ags
bit[64]
bit
Flags
The G
1
and G
2
parameters for the large majority of asteroids will not
be well constrained until later in the survey. We may decide not to ?t for
http://www2.lowell.edu/users/elgb/moid.html
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
26
it at all over the ?rst few DRs and add it later in Operations, or provide
two-parameter G
12
?ts. Alternatively, we may ?t it using strong priors on
slopes poorly constrained by the data. The design of the data management
system is insensitive to this decision, making it possible to postpone it to
Commissioning to ensure it follows the standard community practice at that
time.
The LSST database will provide functions to compute the phase (Sun-
Asteroid-Earth) angle ? for every observation, as well as the reduced, H(?),
and absolute, H, asteroid magnitudes in LSST bands.
4.3.4 Precovery Measurements
When a new DIASource is detected, it's useful to perform PSF photometry
at the location of the new source on images taken prior to discovery. These
are colloquially know as precovery measurements
49
. Performing precovery in
real time over all previously acquired visits is too I/O intensive to be feasible.
We therefore plan the following:
1. For all newly discovered objects, perform precovery PSF photometry
on visits taken over the previous 30 days
50
.
2. Make available a \precovery service" to request precovery for a limited
number of DIASources across all previous visits, and make it available
within 24 hours of the request. Web interface and machine-accessible
APIs will be provided.
The former should satisfy the most common use cases (eg., SNe), while
the latter will provide an opportunity for more extensive yet timely precovery
of targets of special interest.
4.3.5 Reprocessing the Level 1 Data Set
In what we've described so far, the Level 1 database is continually being
added to as new images are taken and DIASources identi?ed. Every time a
49
When Solar System objects are concerned, precovery has a slightly di?erent meaning:
predicting the positions of newly identi?ed SSObjects on previously acquired visits, and
associating with them the DIASources consistent with these predictions.
50
We will be maintaining a cache of 30 days of processed images to support this feature.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
27
new DIASource is associated to an existing DIAObject , the DIAObject record
is updated to incorporate new information brought in by the DIASource .
Once discovered and measured, the DIASources would never be re-discovered
and re-measured at the pixel level.
This would be far from optimal. The instrument will be better understood
with time. Newer versions of LSST pipelines will improve detection and
measurements on older data. Also, precovery photometry should optimally
be performed for all objects, and not just a select few. This argues for
periodic reprocessing of the Level 1 data set.
We plan to reprocess all image di?erencing-derived data (the Level 1
database), at the same time we perform the annual Level 2 data release
productions. This will include all images taken since the start of survey
operations, to the time when the data release production begins. The im-
ages will be reprocessed using a single version of the image di?erencing and
measurement software, resulting in a consistent data set.
As the reprocessing may take as long as ˘ 9 months, more imaging will
be acquired in the meantime. These data will be reprocessed as well, and
added to the new Level 1 database generated by the data release processing.
The reprocessed database will thus \catch up" with the Level 1 database
currently in use, possibly in a few increments. Once it does, the existing
Level 1 database will be replaced with the new one, and all future alerts will
refer to the reprocessed Level 1 database. Alerts for new sources \discovered"
during data release processing and/or the catch-up process will not be issued.
Note that Level 1 database reprocessing and switch will have signi?-
cant side-e?ects on downstream users. For example, all DIASource and
DIAObject IDs will change. Some DIASources and DIAObjects will dis-
appear (eg., if they're image subtraction artifacts that the improved soft-
ware was now able to recognize as such). New ones may appear. The
DIASource / DIAObject / Objects associations may change as well.
While the annual database switches will undoubtedly cause technical
inconvenience (eg., a DIASource detected at some position and associated
to one DIAObject ID on day T ? 1, will now be associated to a di?erent
DIAObject ID on day T + 0), the resulting database will be a more accurate
description of the astrophysics that the survey is seeing (eg., the associa-
tion on day T +0 is the correct one; the associations on T ? 1 and previous
days were actually made to an artifact that skewed the DIAObject lightcurve
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
28
characterization).
To ease the transition, third parties (event brokers) may choose to provide
positional-crossmatching to older versions of the Level 1 database. A set of
best practices will be developed to minimize the disruptions caused by the
switches (eg., when writing event-broker queries, ?lter on position, not on
DIAObject ID, if possible, etc.). A Level 1 database distribution service,
allowing for bulk downloads of the reprocessed Level 1 database, will exist
to support the brokers who will use it locally to perform more advanced
brokering
51
.
Older versions of the Level 1 database will be archived following the same
rules as for the Level 2 databases. The most recent DR, and the penultimate
data release will be kept on disk and loaded into the database. Others will be
archived to tape and made available as bulk downloads. See x 5.5 for more
detail.
4.4 Level 1 Image Products
4.4.1 Visit Images
Raw and processed visit images will be made available for download no later
than 24 hours from the end of visit acquisition.
The images will remain accessible with low-latency (seconds from request
to start of download) for at least 30 days, with slower access afterwards
(minutes to hours).
4.4.2 Di?erence Images
Complete di?erence images will be made available for download no later than
24 hours from the end of visit acquisition.
The images will remain accessible with low-latency (seconds from request
to start of download) for at least 30 days, with slower access afterwards
(minutes to hours).
4.4.3 Image Di?erencing Templates
Templates for di?erence image analysis will be created by coadding 6-months
to a year long groups of visits. The coaddition process will take care to remove
51
A \bulk-download" database distribution service will be provided for the Level 2
databases as well, to enable end-users to establish and run local mirrors (partial or full).
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
29
transient or fast moving objects (eg., asteroids) from the templates.
The input images may be further grouped by airmass and/or seeing
52
.
Therefore, at DR11, we will be creating 11 groups templates: two for the
?rst year of the survey (DR1 and DR2), and then one using imaging from
each subsequent year.
Di?erence image analysis will use the appropriate template given the time
of observation, airmass, and seeing.
4.5 Alerts to DIASources
4.5.1 Information Contained in Each Alert
For each detected DIASource , LSST will emit an \Event Alert" within 60
seconds of the end of visit (de?ned as the end of image readout from the
LSST Camera). These alerts will be issued in VOEvent format
53
, and should
be readable by VOEvent -compliant clients.
Each alert (a VOEvent packet) will at least include the following:
? alertID: An ID uniquely identifying this alert. It can also be used to
execute a query against the Level 1 database as it existed when this
alert was issued
? Level 1 database ID: For example, DR5L1
? Science Data:
{ The DIASource record that triggered the alert
{ The entire DIAObject (or SSObject ) record
{ All previous DIASource records
? Cut-out of the di?erence image centered on the DIASource (10 bytes/pixel,
FITS MEF)
? Cut-out of the template image centered on the DIASource (10 bytes/pixel,
FITS MEF)
52
The number and optimal parameters for airmass/seeing bins will be determined in
Commissioning.
53
Or some other format that is broadly accepted and used by the community at the
start of LSST commissioning.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
30
The cutouts will be sized so as to encompass the entire footprint of the
detected source, but be no smaller than 30 ? 30 pixels. The provided images
will comprise of a ?ux (32 bit ?oat), variance (32 bit ?oat), and mask (16
bit ?ags) planes, and include metadata necessary for further processing (e.g.,
WCS, zero point, PSF, etc.).
The items above are meant to represent the information transmitted with
each alert; the content of the alert packet itself will be formatted to con?rm
to VOEvent (or other relevant) standard. Where the existing standard is
inadequate for LSST needs, LSST will propose extensions and work with the
community to reach a common solution.
With each alert, we attempt to include as much information known to
LSST about the DIASource as possible, to minimize the need for follow-up
database queries. This speeds up classi?cation and decision making at the
user end, and relaxes the requirements on the database on the Project end.
4.5.2 Receiving and Filtering the Alerts
Alerts will be transmitted in VOEvent format, using standard IVOA protocols
(eg., VOEvent Transport Protocol; VTP
54
. As a very high rate of alerts is
expected, approaching ˘ 2 million per night, we plan for public VOEvent
Event Brokers
55
to be the primary end-points of LSST's event streams. End-
users will use these brokers to classify and ?lter events for subsets ?tting
their science goals. End-users will not be able to subscribe to full, un?ltered,
alert streams coming directly from LSST
56
.
To directly serve the end-users, LSST will provide a basic, limited ca-
pacity, alert ?ltering service. This service will run at the LSST U.S. Archive
Center (at NCSA). It will let astronomers create simple ?lters that limit what
alerts are ultimately forwarded to them
57
. These user de?ned ?lters will be
54
VOEvent Transport Protocol is currently an IVOA Note, but we understand work is
under way to ?nalize and bring it up to full IVOA Recommendation status.
55
These brokers are envisioned to be operated as a public service by third parties who
will have signed MOUs with LSST. An example may be the VAO (or its successor).
56
This is due to ?nite network bandwidth available: for example, a 100 end-users sub-
scribing to a ˘ 100 Mbps stream (the peak full stream data rate at end of the ?rst year of
operations) would require 10Gbps WAN connection from the archive center, just to serve
the alerts.
57
More speci?cally, to their VTP clients. Typically, a user will use the Science User
Interface (the web portal to LSST Archive Center) to set up the ?lters, and use their VTP
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
4 LEVEL 1 DATA PRODUCTS
31
possible to specify using an SQL-like declarative language, or short snippets
of (likely Python) code. For example, here's what a ?lter may look like:
# Keep only never-before-seen events within two
# effective radii of a galaxy. This is for illustration
# only; the exact methods/members/APIs may change.
def filter(alert):
if len(alert.sources) > 1:
return False
nn = alert.diaobject.nearest_neighbors[0]
if not nn.flags.GALAXY:
return False
return nn.dist < 2. * nn.Re
We emphasize that this LSST-provided capability will be limited, and is
not intended to satisfy the wide variety of use cases that a full-?edged public
Event Broker could. For example, we do not plan to provide any classi?-
cation (eg., \is the light curve consistent with an RR Lyra?", or \a Type
Ia SN?"). No information beyond what is contained in the VOEvent packet
will be available to user-de?ned ?lters (eg., no cross-matches to other cata-
logs). The complexity and run time of user de?ned ?lters will be limited by
available resources. Execution latency will not be guaranteed. The number
of VOEvents transmitted to each user per user will be limited as well (eg.,
at least up to ˘ 20 per visit per user, dynamically throttled depending on
load). Finally, the total number of simultaneous subscribers is likely to be
limited { in case of overwhelming interest, a TAC-like proposal process may
be instituted.
client to receive the ?ltered VOEvent stream.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
32
5 Level 2 Data Products
5.1 Overview
Level 2 data products result from direct image
58
analysis. They're designed to
enable static sky science (eg., studies of galaxy evolution, or weak lensing),
and time-domain science that is not time sensitive (eg. statistical investi-
gations of variability). They include image products (reduced single-epoch
exposures, called calibrated exposures, and coadds), and catalog products (ta-
bles of objects, sources, their measured properties, and related metadata).
Similarly to Level 1 catalogs of DIAObjects and DIASources , Objects
in the Level 2 catalog represent the astrophysical phenomena (stars, galax-
ies, quasars, etc.), while Sources represent their single-epoch observations.
Sources are independently detected and measured in single epoch exposures
and recorded in the Source table.
The master list of Objects in Level 2 will be generated by associating
and deblending the list of single-epoch source detections and the lists of
sources detected on coadds. We plan to build coadds designed to maximize
depth (\deep coadds") and coadds designed to achieve a good combination of
depth and seeing (\best seeing coadds"). We will also build a series of short-
period (eg. yearly, or multi-year) coadds. The ?ux limit in deep coadds will
be signi?cantly fainter than in individual visits, and the best seeing coadds
will help with deblending the detected sources. The short-period coadds are
necessary to avoid missing faint objects showing long-term variability. These
coadds will be built for all bands, as well as some combining multiple bands
(\multi-color coadds"). Not all of these will be preserved after sources
are detected and measured (see x 5.4.3 for details). We will provide a facility
to regenerate their subsections as Level 3 tasks (x 6).
The deblender will be run simultaneously on the catalog of peaks
59
de-
tected in the coadds, the DIAObject catalog from the Level 1 database, and
one or more external catalogs. It will use the knowledge of peak positions,
58
As opposed to di?erence image, for Level 1.
59
The source detection algorithm we plan to employ ?nds regions of connected pixels
above the nominal S=N threshold in the PSF-likelihood image of the visit (or coadd).
These regions are called footprints. Each footprint may have one or more peaks, and it
is these peaks that the deblender will use to infer the number and positions of objects
blended in each footprint.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
33
bands, time, time variability (from Level 1 and the single-epoch Source de-
tections), inferred motion, Galactic longitude and latitude, and other avail-
able information to produce a master list of deblended Objects . Metadata
on why and how a particular Object was deblended will be kept.
The properties of Objects , including their exact positions, motions, par-
allaxes, and shapes, will be characterized by MultiFit-type algorithms
60
.
Finally, to enable studies of variability, the ?uxes of all Objects will be
measured on individual epochs while keeping their shape parameters and
deblending resolutions constant. This process is known as forced photometry
(see x 5.2.4), and the ?ux measurements will be stored in the ForcedSource
table.
5.2 Level 2 Data Processing
Figure 1 presents a high-level overview of the Level 2 data processing work-
?ow
61
. Logically
62
, the processing begins with single-frame (visit) image
reduction and source measurement, followed by global astrometric and pho-
tometric calibration, coadd creation, detection on coadds, association and
deblending, object characterization, and forced photometry measurements.
The following is a high-level description of steps which will occur during
regular Level 2 data processing:
1. Single Frame Processing: Raw exposures are reduced to calibrated visit
exposures, and Sources are independently detected, deblended, and
measured on all visits. Their measurements (instrumental ?uxes and
shapes) are stored in the Source table.
2. Relative calibration: The survey is internally calibrated, both photo-
metrically and astrometrically. Relative zero point and astrometric
corrections are computed for every visit. Su?cient data is kept to re-
construct the normalized system response function ˚
b
(?) (see Eq. 5,
SRD) at every position in the focal plane at the time of each visit as
required by x 3.3.4 of the SRD.
60
\MultiFit algorithms" are those that ?t a PSF-convolved model to all multi-epoch
observations of an object. This is in contrast to measurement techniques where multi-
epoch images are coadded ?rst, and the properties are measured from the coadded pixels.
61
Note that some LSST documents refer to Data Release Processing, which includes
both Level 1 reprocessing (see x 4.3.5), and the Level 2 processing described here.
62
The actual implementation may parallelize these steps as much as possible.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
34
Associate and Deblend
Multi-Epoch Object
Characterization
Forced Photometry
Objects
ForcedSources
CoaddSources
Level 1
Catalog
External
Catalogs
Visits
Coadd Creation
Coadd Source
Detection and
Characterization
Coadds
Visits
Single Frame
Processing
Visits
Sources
Visits
Relative Calibration
(Photometric and Astrometric)
Figure 1: Level 2 Data Processing Overview
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
35
3. Coadd creation: Deep, seeing optimized, and short-period per-band
coadds are created in ugrizy bands, as well as deeper, multi-color,
coadds
63
. Transient sources (including Solar System objects, explosive
transients, etc), will be rejected from the coadds. See x 5.4.3 for details.
4. Coadd source detection and characterization. Sources will be detected
on all coadds generated in the previous step. The source detection al-
gorithm will detect regions of connected pixels, known as footprints,
above the nominal S=N threshold in the PSF-likelihood image of the
visit. Each footprint may have one or more peaks, and the collec-
tion of these peaks (and their membership in the footprints) are the
output of this stage. This information will be stored in a catalog of
CoaddSources
64
.
5. Association and deblending. The next stage in the pipeline, which
we will for simplicity just call the deblender, will synthesize a list of
unique objects. In doing so it will consider the catalogs of Sources and
CoaddSources , catalogs of DIASources , DIAObjects and SSObjects
detected on di?erence images, and objects from external catalogs.
The deblender will make use of all information available at this stage,
including the knowledge of peak positions, bands, time, time variability
(from Level 1), Galactic longitude and latitude, etc. The output of this
stage is a list of (uncharacterized) Objects
65
.
6. Multi-epoch object characterization. A set of prede?ned model ?ts and
measurements will be performed on each of the Objects identi?ed in
the previous step, taking all available multi-epoch data into account.
Model ?ts will be performed using MultiFit-type algorithms. Rather
than coadding a set of images and measuring object characteristics on
the coadd, MultiFit simultaneously ?ts PSF-convolved models to the
objects multiple observations. This reduces systematic errors, improves
the overall S=N, and allows for ?tting of time-dependent quantities
degenerate with shape on the coadds (for example, the proper motion).
63
We'll denote the \band" of the multi-color coadd as 'M'.
64
The exact contents of this catalog is implementation speci?c and will not be described
here.
65
Depending on the exact implementation of the deblender, this stage may also attach
signi?cant metadata (eg, deblended footprints and pixel-weight maps) to each deblended
Object record.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
36
The models we plan to ?t will not allow for ?ux variability (see the
next item).
7. Forced Photometry. Source ?uxes will be measured on every visit, with
the position, motion, shape, and the deblending parameters character-
ized in the previous step kept ?xed. This process of forced photometry,
will result in the characterization of the light-curve for each object in
the survey. The ?uxes will be stored in the ForcedSource table.
5.2.1 Object Characterization Measures
Properties of detected objects will be measured as a part of the object char-
acterization step described in the previous section and stored in the Object
table. These measurements are designed to enable LSST \static sky" science.
This section discusses at a high level which properties will be measured and
how those measurements will be performed. For a detailed list of quantities
being ?t/measured, see the table in x 5.3.1.
All measurements discussed in this section deal with properties of objects,
and will be performed on multi-epoch coadds, or by simultaneously ?tting
to all epochs. Measurements of sources in individual visits, independent of
all others, are described in x 5.2.3.
To enable science cases depending on observations of non-variable ob-
jects in the LSST-observed sky, we plan to measure the following using the
MultiFit approach:
? Point source model ?t. The observed object is modeled as a point
source with ?nite proper motion and parallax and constant ?ux (al-
lowed to be di?erent in each band). This model is a good description
for stars and other unresolved sources. Its 11 parameters will be simul-
taneously constrained using information from all available observations
in all bands
66
.
? Bulge-disk model ?t. The object is modeled as a sum of a de Vau-
couleurs (Sersic n = 4) and an exponential (Sersic n = 1) component.
This model is intended to be a reasonable description of galaxies
67
.
66
The ?tting procedure will account for di?erential chromatic refraction.
67
We may reconsider this choice if a better suited parametrization is discovered while
LSST is in Construction.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
37
The object is assumed not to move
68
. The components share the same
ellipticity and center. The model is independently ?t to each band.
There are a total of 8 free parameters, which will be simultaneously
constrained using information from all available epochs for each band.
Where there's insu?cient data to constrain the likelihood (eg., small,
poorly resolved, galaxies, or very few epochs), priors will be adopted
to limit the range of its sampling.
In addition to the maximum likelihood values of ?tted parameters and
their covariance matrix, a number (currently planned to be ˘ 200, on
average) of independent samples from the likelihood function will be
provided. These will enable use-cases sensitive to departures from the
Gaussian approximation, with shear measurement being the primary
use case. A permissible descope, in case of insu?cient storage, will be
not to sample the posteror for u and y bands.
? Standard colors. Colors of the object in \standard seeing" (for example,
the third quartile expected survey seeing in the i band, ˘ 0:8") will be
measured. These colors are guaranteed to be seeing-insensitive, suitable
for estimation of photometric redshifts
69
.
? Centroids. Centroids will be computed independently for each band
using an algorithm similar to that employed by SDSS. Information
from all
70
epochs will be used to derive the estimate. These centroids
will be used for adaptive moment, Petrosian, Kron, standard color, and
aperture measurements.
? Adaptive moments. Adaptive moments will be computed using infor-
mation from all epochs, independently for each band. The moments of
the PSF realized at the position of the object will be provided as well.
? Petrosian and Kron ?uxes. Petrosian and Kron radii and ?uxes will be
measured in standard seeing using self-similar elliptical apertures com-
puted from adaptive moments. The apertures will be PSF-corrected
68
I.e., have zero proper motion.
69
The problem of optimal determination of photometric redshift is the subject of in-
tense research. The approach we're taking here is conservative, following contemporary
practices. As new insights develop, we will revisit the issue.
70
Whenever we say all, it should be understood that this does not preclude reasonable
data quality cuts to exclude data that would otherwise degrade the measurement.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
38
and homogenized, convolved to a canonical circular PSF
71
. The radii
will be computed independently for each band. Fluxes will be com-
puted in each band, by integrating the light within some multiple of
the radius measured in the canonical band
72
(most likely the i band).
Radii enclosing 50% and 90% of light will be provided.
? Aperture surface brightness. Aperture surface brightness will be com-
puted in a variable number
73
of concentric, logarithmically spaced,
PSF-homogenized, elliptical apertures, in standard seeing.
? Variability characterization. Two groups of parameters will be pro-
vided, designed to characterize periodic and aperiodic variability fea-
tures (Richards et al. 2011). We caution that the exact features in use
when LSST begins operations are likely to be di?erent compared to the
baseline described here; this is to be expected given the rapid pace of
research in time domain astronomy. However, their number is unlikely
to grow beyond the present estimate.
5.2.2 Supporting Science Cases Requiring Full Posteriors
Science cases sensitive to systematics, departures of likelihood from Gaus-
sianity, or requiring user-speci?ed priors, demand knowledge of the shape of
the likelihood function beyond a simple Gaussian approximation around the
ML value. The estimate of bulge-disk model parameters and the estimate of
photometric redshifts are two examples where knowledge of the full posterior
is likely to be needed for LSST science cases.
We currently plan to provide this information in two ways: a) by pro-
viding independent samples from the likelihood function (in the case of the
71
This is an attempt to derive a de?nition of elliptical apertures that does not depend
on seeing. For example, for a large galaxy, the correction to standard seeing will introduce
little change to measured ellipticity. Corrected apertures for small galaxies will tend to be
circular (due to smearing by the PSF). In the intermediate regime, this method results in
derived apertures that are relatively seeing-independent. Note that this is only the case
for apertures; the measured ?ux will still be seeing dependent and it is up to the user to
take this into account.
72
The shape of the aperture in all bands will be set by the pro?le of the galaxy in the
canonical band alone. This procedure ensures that the color measured by comparing the
?ux in di?erent bands is measured through a consistent aperture. See
org/dr7/algorithms/photometry.html
for details.
73
The number will depend on the size of the source.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
39
bulge-disk model), and b) by providing parametric estimates of the likelihood
function (for the photometric redshifts). As will be shown in Table 4, the
current allocation is ˘ 200 samples (on average) for the bulge-disk model,
and ˘ 100 parameters for the photo-Z likelihood distributions, per object.
The methods of storing likelihood functions (or samples thereof) will con-
tinue to be developed and optimized throughout Construction and Commis-
sioning. The key limitation, on the amount of data needed to be stored, can
be overcome by compression techniques. For example, simply noticing that
not more than ˘ 0:5% accuracy is needed for sample values allows one to
increase the number of samples by a factor of 4. More advanced techniques,
such as PCA analysis of the likelihoods across the entire catalog, may al-
low us to store even more, providing a better estimate of the shape of the
likelihood function. In that sense, what is presented in Table 4 should be
thought of as a conservative estimate , which we plan to improve upon as
development continues in Construction.
5.2.3 Source Characterization
Sources will be detected on individual visits as well as the coadds. Sources
detected on coadds will primarily serve as inputs to the construction of the
master object list as described in x 5.2, and may support other LSST science
cases as seen ?t by the users (for example, searches for objects whose shapes
vary over time).
The following source properties are planned to be measured:
? Static point source model ?t. The source is modeled as a static point
source. There are a total of 3 free parameters (?, ?, ?ux). This model
is a good description of stars and other unresolved sources.
? Centroids. Centroids will be computed using an algorithm similar to
that employed by SDSS. These centroids will be used for adaptive mo-
ment and aperture magnitude measurements.
? Adaptive moments. Adaptive moments will be computed. The mo-
ments of the PSF realized at the position of the object will be provided
as well.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
40
? Aperture surface brightness. Aperture surface brightness will be com-
puted in a variable number
74
of concentric, logarithmically spaced,
PSF-homogenized, elliptical apertures.
Note that we do not plan to ?t extended source Bulge+Disk models to
individual Sources , nor measure per-visit Petrosian or Kron ?uxes. These
are object properties that are not expected to vary in time
75
, and will be
better characterized by MultiFit (in the Object table).
5.2.4 Forced Photometry
Forced Photometry is the measurement of ?ux in individual visits, given a
?xed position, shape, and the deblending parameters of an object. It enables
the study of time variability of an object's ?ux, irrespective of whether the
?ux in any given individual visit is above or below the single-visit detection
threshold.
Forced photometry will be performed on all visits, for all Objects . The
measured ?uxes will be stored in the ForcedSources table. Due to space
constraints, we only plan to measure the PSF ?ux.
5.2.5 Crowded Field Photometry
A fraction of LSST imaging will cover areas of high object (mostly stellar)
density. These include the Galactic plane, the Large and Small Magellanic
Clouds, and a number of globular clusters (among others).
LSST image processing and measurement software, although primarily
designed to operate in non-crowded regions, is expected to perform well in
areas of crowding. The current LSST applications development plan envi-
sions making the deblender aware of Galactic longitude and latitude, and
permitting it to use that information as a prior when deciding how to de-
blend objects. While not guaranteed to reach the accuracy or completeness
of purpose-built crowded ?eld photometry codes, we expect this approach
will yield acceptable results even in areas of moderately high crowding.
74
The number will depend on the size of the source.
75
Objects that do change shape with time would, obviously, be of particular interest.
Aperture ?uxes provided in the Source table should su?ce to detect these. Further per-
visit shape characterization can be performed as a Level 3 task.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
41
Note that this discussion only pertains to processing of direct images.
Crowding is not expected to signi?cantly impact the quality of data products
derived from di?erence images (i.e., Level 1).
5.3 The Level 2 Catalogs
This section presents the contents of key Level 2 catalog tables. As was
the case for Level 1 (see x 4.3), here we present the conceptual schemas for
the most important Level 2 tables (the Object , Source , and ForcedSource
tables).
These convey what data will be recorded in each table, rather than the
details of how. For example, columns whose type is an array (eg., radec ) may
be expanded to one table column per element of the array (eg., ra , decl ) once
this schema is translated to SQL. Secondly, the tables to be presented are
normalized (i.e., contain no redundant information). For example, since the
band of observation can be found by joining a Source table to the table with
exposure metadata, there's no column named band in the Source table. In
the as-built database, the views presented to the users will be appropriately
denormalized for ease of use.
5.3.1 The Object Table
Table 4: Level 2 Catalog Object Table
Name
Type
Unit
Description
ob jectId
uint64
Unique object identi?er
parentObjectId
uint64
ID of the parent Object this
object has been deblended
from, if any.
psRadecTai
double
time
Point source model: Time
at which the object was at
position radec .
psRadec
double[2]
degrees
Point source model: (?;?)
position of the object at
time radecTai .
psPm
?oat[2]
mas/yr
Point source model: Proper
motion vector.
Continued on next page
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
42
Table 4: Level 2 Catalog Object Table
Name
Type
Unit
Description
psParallax
?oat
mas
Point source model: Paral-
lax.
psFlux
?oat[ugrizy] nmgy
Point source model ?uxes
76
.
psCov
?oat[66]
various
Point-source model covari-
ance matrix
77
.
psLnL
?oat
Natural log likelihood of
the observed data given the
point source model.
psChi2
?oat
˜
2
statistic of the model ?t.
psN
int
The number of data points
(pixels) used to ?t the
model.
bdRadec
double[2][ugrizy]
B+D model
78
: (?;?) posi-
tion of the object at time
radecTai , in each band.
bdEllip
?oat[2][ugrizy]
B+D model:
Ellipticity
(e
1
; e
2
) of the object.
bdFluxB
?oat[ugrizy] nmgy
B+D model:
Integrated
?ux of the de Vaucouleurs
component.
bdFluxD
?oat[ugrizy] nmgy
B+D model:
Integrated
?ux of the exponential com-
ponent.
bdReB
?oat[ugrizy] arcsec
B+D model: E?ective ra-
dius of the de Vaucouleurs
pro?le component.
Continued on next page
76
Point source model assumes that ?uxes are constant in each band. If the object is
variable, psFlux will e?ectively be some estimate of the average ?ux.
77
Not all elements of the covariance matrix need to be stored with same precision. While
the variances will be stored as 32 bit ?oats (˘ seven signi?cant digits), the covariances
may be stored to ˘ three signi?cant digits (˘ 1% ).
78
Though we refer to this model as \Bulge plus Disk", we caution the reader that the
decomposition, while physically motivated, should not be taken too literally.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
43
Table 4: Level 2 Catalog Object Table
Name
Type
Unit
Description
bdReD
?oat[ugrizy] arcsec
B+D model: E?ective ra-
dius of the exponential pro-
?le component.
bdCov
?oat[36][ugrizy]
B+D model covariance ma-
trix
79
.
bdLnL
?oat[ugrizy]
Natural log likelihood of
the observed data given the
bulge+disk model.
bdChi2
?oat[ugrizy]
˜
2
statistic of the model ?t.
bdN
int[ugrizy]
The number of data points
(pixels) used to ?t the
model.
bdSamples
?oat16[9][200][ugrizy]
Independent samples of
bulge+disk likelihood sur-
face. All sampled quantities
will be stored with at least
˘ 3 signi?cant digits of
precision.
The number
of samples will vary from
object to object, depending
on how well the object's
likelihood function is ap-
proximated by a Gaussian.
stdColor
?oat[5]
mag
Color of the object mea-
sured in \standard seeing".
While the exact algorithm
is yet to be determined,
this color is guaranteed to
be seeing-independent and
suitable for photo-Z deter-
minations.
stdColorSigma
?oat[5]
mag
Uncertainty of stdColor .
Continued on next page
79
See psCov for notes on precision of variances/covariances.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
44
Table 4: Level 2 Catalog Object Table
Name
Type
Unit
Description
radec
double[6][2] arcsec
Position of the object (cen-
troid), computed indepen-
dently in each band. The
centroid will be computed
using an algorithm similar
to that employed by SDSS.
radecSigma
double[6][2] arcsec
Uncertainty of radec .
E1
?oat[ugrizy]
Adaptive e
1
shape measure.
See Bernstein & Jarvis
(2002) for detailed discus-
sion of all adaptive-moment
related quantities
80
.
E2
?oat[ugrizy]
Adaptive e
2
shape measure.
E1E2cov
?oat[ugrizy][3]
E1 , E2 covariance matrix.
mSum
?oat[ugrizy]
Sum of second adaptive mo-
ments.
mSumSigma
?oat[ugrizy]
Uncertainty in mSum
m4
?oat[ugrizy]
Fourth order adaptive mo-
ment.
petroRad
?oat[ugrizy] arcsec
Petrosian radius, computed
using elliptical apertures de-
?ned by the adaptive mo-
ments.
petroRadSigma
?oat[ugrizy] arcsec
Uncertainty of petroRad
petroBand
int8
The band of the canonical
petroRad
petroFlux
?oat[ugrizy] nmgy
Petrosian ?ux within a de-
?ned multiple of the canon-
ical petroRad
petroFluxSigma ?oat[ugrizy] nmgy
Uncertainty in petroFlux
petroRad50
?oat[ugrizy] arcsec
Radius containing 50% of
Petrosian ?ux.
petroRad50Sigma ?oat[ugrizy] arcsec
Uncertainty of petroRad50 .
Continued on next page
80
http://ls.st/5f4
for a brief summary.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
45
Table 4: Level 2 Catalog Object Table
Name
Type
Unit
Description
petroRad90
?oat[ugrizy] arcsec
Radius containing 90% of
Petrosian ?ux.
petroRad90Sigma ?oat[ugrizy] arcsec
Uncertainty of petroRad90 .
kronRad
?oat[ugrizy] arcsec
Kron radius (computed us-
ing elliptical apertures de-
?ned by the adaptive mo-
ments)
kronRadSigma
?oat[ugrizy] arcsec
Uncertainty of kronRad
kronBand
int8
The band of the canonical
kronRad
kronFlux
?oat[ugrizy] nmgy
Kron ?ux within a de?ned
multiple of the canonical
kronRad
kronFluxSigma
?oat[ugrizy] nmgy
Uncertainty in kronFlux
kronRad50
?oat[ugrizy] arcsec
Radius containing 50% of
Kron ?ux.
kronRad50Sigma ?oat[ugrizy] arcsec
Uncertainty of kronRad50 .
kronRad90
?oat[ugrizy] arcsec
Radius containing 90% of
Kron ?ux.
kronRad90Sigma ?oat[ugrizy] arcsec
Uncertainty of kronRad90 .
apN
int8
Number of elliptical annuli
(see below).
apMeanSb
?oat[6][ apN ] nmgy/as
2
Mean surface brightness
within an annulus
81
.
apMeanSbSigma ?oat[6][ apN ] nmgy/as
2
Standard
deviation
of
apMeanSb .
Continued on next page
81
A database function will be provided to compute the area of each annulus, to enable
the computation of aperture ?ux.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
46
Table 4: Level 2 Catalog Object Table
Name
Type
Unit
Description
extendedness
?oat
A measure of extendedness,
computed using a com-
bination of available mo-
ments and model ?uxes or
from a likelihood ratio of
point/trailed source mod-
els (exact algorithm TBD).
extendedness = 1 implies
a high degree of con?dence
that the source is extended.
extendedness = 0 implies
a high degree of con?dence
that the source is point-like.
lcPeriodic
?oat[6 ? 32]
Periodic features extracted
from light-curves using gen-
eralized Lomb-Scargle peri-
odogram (Table 4, Richards
et al. 2011).
lcNonPeriodic
?oat[6 ? 20]
Non-periodic features ex-
tracted from light-curves
(Table 5, Richards et al.
2011).
photoZ
?oat[2 ? 100]
Photometric redshift likeli-
hood samples { pairs of (z,
logL) { computed using a
to-be-determined published
and widely accepted algo-
rithm at the time of LSST
Commissioning.
?ags
bit[128]
bit
Flags
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
47
5.3.2 Source Table
Source measurements are performed independently on individual visits. They're
designed to enable astrometric and photometric relative calibration, variabil-
ity studies of high signal-to-noise objects, and studies of high SNR objects
that vary in position and/or shape (eg., comets).
Table 5: Level 2 Catalog Source Table
Name
Type
Unit
Description
sourceId
uint64
Unique source identi?er
82
ccdVisitId
uint64
ID of CCD and visit where
this source was measured
ob jectId
uint64
ID of the Object this source
was associated with, if any.
ssOb jectId
uint64
ID of the SSObject this
source has been linked to, if
any.
parentSourceId
uint64
ID of the parent Source this
source has been deblended
from, if any.
sky
?oat
nmgy/asec
2
Estimated sky background
at the position (centroid) of
the source.
skySigma
?oat
nmgy/asec
2
Estimated uncertainty of
sky .
psFlux
?oat
nmgy
Calibrated point source
model ?ux.
psXY
?oat[2]
pixels
Point
source
model:
(column; row)
position
of the object on the CCD.
psCov
?oat[6]
various
Point-source model covari-
ance matrix
83
.
Continued on next page
82
It would be optimal if the source ID is globally unique across all releases. Whether
that's realized will depend on technological and space constraints.
83
Not all elements of the covariance matrix will be stored with same precision. While
the variances will be stored as 32 bit ?oats (˘ seven signi?cant digits), the covariances
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
48
Table 5: Level 2 Catalog Source Table
Name
Type
Unit
Description
psLnL
?oat
Natural log likelihood of
the observed data given the
point source model.
psChi2
?oat
˜
2
statistic of the model ?t.
psN
int
The number of data points
(pixels) used to ?t the
model.
psRadec
double[2]
degrees
Point source model: (?;?)
position of the object,
transformed from psXY
psCov2
?oat[6]
various
Point-source model covari-
ance matrix for psRadec
and psFlux .
xy
?oat[2]
arcsec
Position of the object (cen-
troid), computed using an
algorithm similar to that
used by SDSS.
xyCov
?oat[3]
Covariance matrix for xy .
radec
double[2]
arcsec
Calibrated (?, ?) of the
source, transformed from
xy .
radecCov
?oat[3]
arcsec
Covariance
matrix
for
radec .
E1
?oat
Adaptive e
1
shape measure.
E2
?oat
Adaptive e
2
shape measure.
E1E2cov
?oat[3]
E1 , E2 covariance matrix.
mSum
?oat
Sum of second adaptive mo-
ments.
mSumSigma
?oat
Uncertainty in mSum
m4
?oat
Fourth order adaptive mo-
ment.
apN
int8
Number of elliptical annuli
(see below).
Continued on next page
may be stored to ˘ three signi?cant digits (˘ 1% ).
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
49
Table 5: Level 2 Catalog Source Table
Name
Type
Unit
Description
apMeanSb
?oat[ apN ]
nmgy
Mean surface brightness
within an annulus.
apMeanSbSigma ?oat[ apN ]
nmgy
Standard
deviation
of
apMeanSb .
?ags
bit[64]
bit
Flags
5.3.3 ForcedSource Table
Table 6: Level 2 Catalog ForcedSource Table
Name
Type
Unit
Description
ob jectId
uint64
Unique object identi?er
ccdVisitId
uint64
ID of CCD and visit where
this source was measured
psFlux
?oat
nmgy
Point source model ?ux.
psFluxErr
?oat
nmgy
Point source model ?ux er-
ror, stored to 1% precision.
?ags
bit[8]
bit
Flags
5.4 Level 2 Image Products
5.4.1 Visit Images
Raw exposures, including individual snaps, and processed visit images will be
made available for download as FITS ?les. They will be downloadable both
through a human-friendly Science User Interface, as well as using machine-
friendly APIs.
Required calibration data, processing metadata, and all necessary image
processing software will be provided to enable the user to generate bitwise
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
50
identical processed images from raw images
84
.
5.4.2 Calibration Data
All calibration frames (darks, ?ats, biases, fringe, etc.) will be preserved and
made available for download as FITS ?les.
All auxiliary telescope data, both raw (images with spectra) and pro-
cessed (calibrated spectra, derived atmosphere models), will be preserved
and made available for download.
5.4.3 Coadded Images
In course of Level 2 processing, multiple classes and numerous of coadds will
be created:
? A set of deep coadds. One deep coadd will be created for each of the
ugrizy bands, plus a seventh, deeper, multi-color coadd. These coadds
will be optimized for a reasonable combination of depth (i.e., employ no
PSF matching) and resolution (i.e., visits with signi?cantly degraded
seeing may be omitted). Transient sources (including Solar System
objects, explosive transients, etc), will be removed. Care will be taken
to preserve the astrophysical backgrounds
85
.
These coadds will be kept inde?nitely and made available to the users.
Their primary purpose is to enable the end-users to apply alternative
object characterization algorithms, perform studies of di?use structures,
and for visualization.
? A set of short-period coadds. These will comprise of multiple (ugrizyM)
sets of yearly and multi-year coadds. Each of these sets will be created
using only a subset of the data, and otherwise share the characteristics
of the deep coadds described above. These are designed to enable detec-
tion of long-term variable or moving
86
objects that would be \washed
out" (or rejected) in full-depth coadds. We do not plan to keep and
make these coadds available . We will retain and provide su?cient
metadata for users to re-create them using Level 3 or other resources.
84
Assuming identically performing software and hardware con?guration.
85
For example, using \background matching" techniques;
http://ls.st/l9u
86
For example, nearby high proper motion stars.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
51
? A set of best seeing coadds. One deep coadd will be created for each of
the ugrizy bands, using only the best seeing data (for example, using
only the ?rst quartile of the realized seeing distribution). These will be
built to assist the deblending process. We do not plan to keep and
make these coadds available . We will retain and provide su?cient
metadata for users to re-create them using Level 3 or other resources.
? One (ugrizyM) set of PSF-matched coadds. These will be used to
measure colors and shapes of objects at \standard" seeing. We do
not plan to keep and make these coadds available . We will
retain and provide su?cient metadata for users to re-create them using
Level 3 or other resources.
The exact details of which coadds to build, and which ones to keep, can
change during Construction without a?ecting the processing system design
as the most expensive operations (raw image input and warping) are constant
in the number of coadds produced. The data management system design is
sensitive to the total number and size of coadds to be kept { these are the
relevant constraining variables.
To build the coadds, we currently plan to subdivide the sky into 12 over-
lapping
87
tracts, spanning approximately 75 ? 72 degrees. The sky will be
stereographically projected onto the tracts
88
, and pixelized into (logical) im-
ages 2.0 x 1.9 megapixels in size (3.8 terapixels in all). Physically, these
large images will be subdivided into smaller (e.g. 2k ? 2k), non-overlapping,
patches, though that will be transparent to the users. The users will be able
to request arbitrarily chosen regions
89
in each tract, and receive them back
as a FITS ?le.
We reiterate that not all coadds will be kept and served to the
public
90
, though su?cient metadata will be provided to users to recreate
them on their own. Some coadds may be entirely \virtual": for example,
87
We're planning for 3.5 degrees of overlap, roughly accommodating a full LSST focal
plane.
88
https://dev.lsstcorp.org/trac/wiki/DM/SAT/SkyMap
for details.
89
Up to some reasonable upper size limit; i.e., we don't plan to expect to support creation
of 3.8 Tpix FITS ?les.
90
The coadds are a major cost driver for storage. LSST Data Management system is
currently sized to keep and serve seven coadds, ugrizyM, over the full footprint of the
survey.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
52
the PSF-matched coadds could be implemented as ad-hoc convolutions of
postage stamps when the colors are measured.
We will retain smaller sections of all generated coadds, to support quality
assessment and targeted science. Retained sections may be positioned to
cover areas of the sky of special interest such as overlaps with other surveys,
nearby galaxies, large clusters, etc.
5.5 Data Release Availability and Retention Policies
Over 10 years of operations, LSST will produce eleven data releases: two for
the ?rst year of survey operations, and one every subsequent year. Each data
release will include reprocessing of all data from the start of the survey, up
to the cuto? date for that release.
The contents of data releases are expected to range from a few PB (DR1)
to ˘ 70 PB for DR11 (this includes the raw images, retained coadds, and
catalogs). Given that scale, it is not feasible to keep all data releases loaded
and accessible at all times.
Instead, only the contents of the most recent data release, and
the penultimate data release will be kept on fast storage and with
catalogs loaded into the database . Statistics collected by prior surveys
(eg., SDSS) show that users nearly always prefer accessing the most recent
data release, but sometimes may use the penultimate one (this is especially
true just after the publication of a new data release). Older releases are used
rarely.
To assist with data quality monitoring and assessment small, overlap-
ping, samples of data from older releases will be kept loaded in
the database . The sample size is expected to be on order of ˘ 1? 5% of
the data release data, with larger samples kept early on in the survey. The
goal is to allow one to test how the reported characterization of the same
data varies from release to release.
Older releases will be archived to mass storage (tape). The users will
not be able to perform database queries against archived releases .
They will be made available as bulk downloads in some common format (for
example, FITS binary tables). Database software and data loading scripts
will be provided for users who wish to set up a running copy of an older (or
current) data release database on their systems.
All raw data used to generate any public data product (raw exposures,
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
5 LEVEL 2 DATA PRODUCTS
53
calibration frames, telemetry, con?guration metadata, etc.) will be kept and
made available for download.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
6 LEVEL 3 DATA PRODUCTS AND CAPABILITIES
54
6 Level 3 Data Products and Capabilities
Level 3 capabilities are envisioned to enable science cases that would greatly
bene?t from co-location of user processing and/or data within the LSST
Archive Center. The high-level requirement for Level 3 is established in x 3.5
of the LSST SRD.
Level 3 capabilities include three separate deliverables:
1. Level 3 Data Products and associated storage resources
2. Level 3 processing resources, and
3. Level 3 programming environment and framework
Many scientists' work may involve using two or all three of them in concert,
but they can each be used independently. We describe each one of them in
the subsections to follow.
6.1 Level 3 Data Products and Associated Storage Re-
sources
These are data products that are generated by users on any computing re-
sources anywhere that are then brought to an LSST Data Access Center
(DAC) and stored there. The hardware for these capabilities includes the
physical storage and database server resources at the DAC to support them.
For catalog data products, there is an expectation that they can be "fed-
erated" with the Level 1 (L1) and Level 2 (L2) catalogs to enable analyses
combining them. Essentially this means that either the user-supplied tables
include keys from the L1/L2 catalogs that can be used for key-equality-
based joins with them (example: a table of custom photometric redshifts for
galaxies, with a column of object IDs that can be joined to the L2 Object
catalog), or that there are columns that can be used for spatial (or tempo-
ral, or analogous) joins against L1/L2 tables. The latter implies that such
L3 table's columns must be in the same coordinate system and units as the
corresponding L1/L2 columns.
There is no requirement that Level 3 data products (L3DPs) are derived
from L1 or L2 other than that they be joinable with them. For instance,
a user might have a catalog of radio sources that they might want to bring
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
6 LEVEL 3 DATA PRODUCTS AND CAPABILITIES
55
into federation with the LSST catalogs. That can be thought of as a Level 3
Data Product as long as they have \LSST-ized" it by ensuring compatibility
of coordinate, time, measurement systems, etc. Nevertheless, we do expect
the majority of L3DPs to be derived from processed LSST data.
There could also be L3 image data products; for example, user-generated
coadds with special selection criteria or stacking algorithms.
Any L3DP may have access controls associated with it, restricting read
access to just the owner, to a list of people, to a named group of people, or
allowing open access.
The storage resources for L3DPs come out of the SRD requirement for
10% of LSST data management capabilities to be devoted to user processing.
In general, they are likely to be controlled by some form of a \space allo-
cation committee". Users will probably have some small baseline automatic
allocation, beyond which a SAC proposal is needed. The SAC may take into
account scienti?c merit, length of time for which the storage is requested,
and openness of the data to others, in setting its priorities.
It is to be decided whether users will be required to provide the code
and/or documentation behind their L3DPs, or whether the SAC may include
the availability of this supporting information in its prioritization. Obviously
if a user intends to make a L3DP public or publish it to a group it will be
more important that supporting information be available.
Level 3 data products that are found to be generally useful can be mi-
grated to Level 2. This is a fairly complex process that ultimately involves
the project taking responsibility for supporting and running LSST-style code
that implements the algorithm necessary to produce the data product (it's
not just relabeling an existing L3DP as L2). The project will provide neces-
sary support for such migrations.
6.2 Level 3 Processing Resources
These are project-owned computing resources located at the DACs. They are
available for allocation to all users with LSST data rights. They may be used
for any computation that involves the LSST data and advances LSST-related
science. The distinctive feature of these computing resources is that they are
located with excellent I/O connections to the image and catalog datasets at
Level 1 and Level 2. There may be other co-located but not project-owned,
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
6 LEVEL 3 DATA PRODUCTS AND CAPABILITIES
56
resources available at the LSST DACs
91
; their use is beyond the scope of this
document, except to note that reasonable provisions will be made to ensure
it is possible to use them to process large quantities of LSST data.
Level 3 processing resources will, at least, include systems that can carry
out traditional batch-style processing, probably similarly con?gured to those
LSST will be using for the bulk of data release production processing. It is
to be determined whether any other ?avors of hardware would be provided,
such as large-memory machines; this is likely to be driven by the project
need (or lack thereof) for such resources.
There will be a time allocation committee (TAC) for these resources.
Every LSST-data-rights user may get a small default allocation (enough to
run test jobs). Substantial allocations will require a scienti?c justi?cation.
Priorities will be based on the science case and, perhaps, also on whether the
results of the processing will be released to a larger audience. Requests must
specify what special ?avors of computing will be needed (e.g., GPUs, large
memory, etc.).
A fairly standard job control environment (like Condor), will be available,
and users will be permitted to work with it at a low, generic level. They
will not be required to use the higher levels of the LSST process control
middleware (but they may; see x 6.3).
These processing resources can be available for use in any clearly LSST-
related scienti?c work. It is not strictly required that they be used to process
LSST data, in this context. For instance, it could be acceptable to run
special types of cosmological simulations that are in direct support of an
LSST analysis, if the closeness to the data makes the LSST facility uniquely
suitable for such work. The TAC will take into account in its decisions
whether proposed work makes good use of the enhanced I/O bandwidth
available to LSST data on these systems.
6.3 Level 3 Programming Environment and Frame-
work
As a part of the Level 3 Programming Environment and Framework, the
LSST will make available the LSST software stack to users, to aid in the
analyses of LSST data. This includes all code implementing the core process-
91
For example, the U.S. DAC will be located at the National Petascale Facility building
at NCSA, adjacent to the Blue Waters supercomputer.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
6 LEVEL 3 DATA PRODUCTS AND CAPABILITIES
57
ing algorithms (image characterization and manipulation, building of coadds,
image di?erencing, object detection, object characterization, moving object
detection, etc.), the middleware necessary to run those codes at large scale,
as well as the LSST database management system.
These analyses could be done on LSST-owned systems (i.e., on the Level
3 processing resources) but also on a variety of supported external systems.
We will aim to support common personal Unix ?avors (for example, common
distributions of Linux and Mac OS X) as well as commonly used cluster and
HPC environments. The vision is to enable relatively straightforward use
of major national systems such as XSEDE or Open Science Grid, as well
as some common commercial cloud environments. The decision of which
environments to support will be under con?guration control and we will seek
advice from the user community. We cannot commit to too many ?avors. In-
kind contributions of customizations for other environments will be welcome
and may provide a role for national labs.
The Level 3 environment is intended, when put to fullest use, to allow
users to run their own productions-like runs on bulk image and/or catalog
data, with mechanisms for creating and tracking large groups of jobs in a
batch system.
The Level 3 environment, in asymptopia, has a great deal in common
with the environment that the Project will use to build the Level 2 data
releases. It is distinct, however, as supporting it as a tool meant for the
end-users imposes additional requirements:
? In order to be successful as a user computing environment, it needs to
be easy to use. Experience with prior project
92
has shown that if the
production computing environment is not envisioned from the start as
being shared with users, it will likely evolve into an experts-only tool
that is too complicated, or too work-hardened, to serve users well.
? While it is desirable for the production computing to be portable to
Grid, cloud, etc. resources, this option might not be exercised in prac-
tice and could atrophy. For the user community, it's a far more central
capability. Early community engagement is therefore key to developing
and maintaining these capabilities.
? Not all the capabilities of the LSST production environment need nec-
essarily be exported to the users. LSST-speci?c capabilities associated
92
For example, BaBar.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
6 LEVEL 3 DATA PRODUCTS AND CAPABILITIES
58
with system administration, for instance, are not of interest to end-
users.
6.4 Migration of Level 3 data products to Level 2
? For the migration to be considered, the creator of the L3DP will need
to agree to make their data product public to the entire LSST data-
rights community, along with supporting documentation and code. The
code at ?rst need not be in the LSST framework or even in an LSST-
supported language.
? If the original proponent wrote her/his code in the C++/Python LSST
stack environment (the "Level 3 environment"), it will be easier to mi-
grate it to Level 2 (though, obviously, using the same languages/frameworks
does not guarantee that the code is of production quality).
? If the original code was written in another language or another data
processing framework, the project will have the resources to rewrite it
to required LSST standards.
? Taking on a new Level 2 DP means that the project is committing to
code maintenance, data quality review, space allocation, and continuing
production of the new L2DP through DR11.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
7 DATA PRODUCTS FOR SPECIAL PROGRAMS
59
7 Data Products for Special Programs
LSST Survey Speci?cations (LSST SRD, x 3.4) specify that 90% of LSST
observing time will be spend executing the so-called \universal cadence".
These observations will result in Level 1 and 2 data products described earlier
in this document.
The remaining 10% of observing time will be devoted to special programs,
obtaining improved coverage of interesting regions of observational parameter
space. Examples include very deep (r ˘ 26, per exposure) observations,
observations with very short revisit times (˘1 minute), and observations of
\special" regions such as the Ecliptic, Galactic plane, and the Large and
Small Magellanic Clouds. A third type of survey, micro-surveys, that would
use about 1% of the time, may also be considered.
The details of these special programs or micro surveys are not yet de-
?ned
93
. Consequently, the speci?cs of their data products are left unde?ned
at this time. Instead, we just specify the constraints on these data products,
given the adopted Level 1/2/3 architecture. It is understood that no special
program will be selected that does not ?t these constraints
94
. This allows
us to size and construct the data management system, without knowing the
exact de?nition of these programs this far in advance.
Processing for special programs will make use of the same software stack
and computing capabilities as the processing for universal cadence. The
programs are expected to use no more than ˘10% of computational and
storage capacity of the LSST data processing cluster. When special products
include time domain event alerts, their processing shall generally be subject
to the same latency requirements as Level 1 data products.
For simplicity of use and consistency, the data products for special pro-
grams will be stored in databases separate from the \main" (Level 1 and 2)
databases. The system will, however, allow for simple federation with Level
1/2/3 data products (i.e., cross-queries and joins).
As a concrete example, a data product complement for a \deep drilling"
?eld designed for supernova discovery and characterization may consist of: i)
alerts to events discovered by di?erencing the science images against a special
deep drilling template, ii) a Level 1-like database iii) one or more \nightly
93
The initial complement is expected to be de?ned and selected no later than Science
Veri?cation.
94
Or will come with additional, external, funding, capabilities, and/or expertise.
LSST Data Products Definition Document
LSE-163
Latest Revision 10/7/2013
The contents of this document are subject to configuration control and may not be changed, altered, or their provisions waived without prior
approval of the LSST Change Control Board.
7 DATA PRODUCTS FOR SPECIAL PROGRAMS
60
co-adds" (co-adds built using the data from the entire night), produced and
made available within ˘ 24 hours, and iv) special deep templates, built us-
ing the best recently acquired seeing data, produced on a fortnightly cadence.
Note that the data rights and access rules apply just as they would for
for Level 1/2/3. For example, while generated event alerts (if any) will be
accessible world-wide, the image and catalog products will be restricted to
users with LSST data rights.