Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

A publishing partnership

Tails: Chasing Comets with the Zwicky Transient Facility and Deep Learning

, , , , , , , , , , , , , , and

Published 2021 April 8 © 2021. The American Astronomical Society. All rights reserved.
, , Citation Dmitry A. Duev et al 2021 AJ 161 218 DOI 10.3847/1538-3881/abea7b

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

1538-3881/161/5/218

Abstract

We present Tails, an open-source deep-learning framework for the identification and localization of comets in the image data of the Zwicky Transient Facility (ZTF), a robotic optical time-domain survey currently in operation at the Palomar Observatory in California, USA. Tails employs a custom EfficientDet-based architecture and is capable of finding comets in single images in near real time, rather than requiring multiple epochs as with traditional methods. The system achieves state-of-the-art performance with 99% recall, a 0.01% false-positive rate, and a 1–2 pixel rms error in the predicted position. We report the initial results of the Tails efficiency evaluation in a production setting on the data of the ZTF Twilight survey, including the first AI-assisted discovery of a comet (C/2020 T2) and the recovery of a comet (P/2016 J3 = P/2021 A3).

Export citation and abstract BibTeX RIS

1. Introduction

Comets have mesmerized humans for millennia, frequently offering, arguably, some of the most spectacular sights in the night sky. Containing the original materials from when the solar system first formed, comets provide a unique insight into the distant past of our solar system. The recent discovery of the first interstellar comet 2I/Borisov by amateur astronomer Gennadiy Borisov predictably sparked much excitement and enthusiasm among astronomers and the general public alike (e.g., Bolin et al. 2020; Fitzsimmons et al. 2019; Guzik et al. 2020). Such objects could potentially provide important information on the formation of other stellar systems. It is a very exciting time to look for comets: the large-scale time-domain surveys that are currently in operation, such as the Zwicky Transient Facility (ZTF; Bellm et al. 2019a; Graham et al. 2019), Panoramic Survey Telescope & Rapid Response System (Pan-STARRS; Chambers et al. 2016), or Asteroid Terrestrial-impact Last Alert System (ATLAS; Tonry et al. 2018), and the upcoming ones such as BlackGEM (Bloemen et al. 2016) and Vera Rubin Observatory/Large Synoptic Survey Telescope (LSST; Ivezić et al. 2008) offer the richest data sets ever available to mine for comets.

Traditional comet detection algorithms rely on multiple observations of cometary objects that are linked together and used to fit an orbital solution. To the best of our knowledge, the previous attempts to take the comet's morphology in the optical image data into consideration in the detection algorithms have not led to reliable and robust results.

In this work, we present Tails—a state-of-the-art open-source deep-learning-based system for the identification and localization of comets in the image data of ZTF. 11 Tails employs an EfficientDet-based architecture (Tan et al. 2019) and is thus capable of finding comets in single images in near real time, rather than requiring multiple epochs as with traditional methods.

1.1. The Zwicky Transient Facility

The ZTF 12 is a state-of-the-art robotic time-domain sky survey capable of visiting the entire visible sky north of −30° decl. every night. ZTF observes the sky in the g, r, and i bands at different cadences depending on the scientific program and sky region (Bellm et al. 2019a; Graham et al. 2019). The 576 megapixel camera with a 47 deg2 field of view, installed on the Samuel Oschin 48 inch (1.2 m) Schmidt Telescope, can scan more than 3750 deg2 per hour, to a 5σ detection limit of 20.7 mag in the r band with a 30 s exposure during a new moon (Masci et al. 2019; Dekany et al. 2020).

The ZTF partnership has been running a specialized survey, the Twilight Survey (ZTF-TS) that operates at solar elongations down to 35 degrees with an r-band limiting magnitude of 19.5 (Ye et al. 2020; Bellm et al. 2019b). ZTF-TS has so far resulted in the discovery of a number of Atira asteroids (orbits interior to Earth's) as well as the first inner-Venus object, 2020 AV2 (Ip et al. 2020). Motivated by the success, ZTF-TS will be expanded in Phase II of the project, which commenced in 2020 December.

Comets become more easily detectable when close to the Sun as they become brighter and start exhibiting more pronounced coma and tails. Furthermore, it has been shown that the most detectable direction of approach of an interstellar object is from directly behind the Sun because of observational selection effects (Jedicke et al. 2016) and the fact that this direction has a greater cross section for asteroids to bend around and pass into the visibility volume (Engelhardt et al. 2017; Do et al. 2018).

Tails automates the search for comets with detectable morphology. While trained and evaluated on a large corpus of ZTF data, in this work we focus on Tails' performance when applied to the ZTF-TS data.

2. Tails: A Deep-learning Framework for the Identification and Localization of Comets

Deep learning (DL) is a subset of machine learning that employs artificial multi-layer neural networks (McCulloch & Pitts 1943). DL systems are able to discover, in a highly automated manner, efficient representations of the data, simplifying the task of finding the meaningful sought-after patterns in them. We refer the reader to a brilliant introduction into DL given in Géron (2019).

DL systems often reach near-optimal performance for a given task and are able to learn even very complicated, highly nonlinear mappings between the input and output spaces. The art of building applied DL systems involves two major challenges: finding a suitable network architecture and, more importantly, constructing a large, labeled representative data set for the network training. In the case of comet detection, the training set must reflect the possible variations across different seeing conditions, filters, sky location, CCDs, and include data artifacts caused by, for example, cross-talk or telescope reflections.

2.1. Data Set

To build a seed sample for labeling, we first identified all potential observations of known comets conducted with ZTF from 2018 March 5 to 2020 March 4, based on their predicted position and brightness. The code for accomplishing that is based on the Python libraries pypride (Duev et al. 2016) and solarsyslib (Jensen-Clem et al. 2018) and uses the comet ephemerides obtained from the Minor Planet Center (MPC) 13 for a coarse search, followed by a JPL Horizons 14 (Giorgini et al. 1996) query for precision.

To provide more contextual information, epochal image data are supplemented by properly aligned reference images of the corresponding patches of sky and difference (epochal minus reference) images generated with the ZOGY algorithm (Zackay et al. 2016), all produced by the ZTF Science Data System at Caltech's IPAC (Masci et al. 2019). Finally, we generate image triplet cutouts 256 by 256 pixel in size, which in angular measure translates into 4farcm3 by 4farcm3 at ZTF's pixel scale of 1farcs01/pixel.

We selected over 60,000 individual observations with the total comet magnitude ranging from 10 to 23 (as reported by JPL Horizons; see Figure 1), out of which about 20,000 were sourced for manual annotation. This resulted in an initial sample of 3000 examples with identifiable morphology.

Figure 1.

Figure 1. Distribution of over 60,000 individual observations of comets as a function of the predicted total magnitude (as reported by JPL Horizons) used in the seed sample.

Standard image High-resolution image

We also compiled a set of approximately 20,000 negative examples consisting of point-like cometary detections, patches of sky with no identified transient or variable sources, CCD-edge cases, and a wide range of real (point-source) transient and bogus (e.g., artifacts due to bright stars, optical ghosts, and "dementors") samples from the Braai data set (Duev et al. 2019).

To expand the data set, we then assembled a standard ResNet-based (He et al. 2015) classifier for comet identification. With this basic classification model, we ran several rounds of an active-learning-like procedure, where we would first train the classifier, evaluate it on the whole data set, sample both confident predictions and the cases close to the classifier's decision boundary, manually inspect and label those examples, and add them to the training set. Roughly 2000 positive and 2000 negative examples were added to the training set via this method.

The resulting training data set contains about 5000 positive and 22,000 negative examples (see Figure 2). Each triplet in the set has been assigned a label [pc , x, y], where pc marks the presence of a comet in the image and x, y ∈ [0, 1] is the relative positions of the comet's center of mass, as reported by JPL Horizons. For positive examples this translates into [1, xJPL, yJPL], and for negative ones [0, ?, ?], where the question marks denote that these do not affect the loss in this case.

Figure 2.

Figure 2. Comet examples from the training set. The first image in a triplet is the epochal science image, the second is the reference image, and the third is the ZOGY difference image.

Standard image High-resolution image

2.2. Deep Neural Network Architecture and Training

Tails adopts a custom architecture (see Figure 3) based on EfficientDet D0 (Tan et al. 2019), a variant of a state-of-the-art architecture designed for object detection—a computer-vision technique for the identification and location of objects in image data.

Figure 3.

Figure 3. Tails architecture: a custom EfficientDet D0-based network (Tan et al. 2019). A batch of duplet or triplet image stacks of size (nb , 256, 256, [2∣3]), correspondingly, is passed through an EfficientNet B0 backbone, where nb is the number of stacks in the batch. The extracted features from the last five blocks/levels of the backbone network are passed through a bidirectional feature pyramid network (BiFPN). The resulting five output tensors denoted in colored circles are fed into the head network, which outputs the probability of the image containing a comet pc and its predicted relative position (x, y).

Standard image High-resolution image

This architecture delivers best-in-class object detection efficiency and performance across a wide range of resource constraints. This is achieved by using EfficientNet—state-of-the art backbone networks for feature extraction, a weighted bidirectional feature pyramid network (BiFPN), which allows easy and fast multiscale feature fusion, and a compound scaling method that simultaneously and uniformly scales the resolution, depth, and width for all backbone, feature, and location/class prediction networks (Tan et al. 2019).

The use of a BiFPN, which effectively represents and processes multiscale features, makes this architecture particularly well suited for the problem of morphology-based comet identification and localization.

A batch of triplet image stacks (nb , 256, 256, 3) in size, where nb is the number of stacks in the batch, is passed through an EfficientNet B0 backbone (Tan & Le 2019). The extracted features from the last five blocks/levels of the network are passed through the BiFPN. The resulting five output tensors denoted by the colored circles in Figure 3 are fed into the head network, which outputs the probability of the image containing a comet pc and its centroid's predicted relative (x, y) position. 15

We defined the loss function as

Equation (1)

where Lc denotes the binary cross-entropy function for the label c (1—there is a comet in the image, 0—there is no comet) and the predicted probability pc . If ⌊pc ⌉ = 1, Lp is computed as an L1 loss for the relative position (x, y) and its prediction (xp , yp ) with a small L2 regularizing term (with epsilon = 10−3), and wc and wp denote the weights of the two terms, respectively,

Equation (2)

We employed the Adam optimizer (Kingma & Ba 2014), a batch size of 32, and a 81%/9%/10% training/validation/test data split. For data augmentation, we applied random horizontal and vertical flips of the input data; no random rotations and translations were added. We note that the test/validation sets did not contain augmented data from the training set. We used standard techniques to maximize training performance: if no improvement in validation loss was observed for 10 epochs, the learning rate was reduced by a factor of 2, and training was stopped early if no improvement was observed for 30 epochs.

The EfficientNet's weights were randomly initialized. 16 We first set wc = 10, wp = 1 to allow for a fast convergence of the feature-extracting part of the network. To fine-tune the performance, we trained Tails on a balanced data set setting wc = 1.1, wp = 1 and monitored the validation loss for early stopping, then bumped wp = 2 and monitored the validation positional rms error, and finally, we added the omitted negative examples and again monitored the validation loss for early stopping.

The resulting classifiers were put through the same active-learning-like procedure as was employed in the initial data set assembly, using several months of ZTF-TS data.

3. Tails Performance

Evaluated on the test set, with a score pc threshold of 0.5, Tails demonstrates false-positive (FPR) and false-negative rates (FNR) of 1.7%, and a ∼1–2 pixel median rms error of the predicted comet centroid position versus that acquired from JPL Horizon (see Figure 4).

Figure 4.

Figure 4. Test set performance of Tails. The set contains 1400 negative and 650 positive examples.

Standard image High-resolution image

The ZTF instrument's CCD mosaic has 16 individual 6k × 6k science CCDs. The raw ZTF image data are split into four readout quadrants per CCD and all processing is conducted independently on each CCD readout quadrant. We tessellate each 3k × 3k CCD-quadrant image into a 13 × 13 grid of overlapping 256 × 256 pixel tiles and evaluate Tails on those. 17

Tails has been deployed in production since late 2020 June. We have implemented a sentinel service 18 that processes the incoming data in real time and posts the plausible candidates to Fritz 19 , the ZTF Phase II open-source science data platform (van der Walt et al. 2019; Duev et al. 2019; Kasliwal et al. 2019), for further manual inspection and vetting. The candidates are auto-annotated with the detailed information on the detection such as the score, CCD, and sky positions, and cross-matches with known solar system objects. Figure 5 shows screenshots of the Fritz user interfaces used in the process.

Figure 5.

Figure 5. Screenshots of the Fritz user interfaces used for Tails candidate inspection and vetting.

Standard image High-resolution image

It takes about 5 hr to run inference on a typical set of nightly ZTF-TS data (∼45 30 s exposures) on an e2-highcpu-32 virtual machine instance (32 vCPU, 32 GB memory, SSD disk) on the Google Cloud Platform, including I/O operations.

Consistently with the expected rate of comet observations, a typical run on nightly Twilight data yields a few dozen candidates, which, given the typical number of processed tiles, gives an empirical FPR value of about 0.01%.

The scanning results are accumulated and used to expand the training set and improve Tails' performance.

We have evaluated Tails' performance on a random sample of 200 observations of known comets with identifiable morphology in 2020 July–August and found an empirical recall value of 99%.

Figure 6 shows a number of comet candidates not from the training set identified by Tails, including some of the ZTF observations of the comet 2I/Borisov. Optical artifacts resembling cometary objects are the main source of contamination.

Figure 6.

Figure 6. Candidates identified by Tails. Panels (a), (b), and (c) on the left show the detections of real comets. Typical false positives are shown on panels (d), (e), and (f) on the right. For each image triplet, the left pane shows the epochal science exposure, the middle pane shows the reference image of the corresponding patch of sky, and the right pane shows the ZOGY difference image.

Standard image High-resolution image

3.1. Discovery of Comet C/2020 T2

On 2020 October 7, Tails discovered a candidate that was posted to MPC's Possible Comet Confirmation Page (PCCP) 20 as ZTFDD01 (see Figure 7). It was later confirmed to be a long-period comet and designated C/2020 T2 (Palomar), marking the first DL-assisted comet discovery (Duev et al. 2020). The candidate was found in the Twilight survey data; it was at 19.3 mag in the ZTF r band. The FWHM of the object was approximately 2farcs5–3'', compared to nearby background stars that have an FWHM of ∼2''. The object showed a tail extending up to 5'' in the westward direction. Table 1 summarizes the orbital elements of C/2020 T2 provided by the MPC and Figure 8 shows its orbit as of the discovery date.

Figure 7.

Figure 7. Discovery image of C/2020 T2 (Palomar), the first DL-assisted comet discovery by Tails. Taken on 2020 October 7 with the ZTF camera on the 48 inch Schmidt telescope at Palomar. The left panel shows the epochal science exposure (256 × 256 pixel cutout), the middle panel shows the reference image, and the right panel shows the ZOGY difference image. East is to the left, north is down.

Standard image High-resolution image
Figure 8.

Figure 8. The orbit of comet C/2020 T2 as of 2020 October 7. Image credit: NASA/JPL-Caltech / D. Duev.

Standard image High-resolution image

Table 1. Orbital Elements of C/2020 T2 Provided by the MPC

ElementValue
e 0.9934213
Incl.27.87307
Peri.150.38279
Node83.04834
q 2.0546940
T 2021 July 11.14638 TT

Download table as:  ASCIITypeset image

To determine if Tails could have discovered C/2020 T2 before 2020 October 7, we searched the ZTF archive for all Twilight Survey data covering the ephemeris position of the comet with the ZChecker software (Kelley et al. 2019). Eleven nights of data were found between 2020 June 11–20 (evening twilight) and October 7–21 (morning twilight). The comet was in conjunction with the Sun between the two sets, and not observable by ZTF. We measured the brightness of the coma in 4 pixel radius apertures, and aperture corrected the photometry according to the ZTF pipeline documentation. The data are shown in Figure 9. Typical seeing was 2'' in June, and the comet was very faint (r = 20.2 mag), near the single-image detection limit (r = 20.4–20.9 mag, 5σ point source), and had no morphological features for Tails to pick up. Thus 2020 October 7 was really the first opportunity for Tails to discover the comet.

Figure 9.

Figure 9. Photometry of comet C/2020 T2 (Palomar) derived from ZTF-TS images (r band) vs. time from perihelion. A best-fit model lightcurve is also shown: $r=9.85+9.54{\mathrm{log}}_{10}({r}_{{\rm{h}}})+5{\mathrm{log}}_{10}({\rm{\Delta }})-{\rm{\Phi }}(\theta )$, where rh is the heliocentric distance in astronomical units, Δ is the comet-observer distance in astronomical units, and Φ(θ)) is the phase angle correction from Schleicher et al. (1998). Tp denotes the time of perihelion passage (2021 July 11).

Standard image High-resolution image

3.2. Recovery of Comet P/2016 J3 = P/2021 A3 (STEREO)

A comet candidate was identified by a combination of Tails and the ZTF Moving Object Detection Engine (Masci et al. 2019) on 2020 January 4 UTC and was submitted to the PCCP as ZTF0Ion (see Figure 10). It was later identified as a recovery of comet P/2016 J3 (STEREO) and given the designation P/2016 J3 = P/2021 A3 (STEREO; Bolin 2021). P/2021 A3 was identified in the evening Twilight survey data at r 19.3 mag with a clearly extended appearance scoring 0.9 with a coma ∼10'' wide and a tail extending past 20'' in the northeast direction.

Figure 10.

Figure 10. Recovery detection of P/2021 A3 (STEREO) identified by Tails in the ZTF r-band data taken on 2021 January 4. The left panel shows the epochal science exposure (256 × 256 pixel cutout), the middle panel shows the reference image, and the right panel shows the ZOGY difference image. East is to the left, north is down.

Standard image High-resolution image

4. Discussion

This work demonstrates the potential of the state-of-the-art deep-learning computer-vision architecture designs when applied to the problem of astronomical source detection and localization, with a specific focus on comets.

We experimented with the input data and trained a version of Tails that instead of triplet image stacks uses duplets—epochal/reference images, omitting the ZOGY difference images. Our tests show that this version achieves essentially the same performance as the one trained on triplets without requiring image differencing, expanding the range of potential use cases of Tails.

While Tails is trained only on ZTF data, with transfer learning it can be adapted to other sky surveys, including the upcoming Vera Rubin Observatory's LSST (Ivezić et al. 2008).

D.A. Duev would like to thank Ivan Duev for assistance with data labeling. D.A. Duev acknowledges support from Google Cloud and from the Heising-Simons Foundation under grant No. 12540303.

Based on observations obtained with the Samuel Oschin Telescope 48 inch and the 60 inch Telescope at the Palomar Observatory as part of the Zwicky Transient Facility project. ZTF is supported by the National Science Foundation under grant No. AST-1440341 and a collaboration including Caltech, IPAC, the Weizmann Institute for Science, the Oskar Klein Center at Stockholm University, the University of Maryland, the University of Washington, Deutsches Elektronen-Synchrotron and Humboldt University, Los Alamos National Laboratories, the TANGO Consortium of Taiwan, the University of Wisconsin at Milwaukee, and Lawrence Berkeley National Laboratories. Operations are conducted by COO, IPAC, and UW.

This research has made use of data and/or services provided by the International Astronomical Union's Minor Planet Center.

The authors would like to express gratitude to the anonymous referee.

Facilities: PO:1.2 m - , ZTF - .

Software: astropy (Astropy Collaboration et al. 2018), Fritz (https://github.com/fritz-marshal/fritz), Kowalski (Duev et al. 2019), matplotlib (Hunter 2007), numpy (Harris et al. 2020), pandas (Pandas Development Team 2020), pypride (Duev et al. 2016), SEP (Barbary 2016), sbpy (Mommert et al. 2019), TensorFlow (Abadi et al. 2016), ZChecker (Kelley et al. 2019).

Footnotes

Please wait… references are loading.
10.3847/1538-3881/abea7b