This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy.

The following article is Open access

ExoMiner: A Highly Accurate and Explainable Deep Learning Classifier That Validates 301 New Exoplanets

, , , , , , , , , , , , and

Published 2022 February 17 © 2022. The Author(s). Published by the American Astronomical Society.
, , Citation Hamed Valizadegan et al 2022 ApJ 926 120 DOI 10.3847/1538-4357/ac4399

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

0004-637X/926/2/120

Abstract

The Kepler and Transiting Exoplanet Survey Satellite (TESS) missions have generated over 100,000 potential transit signals that must be processed in order to create a catalog of planet candidates. During the past few years, there has been a growing interest in using machine learning to analyze these data in search of new exoplanets. Different from the existing machine learning works, ExoMiner, the proposed deep learning classifier in this work, mimics how domain experts examine diagnostic tests to vet a transit signal. ExoMiner is a highly accurate, explainable, and robust classifier that (1) allows us to validate 301 new exoplanets from the MAST Kepler Archive and (2) is general enough to be applied across missions such as the ongoing TESS mission. We perform an extensive experimental study to verify that ExoMiner is more reliable and accurate than the existing transit signal classifiers in terms of different classification and ranking metrics. For example, for a fixed precision value of 99%, ExoMiner retrieves 93.6% of all exoplanets in the test set (i.e., recall = 0.936), while this rate is 76.3% for the best existing classifier. Furthermore, the modular design of ExoMiner favors its explainability. We introduce a simple explainability framework that provides experts with feedback on why ExoMiner classifies a transit signal into a specific class label (e.g., planet candidate or not planet candidate).

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 4.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

Recent space missions have revolutionized the study of exoplanets in astronomy by observing and comprehensively analyzing hundreds of thousands of stars in search of transiting planets. The CoRoT mission (Baglin et al. 2006) observed 26 stellar fields of view between 2007 and 2012, detecting more than 30 exoplanets and more than 550 candidates (Deleuil et al.2018). Kepler (Borucki et al. 2010) continuously observed 112,046 stars 11 in a 115 deg2 region of the sky for 4 yr and identified over 4000 planet candidates (PCs; Thompson et al. 2018), among which there are more than 2300 confirmed or statistically validated exoplanets. 12 K2, using the repurposed Kepler spacecraft (Howell et al. 2014), detected more than 2300 candidates with over 400 confirmed or validated exoplanets. More recently, the Transiting Exoplanet Survey Satellite (TESS) mission (Ricker et al. 2015) surveyed ∼75% of the entire sky in its first 2 yr of flight starting in 2018 June, an area 300 times larger than that monitored by Kepler, and detected 2241 candidates and ∼130 confirmed exoplanets (Guerrero et al. 2021). All of these large survey missions have produced many more candidate exoplanets than can easily be confirmed or identified as false positives (FPs) using conventional approaches.

The most common contemporary 13 approach to detecting and vetting exoplanet candidates involves a three-step process: (1) The imaging data from a transit survey are processed on complex data processing and transit search pipelines 14 (see, e.g., Jenkins et al. 2010b, 2016), which identify transit-like signals (called threshold-crossing events—TCEs), perform an initial limb-darkened transit model fit for each signal, and conduct a suite of diagnostic tests to help distinguish between exoplanet signatures and nonexoplanet phenomena (such as background eclipsing binaries; Twicken et al. 2018; Li et al. 2019). For the Kepler and TESS science pipelines, the results of the transit search, model fitting, and diagnostic tests are presented in data validation (DV) reports (Twicken et al. 2018; Li et al. 2019) that include one-page summary reports (Figure 1) along with more comprehensive reports. (2) The TCEs are filtered by either predefined if-then vetting rules (Thompson et al. 2018; Coughlin 2017) or machine learning (ML) classifiers (AstroNet; Shallue & Vanderburg 2018; Yu et al. 2019) to identify those most likely to be exoplanets. (3) The DV reports for top-tier TCEs that survive the filtering process in step 2 are then typically reviewed by vetting teams and released as Kepler or TESS Objects of Interest (KOI/TOI) for follow-up observations (Thompson et al. 2016a; Guerrero et al. 2021).

Figure 1.

Figure 1. Example DV one-page summary report. It includes multiple diagnostic plots and variables: (0) stellar parameters, (1) full time series flux, (2) full-orbit phase-folded flux, (3) transit-view phase-folded secondary eclipsing flux, (4) phase-folded transit view of the flux, (5) phase-folded transit view of the whitened flux, (6) phase-folded transit view of the odd & even flux, (7) difference image (out-of-transit) centroid offsets, and (8) DV analysis table of variables.

Standard image High-resolution image

ML methods are ideally suited for probing these massive data sets, relieving experts from the time-consuming task of sifting through the data and interpreting each DV report, or comparable diagnostic material, manually. When utilized properly, ML methods also allow us to train models that potentially reduce the inevitable biases of experts. Among many different ML techniques, deep neural networks (DNNs) have achieved state-of-the-art performance (LeCun et al. 2015) in areas such as computer vision, speech recognition, and text analysis and, in some cases, have even exceeded human performance. DNNs are especially powerful and effective in these domains because of their ability to automatically extract features that may be previously unknown or highly unlikely for human experts in the field to grasp (Bengio et al. 2012). This flexibility allows for the development of ML models that can accept raw data and thereby increase the productiveness of the transit search and vetting process.

The availability of many Kepler and TESS pipeline products has provided researchers with a wide spectrum of input choices for machine-based transit signal classification, from the unprocessed raw pixels to the middle-level pipeline products (e.g., processed flux data and centroid data) and final tabular data sets of diagnostic test values. Several transit signal classifiers have been previously developed and have each been designed to target inputs at different stages along this data processing pipeline.

Robovetter (Coughlin et al. 2016) and Autovetter (Jenkins et al. 2014; McCauliff et al. 2015) both utilize the final values generated via diagnostic statistics as inputs but differ from each other in their respective model designs: Robovetter relies heavily on domain knowledge to create an expert-type classifier system, while Autovetter exploits a fully automated ML approach to build a random forest classifier. The main drawback of these classifiers is their high level of dependency on the pipeline and its final products, i.e., the features that are engineered and built from the diagnostic tests and transit fit.

More recent works use the flux data directly, including Ansdell et al. (2018), Shallue & Vanderburg (2018), Dattilo et al. (2019), and Osborn et al. (2020), which train DNN models using flux and/or centroid time series, and Armstrong et al. (2021), which combined diagnostics scalar values and flux data to train more accurate non-DNN models. Such models not only learn how to automatically classify transit signals but are also capable of extracting important features that allow the machine to classify these signals (unlike human-extracted features, which may not be suitable or sufficient for use by a machine). While the final Kepler data release DR25 applied a rule-based system, Robovetter (Coughlin et al. 2016; Thompson et al. 2018), to generate the final PC catalog, ML models are gaining trust to be employed in practice to help identify PCs. For example, AstroNet is used to assist in the identification of TOIs for MIT's Quick-Look Pipeline, which operates on TESS Full Frame Image data (Guerrero et al. 2021), and it was also used to validate two new exoplanets (Shallue & Vanderburg 2018). More recently, Armstrong et al. (2021) utilized ML methods to systematically validate 50 new exoplanets. Thus far, the primary reasons for the remaining hesitancy about utilizing fully data-driven DNN approaches for vetting and validating transit signals are the models' lack of explainability and acceptable accuracy.

The inefficacy of existing DNN classifiers used to vet transit signals has signaled researchers that the flux data alone may not be enough as inputs for these models, demonstrating the need to potentially use the minimally processed raw pixel data. Although we agree that the flux data alone are insufficient, we argue that using raw pixels as inputs may prove to be impractical owing to the magnitude of the Kepler data set (about 68,000 images × 32 pixels/image on average ≈ 2.2 million pixel values for each transit signal for the Kepler mission) 15 and the lack of enough labeled data. Learning an effective classifier from data in such a high-volume space requires exponentially more data than are available (curse of dimensionality; Bishop 2006). For the final Kepler Data Release (Q1–Q17 DR25), for example, there are only approximately 34,000 labeled TCEs available (and approximately 200,000 transit signals if non-TCEs are included).

To understand the difficulty of developing a model based on the raw pixel data, note that a complex stand-alone DNN intended to classify raw pixels must learn all of the steps in the raw pixel processing pipeline, which includes extracting photometry data from the pixel data, removing systematic errors, extracting other useful features from pixel and flux data, and generating final classifications from the ground up with a limited amount of supervision in the form of labeled data. Considering that there is a whole pipeline dedicated to processing these data, learning each of these complex steps in a data-driven manner is a difficult task. Domain-specific techniques such as specialized augmentation to generate more labeled data or self-supervised learning (Dosovitskiy et al. 2014) could be employed to partially mitigate the curse of dimensionality; however, we believe that more accurate DNN classifiers can be developed if the abundant amount of domain knowledge utilized in the design of the pipeline is used as a guide for designing the DNN architecture. Another advantage of this approach is that it helps to build DNN models that are more explainable, as we will show in this paper. By examining DV reports and how they are used by experts to vet transit signals, we introduce a new DNN model that can be used to build catalogs of PCs and mine new exoplanets. We show that our model is highly accurate in classifying transit signals of the Kepler mission and can be transferred to classify TESS signals.

This paper is organized as follows. We start Section 2 by discussing the manual (Section 2.1) and machine (Section 2.2) classification of transit signals and finish it with Section 2.3, which discusses exoplanet validation, and Section 2.4, which summarizes the contributions/novelties of this work. In order to justify the choices made in the design of our ML approach, we first provide a brief background of machine classification in Section 3.1, DNNs in Section 3.2, and explainability in Section 3.3. This helps us to provide the details of the proposed DNN, ExoMiner, in Section 3.4, relevant data preparation in Section 3.5, and optimizing the model in Section 3.6. The second half of the paper is focused on experimental studies. After discussing the Kepler data set used in this paper, baseline classifiers, and evaluation metrics in Section 4, we study the performance of ExoMiner in Section 5. Section 6 is focused on reliability and stability of ExoMiner, and Section 7 introduces a framework for explainability of ExoMiner. We prepare the ground for validating new exoplanets using ExoMiner in Section 8 and validate over 301 new exoplanets in Section 9. We report a preliminary study of applying ExoMiner on TESS data in Section 10. Finally, Section 11 concludes the paper by listing the caveats of this work and plotting the future directions.

2. Related Work

In this section, we discuss the manual and automatic classification of transit signals. Because the design of the proposed model in this work is greatly influenced by the manual vetting process, we first review how experts vet transit signals. We then review the related works on automatic classification and validation of new exoplanets. We conclude this section by enumerating our contributions.

2.1. Manual Classification of TCEs

The Kepler/TESS pipeline might detect FP transit events due to sources such as eclipsing binaries (EBs), background eclipsing binaries (BEBs), planets transiting background stars, stellar variability, and instrument-induced artifacts (Borucki et al. 2011). Manual classification of TCEs consists of the time-consuming evaluation of multiple diagnostic tests and fit values to determine whether a TCE is a PC or FP. The FP category may be further subdivided into astrophysical FP (AFP) and nontransiting phenomena (NTP).

Several tests have been developed to diagnose different types of FPs. Combined with the flux data in multiple forms, these diagnostic tests are the main components of the one-page DV summary and the longer DV report used by experts to vet transit signals (Twicken et al. 2018, 2020). The one-page DV summary report includes (1) stellar parameters regarding the target star to help determine the type of star associated with the TCE (Figure 1, 1), (2) unfolded flux data to understand the overall reliability of the transit signal (Figure 1, 2), (3) the phase-folded full-orbit and transit view of the flux data to determine the existence of a secondary eclipse and shape of the signal (Figure 1, 3), (4) the phase-folded transit view of the secondary eclipse to check the reliability of the eclipsing secondary (Figure 1, 4), (5) the phase-folded whitened transit view to check how well a whitened transit model is fitted to the whitened light curve (Figure 1, 5), (6) the phase-folded odd and even transit view to check for the FP due to a circular EB target or a BEB when the detected period is half of the period of EB (Figure 1, 6), (7) the difference image to check for the FP due to a background object being the source of the transit signature (Figure 1, 7), and (8) an analysis table that provides values related to the model fit and various DV diagnostic parameters to better understand the signal (Figure 1, 8).

The information in both the odd & even and the weak secondary tests is used in detecting EB FPs. The odd & even depth test is useful when the pipeline detects only one TCE for a circular binary system; in this case, the depth of the odd and even views could be different. The weak secondary test is used when the pipeline detects a TCE for the primary eclipses and a TCE for the secondary in an eccentric EB system, or when there is a significant secondary that does not trigger a TCE. However, secondary events could also be caused by secondary eclipses of a giant, short-period transiting planet exhibiting reflected light or thermal emission. To distinguish between these two types of secondary events, experts use the secondary geometric albedo, planet effective temperature, and multiple event statistic (MES) of the secondary event (Twicken et al. 2018; Jenkins 2020).

Moreover, note that a real exoplanet could be confused with an EB if the pipeline incorrectly measures the period for the corresponding TCE. Depending on whether the pipeline's returned period for an exoplanet is twice 16 or half 17 the real period, the weak secondary or odd & even tests might get flagged. Examples of exoplanets with incorrectly returned periods that resulted in statistically significant weak secondary or odd & even tests are Kepler-1604 b and Kepler-458 b, respectively, as shown in Figures 2(a) and (b). Such an incorrect period can lead models to erroneously classify the corresponding exoplanet as an FP. Therefore, it is difficult for the machine to correctly classify a TCE as a PC without the correct orbital period (available for these cases in the Cumulative Kepler Object of Interest (KOI) catalog; Thompson et al. 2018). 18 Furthermore, circumbinary transiting planets and planets with significant transit timing variations are difficult to classify correctly, since the pipeline was not designed to detect their aperiodic signatures.

Figure 2.

Figure 2. An incorrect period can lead to the misclassification of a TCE. In the top panel, TCE KIC 8416523.1 was detected at twice the actual period of the transiting planet, resulting in half the transits being considered secondary events. In the bottom panel a TCE KIC 9663113.2 was detected at half the true period, resulting in no transit events for the even-numbered transit phases.

Standard image High-resolution image

After examining the DV summary, experts may also have to check the full DV report, which provides a more thorough analysis, in order to classify a TCE. An initial PC label obtained using the manual vetting process should then be subsequently submitted to a follow-up study by the rest of the community (using ground-based telescope observations, for example) in order to confirm the classification.

The manual vetting process has several flaws: not only is it very slow and a bottleneck to efficiently identifying new planets, but it is also not always consistent or effective because of the biases that inevitably come with human labeling. Moreover, in order to make the manual process feasible and reduce the number of transit signals to examine, the pipeline rejects many signals that do not meet certain conditions, including signals below a defined signal-to-noise ratio (S/N). Importantly, some of the most interesting transits of small long-period planets may be just below the threshold but could be successfully vetted with the inclusion of information in the image data that is complementary to the information in the extracted light curve. We risk missing these planets because they are discarded by the pipeline before all the available information in the observed data has been considered. Unlike the manual vetting process, automatic classification is fast and can be used to rapidly vet transit signals, potentially reducing the biases and inconsistencies that come with human vetting. Furthermore, automatic classification can be used to vet all transits, including low-S/N signals near the threshold, in order to identify more small PCs.

2.2. Existing Methods for Automatic Classification of TCEs

Some of the well-known existing methods for automatic classification of transit signals are summarized in Table 1. Even though they all can be considered machine classifiers, they differ in their design objective, inputs, and construction. Because these different models aim to address different problems, they should be compared with their objective in mind. Nonetheless, given that all these models are able to generate dispositions and disposition scores, they can be studied from the vetting and validation perspective. These approaches can be categorized into three groups:

  • 1.  
    The expert system category includes only Robovetter (Coughlin 2017) and a ported version of Robovetter for the TESS mission called TEC (Guerrero et al. 2021). Robovetter was designed to build a KOI catalog by leveraging manually incorporated domain knowledge in the form of if-then conditions. These if-then rules check different types of diagnostic and transit fit values. The developer of such an expert system must recognize special cases and completely understand the decision-making process of an expert in order to develop effective systems. As such, these systems are not fully automated data-driven models. They cannot also easily accept complex data inputs such as time series or images (the nonscalar inputs in Table 1).
  • 2.  
    The generative category includes models such as vespa (Morton & Johnson 2011; Morton 2012; Morton et al. 2016) that require the class priors p(y = 1) and p(y = 0) and likelihoods P(Xy = 1) and P(Xy = 0) to estimate the posterior P(y = 1∣X) according to Bayes's theorem:
    Equation (1)
    where y = 1 represents an exoplanet, y = 0 represents an FP, and X is a representation of the transit signal. Such generative approaches require the detailed knowledge of the likelihood, P(Xy), and prior, P(y), for each class (exoplanet vs. FP) and class scenario (e.g., BEB). While, in general, both the likelihood and priors can be learned in a data-driven approach (using ML), vespa estimates them by simulating a representative population with physically and observationally informed assumptions (Morton 2012; Morton et al. 2016). vespa FP probability (fpp) has been used to validate exoplanets (Morton et al. 2016) when the posterior probability P(y = 1∣X) is >0.99.
  • 3.  
    The discriminative approach directly calculates the posterior probability by considering a functional form for P(yX). It includes all other models listed in Table 1, i.e., Autovetter (Jenkins et al. 2014; McCauliff et al. 2015), AstroNet (Shallue & Vanderburg 2018), ExoNet (Ansdell et al. 2018), GPC/RFC (Armstrong et al. 2021), and ExoMiner, the model proposed in this work. These models differ in the input data and function they use to estimate P(yX). The functional form of P(yX) in Autovetter and RFC is a random forest classifier. As such, they are limited to accepting only scalar values. The functional form for AstroNet, ExoNet, and ExoMiner is a DNN that can handle different types of data input. AstroNet only uses phase-folded flux data. ExoNet adds stellar parameters and centroid motion time series as input. ExoMiner uses the unique components of data from the one-page DV summary report, including scalar and nonscalar data types. GPC is a Gaussian process classifier able to use both scalar and nonscalar data types as input once a kernel (similarity) function between data points is defined. GPC combines different scalar data types with the phase-folded flux time series. The scalar and nonscalar input types used for all these classifiers are summarized in Table 1. Similar to the generative approach, one can validate exoplanets using discriminative models if the model is calibrated and P(y = 1∣X) is sufficiently large.

Table 1. Comparison of Different Models

   vespa Robovetter Autovetter AstroNet ExoNet GPC RFC ExoMiner
ObjectiveVal. a 3C-Vet. b 3C-Vet.2C-Vet. c 2C-Vet.2C-Vet.2C-Vet.2C-Vet. & Val.
Model ConstructionGen. d Exp. e Dis. f Dis.Dis.Dis.Dis.Dis.
Scalar inputsPrimary properties g   
 Transit fit model   
 Stellar parameters   
 Optical ghost test   
 Bootstrap fap   
 Rolling band-fgt     
 Secondary properties h    
Nonscalar inputsUnfolded flux +      
 Phase-folded flux +++
 Odd and even views ++  ++
 Weak secondary flux ++  ++
 Difference image ++  + +
 Centroid motion test ++   

Notes. ✓ Means that the model uses the original nonscalar time series or image as its input; + means that the model uses only a few scalar values summarizing the time series or image as its input.

(a) Validation. (b) Three classes vetting. (c) Two classes vetting. (d) Generative. (e) Expert system. (f) Discriminative. (g) Includes MES, planet radius, depth, duration, etc. (h) Geometric albedo, planet effective temperature, and MES for secondary.

Download table as:  ASCIITypeset image

The advantage of generative approaches is that they can also be used to generate new transit signals, hence the generative name. However, the accuracy of a generative model is directly affected by the validity of the assumptions made about the form of the likelihood function and the values of priors. These can be obtained in a fully data-driven manner (through ML methods) or using the existing body of domain knowledge, or a combination of both (e.g., vespa). For example, to make the calculation of likelihood tractable, Morton (2012) represented a transit signal X by a few parameters: transit depth, transit duration, and duration of the transit ingress and egress, and generated the likelihood P(Xy) using (1) a continuous power law for exoplanet and (2) simulated data from the galactic population model, TRILEGAL (Girardi et al. 2005), for FP scenarios. Such an approach relies on the existing body of knowledge regarding the form of the likelihood and the values of priors, which could be paradoxical, as mentioned in Morton (2012). For example, the occurrence rate calculation requires the posterior probability, which itself requires the occurrence rate values. Moreover, reducing the shape to only a few parameters defining the vertices of a trapezoid fitted to the transit or transit model is very limiting. The shape of a transit is continuous and cannot be defined by a small number of parameters when you consider the effect of starspots. More generally, a transit signal X is summarized by only a few scalars, whereas to fully represent a transit signal for the task of classification, more data about the signal are needed, as we listed in Section 2.1. Full representation of a transit signal using the approach taken by Morton (2012) is not tractable. To be precise, Morton (2012) answers the following question: what is the fpp of a signal that has a specific depth, duration, and shape? Adding more information will change this probability. Refer to Morton & Johnson (2011) and Morton (2012) for the list of other assumptions.

Even though both the likelihood p(Xy) and priors p(y) can be learned using data with a more flexible representation for transit signal X (e.g., flux) and modern generative approaches to ML (e.g., DNNs), discriminative classifiers are preferred if generating data is not needed. The compelling justification of Vapnik (1998) to propose the discriminative approach was that "one should solve the [classification] problem directly and never solve a more general problem as an intermediate step [such as modeling p(Xy)]."

The performance of existing discriminative approaches for transit signal classification depends on their flexibility in accepting and including multiple data types (as provided in DV reports) and on their ability to learn a suitable representation (or feature learning) for their input data. Among the existing machine classifiers, the ones introduced in Armstrong et al. (2021), e.g., RFC and GPC, are the most accurate ones, as we will see in the experiments. Existing DNN models such as AstroNet and ExoNet are not as successful as the former since they do not use many diagnostic tests required to identify different types of FPs (check Table 1). In contrast, RFC and GPC are relatively inclusive of different diagnostic tests (except the centroid test) required for the classification of transit signals. However, these classifiers (1) are not able to learn a representation for nonscalar data automatically—GPC needs to be provided with a kernel function (similarity) prior to training, and RFC is not able to directly receive nonscalar data such as flux—and (2) are not inclusive of nonscalar (time series) data types and only use the results of statistical tests for most time series data such as odd & even, weak secondary, and centroid.

Similar or small modifications have been made to the aforementioned classifiers to perform classification/evaluation tasks for K2 (Dattilo et al. 2019), TESS (Armstrong et al. 2016; Yu et al. 2019; Osborn et al. 2020), and the Next-Generation Transit Survey (NGTS; Chaushev et al. 2019).

2.3. Validation of Exoplanets

Given that traditional confirmation of PCs was not possible or practical owing to the increase in the number of candidates and their specifics (e.g., small planets around faint stars), multiple works were developed aimed at statistically quantifying the probability that signals are FPs (Torres et al. 2004; Sahu et al. 2006; Morton 2012). Technically, any classifier that can classify transit signals into PC/non-PC can be used for validation. However, such classifiers need to be employed with extra scrutiny to make sure their validation is reliable.

Existing approaches to planet validation use extra vetoing criteria in order to improve the precision of the generated catalog of validated planets. These criteria are often model specific and designed to address issues related to each model. For example, vespa (Morton et al. 2016) is not designed to capture FPs due to stellar variability or instrumental noise, and it might fail under other known scenarios such as offset blended binary FPs and EBs that contaminate the flux measurement through instrumental effects, such as charge transfer inefficiency (Coughlin et al. 2014). Based on Morton et al. (2016), vespa is only reliable for KOIs that passed existing Kepler vetting tests. To avoid FPs due to systematic noise, vespa might not work well for the low-MES region, and this is why KOIs in the low-MES region are rejected for validation.

In order to improve the reliability of validating new exoplanets, Armstrong et al. (2021) required the new exoplanets to satisfy the following conditions: (1) the corresponding KOI has MES > 10.5σ, as recommended by Burke et al. (2019); (2) its vespa fpp score is lower than 0.01; (3) all of their four classifiers assign a probability score higher than 0.99 to that TCE; and (4) the corresponding KOI passes two outlier detection tests.

2.4. Contributions/Novelty of This Work

To classify transit signals and validate new exoplanets in this work, we propose a new deep learning classifier called ExoMiner that receives as input multiple types of scalar and nonscalar data as represented in the one-page DV summary report and summarized in Table 1. ExoMiner is a fully automated, data-driven approach that extracts useful features of a transit signal from various diagnostic scalar and nonscalar tests. The design of ExoMiner is inspired by how an expert examines the diagnostic test data in order to identify different types of FPs and classify a transit signal.

Additionally, ExoMiner benefits from an improved ML data preprocessing pipeline for nonscalar data (e.g., flux data) and an explainability framework that helps experts better understand why the model classifies a signal as a planet or FP.

We perform extensive studies to evaluate the performance of ExoMiner on different data sets and under different imperfect data scenarios. Our analyses show that ExoMiner is a highly accurate and robust classifier. We use ExoMiner to validate 301 new exoplanets from a subset of KOIs that are not confirmed planets or certified as FPs to date.

3. Methodology

3.1. Machine Classification

Belonging to the general class of supervised learning in ML and statistics, machine classification refers to the problem of identifying a class label of a new observation by building a model based on a prelabeled training set. Let us denote by x and y the instance (e.g., a transit signal) and its class label (e.g., exoplanet or FP), respectively. x and y are related through function f as y = f(x) + epsilon, with noise epsilon having mean zero and variance σ2. Provided with a labeled training data set ${ \mathcal D }=({x}_{1},{y}_{1}),\,({x}_{2},{y}_{2}),...({x}_{N},{y}_{N})$, the task of classification in ML is focused on learning $\hat{f}(x,D)$ that approximates f(x) to generate label y given instance x.

A good learner, discriminative or generative, should be able to learn from the training set and then generalize well to unseen data. In order to generalize well, a supervised learning algorithm needs to trade off between two sources of error: bias and variance (Bishop 2006). To explain this, note that the expected error of $\hat{f}(x,{ \mathcal D })$ when used to predict the output y for x can be decomposed as (Bishop 2006)

Equation (2)

where the expectation ${\mathbb{E}}$ is taken over different choices of training set ${ \mathcal D }$. The first term on the right-hand side, i.e., ${\left({{\mathbb{E}}}_{{ \mathcal D }}[\hat{f}(x,{ \mathcal D })]-f(x)\right)}^{2}$, is called bias and refers to the difference between the average prediction of the model learned from different training sets and the true prediction. A learner with high bias error learns little from the training set and performs poorly on both training and test sets; this is called underfitting. The second term on the right-hand side of this equation, i.e., ${{\mathbb{E}}}_{{ \mathcal D }}[{{\mathbb{E}}}_{{ \mathcal D }}[\hat{f}{(x,{ \mathcal D })]-\hat{f}(x,{ \mathcal D })}^{2}]$, is called variance and is the prediction variability of the learned model when the training set changes. A learner with high variance overtrains the model to the provided training set. Therefore, it performs very well on the training set but poorly on the test set; this is called overfitting. The last term, i.e., σ2, is the irreducible error.

A successful learner needs to find a good trade-off between bias and variance by understanding the relationship between input and output variables from the training set. Such learners are able to train models that perform equally well for both training and unseen test sets. There are three factors that determine the amount of bias and variance of a model: (1) the size of the training set, (2) the size of x (its complexity or number of features), and (3) the size of the model in terms of the number of parameters. Complex models with many learnable variables, relative to the size of training set, can always be trained to perform very well on the training set but poorly on unseen cases. Different regularization mechanisms are designed in ML to reduce the complexity of the model automatically when there are not enough training data. This includes (1) early stopping of the learning process by monitoring the performance of the model on a held-out part of the training set, called a validation set; and (2) penalizing the effective number of model parameters in order to reduce the model complexity. We will utilize both in this work.

3.2. DNNs

Generally considered as discriminative models (Ng & Jordan2002), neural networks (NNs) are nonlinear computing systems that can approximate any given function (Bishop 2006). An NN consists of many nonlinear processing units, called neurons. A typical neuron computes a weighted sum of its inputs and then passes it through a nonlinear function, called an activation function (Figure 3(a)), to generate its output:

Equation (3)

Figure 3.

Figure 3. Main elements of the NN used in this work.

Standard image High-resolution image

Neurons in an NN are organized in layers. In the most common NNs, neurons in each layer are of a similar type, and the output of neurons in one layer is the input to the neurons in the next layer. Neuron types are characterized by the restrictions on their weights (how many nonzero weights) and the type of activation function, f (Equation (3)), that they use. In this work, we use the following layers:

  • 1.  
    Fully connected (FC) layers: Shown in Figure 3(b), a neuron in an FC layer receives the outputs of all neurons from the previous layer (no restriction on the weights). We use the Parametric Rectified Linear Unit (PReLU) activation function for the neurons in the FC layer because it is well behaved in the optimization process. PReLU, defined as follows, is a variation of the ReLU function ($f(x)={x}^{+}=\max (0,x)$) that avoids the dying neuron 19 problem:
    Equation (4)
    where a is a parameter learned from the data.
  • 2.  
    Dropout layers: Shown in Figure 3(b), these layers are a form of regularization designed to prevent overfitting. The number of neurons in this layer is equivalent to the number of neurons from the previous layer. This layer is designed to drop a percentage of processing neurons (based on a rate called the dropout rate) in the previous layer to make the model training noisier (by not optimizing all weights in each iteration 20 ) and so more robust.
  • 3.  
    One-dimensional (1D) convolutional layers: As shown in Figure 3(c), the neurons in convolutional layers, which are called filters, operate only on a small portion of the output of the previous layer (i.e., only a subset of weights are nonzero in Equation (3)). The neurons in a convolutional layer extract features locally using small connected regions of the input. They are designed specifically for temporal or spatial data where there is locality information between inputs. Each neuron extracts features locally on many patches of the input (e.g., sliding windows on a time series). Filters in a convolutional layer produce a feature map from the previous layer. Similar to the FC layer, we use PReLU as the activation function for the neurons in this layer.
  • 4.  
    Pooling layers: Shown in Figure 3(c), pooling layers are a form of regularization and designed to downsample feature maps generated by the convolutional layers. These layers perform the same exact operation over patches of the feature map. Normally, the operation performed by the pooling layer is fixed and not trainable. The maximum and mean are the most common operations used in pooling layers. In this paper, we use 1D maxpooling layers that calculate the maximum values of sliding windows over the input.
  • 5.  
    Subtraction layers: These layers receive two sets of input variables and perform element-wise subtraction of one set from the other. The weights in this layer are fixed, and each neuron has two inputs: the weight is +1 for one and −1 for the other.

A DNN is an NN with a larger number of layers. Each layer of neurons extracts features on top of the features extracted from the previous layer. A set of layers combined together to follow a specific purpose is called a block. In this paper, we use two types of blocks: a convolutional block that consists of one or multiple convolutional layers with a final maxpooling layer (Figure 5(b)), and an FC block that consists of one or more FC layers, each followed by a dropout layer (Figure 5(c)). A DNN with convolutional layers is called a CNN (LeCun et al. 1989).

3.3. Explainability of DNNs

Explainability remains a central challenge for adopting and scaling DNN models across all domains and industries. Various perturbation- and gradient-based methods have been developed for providing insight into the predictions made by DNN classifiers.

Occlusion sensitivity mapping is among the simplest of these methods, and it provides a model-agnostic approach for localizing regions of the input data that have the greatest impact on a model's classifications (Zeiler & Fergus 2014). This region can then be considered in the context of the classification problem to illuminate how the model makes its predictions. For time series data, occlusion sensitivity mapping involves systematically masking the points of the input data that fall within a sliding window and analyzing the changes in the model predictions as different sliding windows are occluded.

Noise sensitivity mapping is a similar perturbation-based explainability method where noise is added to the points of the input data that fall within each sliding window and a heat map is generated to identify the input regions where the addition of noise has the greatest effect on the model predictions (Greydanus et al. 2018). The advantage of this technique is that no sharp artifacts or edges that may bias the classification are introduced into the input data when the sliding window is applied.

Finally, Gradient-weighted Class Activation Mapping (Grad-CAM) uses the gradients of the final convolutional layer, rather than the input data, to develop class-specific heat maps of the regions most important to the model classifications (Selvaraju et al. 2019). Compared to perturbation-based explainability methods, Grad-CAM is much less computationally expensive because it does not require repeatedly making predictions as the sliding window is moved and it eliminates the subjectivity of choosing the size of the sliding window.

3.4. Proposed DNN Architecture

To specify a DNN model, one needs two sets of parameters: (1) parameters related to the general architecture of DNNs (e.g., how to feed inputs to the model, the number of blocks and layers, the type of layers/filters, etc.)—these parameters are usually decided using domain knowledge, hyperparameter optimization tools, or a combination of both—and (2) the particular weights between neurons. The weights are variables optimized using a learning algorithm that optimizes a data-driven objective function (using, e.g., gradient descent; Bishop 2006). The parameters in the first set and those related to learning (such as batch size, type of optimizer, and loss function) are called hyperparameters and must be fixed prior to learning the second set.

We utilize domain knowledge to preprocess and build representative data to feed to our DNN model and also to decide about its general architecture. Given that an expert vets a transit signal by examining its DV report or similar diagnostics, we will use unique components of the DV report as inputs to our DNN model. This includes full-orbit and transit-view flux data, full-orbit and transit-view centroid motion data, transit-view secondary eclipse flux data, transit-view odd & even lux data, stellar parameters, optical ghost diagnostic (core and halo aperture correlation statistics), bootstrap false-alarm probability (bootstrap-pfa), rolling band contamination histogram for level zero (rolling band-fgt), orbital period, and planet radius (Twicken et al. 2018). We also include scalar values related to each test to make sure the test is effective. These scalar values include the following:

  • 1.  
    Secondary event scalar features: geometric albedo comparison (Section 2.1) and planet effective temperature comparison statistics (Section 2.1), weak secondary maximum MES, and weak secondary transit depth.
  • 2.  
    Centroid motion scalar features: flux-weighted centroid motion detection statistic, centroid offset to the target star according to the Kepler Input Catalog (KIC; Brown et al.2011), centroid offset to the out-of-transit centroid position (OOT), 21 and the respective uncertainties for these two centroid offsets. These scalar values provide complementary information to the transit and full-orbit views of the centroid motion time series, which are helpful for correctly classifying FPs (mainly BEBs).
  • 3.  
    Transit depth, which is important to understand the size of the potential planet. The depth information is lost during our normalization of the flux data. 22 We explain the details of how we create these data inputs in Section 3.5.

The general architecture of the proposed DNN model is depicted in Figure 4, in which we use a concept called convolutional branch to simplify the discussion. A convolutional branch (Figure 5(a)) consists of multiple convolutional blocks (Figure 5(b)). After each convolutional branch (yellow and orange boxes for full-orbit and transit views, respectively), an FC layer (light-blue boxes) is utilized to combine extracted features and scalar values related to a specific diagnostic test to produce a final unified set of features. These final sets of features are then concatenated with other scalar values (stellar parameters and other DV diagnostic tests) to feed into an FC block (dark-green box) that extracts useful, high-level features from all these diagnostic test-specific features. The results are fed to a logistic sigmoid output to produce a score in [0, 1] that represents the DNN's likelihood that the input is a PC. A default threshold value of 0.5 is used to determine the class labels but can be changed based on the desired level of precision/recall. Below, we enumerate some more details about the ExoMiner architecture:

  • 1.  
    The sizes of the inputs to the full-orbit and transit-view convolutional branches are 301 and 31, respectively. This is different from the previous works (Ansdell et al. 2018; Shallue & Vanderburg 2018; Armstrong et al. 2021) that use input sizes of 2001 and 201. Our design is inspired by the binning scheme utilized in the DV reports, which yields smoother plots. Note that the binning for the full-orbit and transit views on the DV one-page summary reports employs five bins per transit duration (width =duration × 0.2). Although this generates 31 bins for transit views of most TCEs, this may produce many more bins than 301 for the full-orbit view. However, we fixed the number to 301 because the size of the input to the classifier must be the same for all TCEs.
  • 2.  
    Inspired by the design of the DV reports, which only include transit views for the secondary and odd & even tests, we only used the transit views for these tests; note that the full-orbit view does not provide useful information for these diagnostic tests.
  • 3.  
    Unlike ExoNet, which stacked the flux and centroid motion time series into different channels of the same input and fed them to a single branch (Figure 6), we feed time series related to each test to separate branches because the types of features extracted from each diagnostic test are unique to that test.
  • 4.  
    For the odd & even transit depth diagnostic test, we use the odd and even views as two parallel inputs to the same processing convolutional branch. This design ensures that the same exact features are extracted from both views. The extracted features from the two views are fed to a untrainable subtraction layer that subtracts one set of features from the other. The use of a subtraction layer is inspired by the fact that domain experts check the odd and even views for differences in the transit depth and shape.
  • 5.  
    The FC layer after each convolutional branch is designed with two objectives in mind: (1) it allows us to merge the extracted features of a diagnostic time series with its relevant scalar values, and (2) by setting the same number of neurons in this layer, we ensure that the full-orbit and transit-view branches have the same number of inputs to the FC block. This is different from Ansdell et al. (2018) and Shallue & Vanderburg (2018), in which the output sizes of the full-orbit and transit-view convolutional branches are different by design.

Figure 4.

Figure 4.  ExoMiner architecture.

Standard image High-resolution image
Figure 5.

Figure 5. (a) Each convolutional branch consists of N convolutional blocks, (b) each convolutional block consists of M convolutional layers and one maxpooling layer, and (c) the FC block consists of multiple FC layers, each followed by a dropout layer.

Standard image High-resolution image
Figure 6.

Figure 6.  ExoNet: DNN architecture used in Ansdell et al. (2018).

Standard image High-resolution image

So far, we have utilized domain knowledge to design the general structure of our DNN. However, the total number of convolutional blocks per branch, the number of convolutional layers per block, and the number of FC layers in the FC block are hyperparameters of the model that need to be decided prior to training the model. Other hyperparameters include the kernel size for each convolutional layer, the number of kernels for the convolutional layers, the number of neurons for the FC layers, the type of optimizer, and the learning rate (step size in a gradient descent algorithm). We use a hyperparameter optimizer called BOHB (Bayesian Optimization and HyperBand, Falkner et al. 2018) to set these parameters prior to training.

The above process of combining domain knowledge and ML tools, such as hyperparameter optimization, provides us with a DNN model architecture that can be trained in a data-driven approach to classify transit signals. In a sense, this is similar to Robovetter, which leverages domain knowledge in order to design the general classification process in the form of if-then rules with if-then condition threshold variables that are partially optimized in a data-driven approach. The difference is the amount of automation in the process. Robovetter relies fully and comprehensively on the domain knowledge not only for feature extraction (the results of diagnostic tests) and classification rules (if-then conditions) but also to set the values for most of the variables. By contrast, our ML process limits the use of domain knowledge to the choice of inputs, the preprocessing of those inputs, and the general architecture of the DNN. Then, by optimizing many different aspects of the architecture through hyperparameter optimization and the optimization of the model's connection weights in a data-driven training process, our DNN learns useful features from raw diagnostic tests to classify transit signals.

The complete ML process ensures that the model is able to correctly classify the training data and also generalize to future unseen transit signals. This results in a machine classifier that can be quickly and easily used to classify other data sets (e.g., a model learned on Kepler data can be used to classify transit-like signals in TESS data). Overall, Robovetter is much closer to the manual vetting process than the machine classifier presented in this paper.

Algorithm 1. Detrending: Removing Colored Noise

Input: Time Series with Colored Noise
Output: Detrended Time Series
1: Remove missing values from the time series.
2: Split the array into new arrays if the time interval between adjacent cadences is larger than 0.75 days.
3: Remove the transits using adjusted padding min(3 × transit duration, ∣weak secondary phase∣, period) around the center of the transits and then linearly interpolate across the transits (see footnote a; this ensures that we remove all the transits for the primary TCE while minimizing changes to the rest of the time series, e.g., deformation of a potential secondary event).
4: Fit a spline for each gapped and linearly interpolated array.
5: Generate detrended arrays by dividing each array in the original time series of step 2 by the fitted spline.
6: Return the detrended arrays.

Download table as:  ASCIITypeset image

3.5. Data Preparation

In this section, we describe the steps performed to build the diagnostic tests required as inputs to the proposed DNN in Figure 4. The process includes the following two major steps:

  • 1.  
    In the first step, we create each diagnostic test time series from the Presearch Data Conditioning flux or the Moment of Mass centroid time series in the Q1–Q17 DR25 light-curve FITS files (Thompson et al. 2016b) available at the Mikulski Achive for Space Telescopes (MAST). 23 This process consists of (1) a preprocessing and detrending step (Algorithm 1) that removes low-frequency variability usually associated with stellar activity while preserving the transits of the given TCE and (2) a phase-folding and binning step for each test in order to generate full-orbit and transit views (Algorithm 2). Phase-folding uses the period signature of the signal to convert the time series to the phase domain, while binning creates a more compact and smoother representation of the transit signal by averaging out other signals present in the time series that are not close to the harmonic frequencies. Both of these steps are similar to Shallue & Vanderburg (2018); however, there are some minor but important differences in our algorithms. In detrending, we use a variable padding as a function of the period and duration of the transit to make sure we remove the whole transit prior to the spline fit (Algorithm 1). In phase-folding and binning, we use fewer bins for the transit and full-orbit views (31 and 301 instead of 201 and 2001, respectively) to obtain smoother time series. Similar to Shallue & Vanderburg (2018), we use (bin width = duration × 0.16) for the transit views. 24 However, for the full-orbit view we adjust the bin width to make sure the bins cover the whole time series and converge to the bin width of the transit views when possible (Algorithm 2).
  • 2.  
    In the final step, we rescale different diagnostic tests to the same range to improve the numerical stability of the model and reduce the training time. The rescaling also helps to ensure that different features have an equal and unbiased chance in contributing to the final output; however, if the range of values provides critical information, rescaling leads to the loss of information. In what follows, we introduce rescaling treatments specific for different time series data and provide a solution for cases where the rescaling leads to the loss of some critical information. Algorithm 3 describes the general normalization approach for most scalar features that follow a normal distribution. As we explain later in this section, we introduce a different normalization when the normal distribution assumption does not hold for a given scalar feature.

Here we detail the specific preprocessing steps for each diagnostic test fed to the DNN:

  • 1.  
    Full-orbit and transit view of the flux data: We first perform standard preprocessing in Algorithm 1 on the flux data and then run Algorithm 2 to generate both full-orbit and transit views. To normalize the resulting views of each TCE, we subtract the median medf =median(flux) from each view and then divide it by the absolute value of the minimum: ${\min }_{f}=| \min ({flux})| $. This generates normalized views with median 0 and minimum value −1 (fixed depth value for all TCEs). As a result of this rescaling, the depth information of the transit is lost. Thus, we add the transit depth as a scalar input to the FC layer after the transit-view convolutional branch (Figure 4). We normalize the transit depth values using Algorithm 3.
  • 2.  
    Transit view of secondary eclipse flux: First, we use the standard preprocessing outlined in Algorithm 1 on the flux data. We then remove the primary transit by using a padding size that ensures that the whole primary is removed without removing any part of the secondary using the following padding strategy: if the transit duration < ∣secondary phase∣, we remove 3 ×transit duration around the center of the primary transit. Otherwise, we take a more conservative padding size of ∣secondary phase∣ + transit duration, which removes the primary up to the edge of the secondary event. This padding strategy is large enough to make sure that the whole transit is removed; however, it is not so large that it removes part of the secondary. Also note that this padding is different from the one we used in Algorithm 1; here, we aim to gap the primary without damaging the secondary event. For spline fitting in Algorithm 1, we aim to make sure we completely gap the transits associated with the TCE.

    Algorithm 2. Generating Full-orbit and Transit Views (see footnote a; note that the bins may or may not overlap depending on the transit duration and period)

    Input: Original Time Series
    Output: Phase-folded Time Series
    1: Fold the data over the period for each TCE with the transit centered.
    2: Use 301 bins of $width=max\left(period/301,transit\,duration\,\times \,0.16\right)$ to bin data for full-orbit view.
    3: Use 31 bins of $width=transit\,duration* .16$ to bin data for transit view. The transit view is defined by $5\times transit\,duration$ centered around the transit.
    4: Return full-orbit and transit views.

    Download table as:  ASCIITypeset image

    Algorithm 3. Normalizing Scalar Feature

    Input: X: A single scalar feature of all TCEs in training set
    Output: $X^{\prime} $: Normalized feature of all TCEs
    1: Compute the median medX and standard deviation ${\delta }_{X}$ from the training set.
    2: Compute Xtemp by subtracting medX from values in X and divide them by ${\delta }_{X}$ (i.e., ${X}_{\mathrm{temp}}=\tfrac{X-{{med}}_{X}}{{\delta }_{X}}$).
    3: Remove the outliers from Xtemp by clipping the values outside 20δ to produce $X^{\prime} $.
    4: Return (1) $X^{\prime} $ and (2) medX and ${\delta }_{X}$ for normalizing future test data.

    Download table as:  ASCIITypeset image

    We obtain the phase for the secondary from the Kepler Q1–Q17 DR25 TCE catalog available at the NASA Exoplanet Archive. 25 Using the phase in conjunction with the ephmerides for the primary transit, we then generate the transit view using Algorithm 2. We normalize this view by subtracting the medf and dividing by $\max ({\min }_{f},| {\min }({secondary})| )$. This normalization method preserves the relative depth of the secondary with respect to the primary event. As with the flux view, we add the scalar transit depth value of the secondary as an input to the FC layer. Besides the transit depth of the secondary event, three other scalar values that help distinguish hot Jupiters from EBs are added. These include the geometric albedo and planet effective temperature comparison statistics and the weak secondary maximum MES. We normalize these four scalar values using Algorithm 3.
  • 3.  
    Transit view of odd & even: We perform standard preprocessing using Algorithm 1 for the flux data. We then use Algorithm 2 separately for the odd and even transit time series to obtain the odd and even transit views. We normalize each odd and even time series by subtracting medf and dividing by minf . This rescaling using the full flux time series rescaling parameters ensures that the relative sizes of the odd and even time series are preserved.
  • 4.  
    Full-orbit and transit views of centroid motion data: Unlike Ansdell et al. (2018), which used heuristics to compute the centroid motion time series, we follow Jenkins et al. (2010a) and Bryson et al. (2013) to generate centroid motion data that preserve the physical distance (e.g., in arcseconds) of the transiting object from the target. Assuming that centroids are provided in celestial coordinates α (R.A.) and δ (decl.), we denote the R.A. and decl. components of centroid at cadence n by Cα (n) and Cδ (n) and their out-of-transit averages by ${\bar{C}}_{\alpha }^{\mathrm{out}}$ and ${\bar{C}}_{\delta }^{\mathrm{out}}$, respectively. If the observed depth of the transit is dobs, the location of the transit source can be approximated as
    Equation (5)
    By subtracting the location of the target stars, i.e., αtarget(n) and δtarget(n), from the location of the transit source, we obtain the amount of shift for the R.A. and decl. components of the centroid motion for each cadence as follows:
    The overall centroid motion shift can be computed as
    Equation (6)
    Algorithm 4 summarizes the steps performed to obtain the final centroid motion time series used to feed into the DNN. To normalize this resulting motion time series for each transit, we subtract the median of the view (centering time series) and divide by the standard deviation across the training set.This specific normalization ensures that the relative magnitudes of different centroid times series, which provide critical information for diagnosing BEBs, are preserved. Centering the time series makes it easier for the DNN to find relevant patterns; however, it leads to the loss of the centroid shift related to the target star position. Thus, we feed the centroid offset information generated in DV to the FC layer after the transit-view convolutional branch (fourth branch in Figure 4). The centroid shift scalar features utilized in this branch are the flux-weighted centroid motion detection statistic, difference image centroid offset from KIC position and OOT location, and associated uncertainties.

    Algorithm 4. Constructing Centroid Motion Full-Orbit and Transit Views

    Input: Raw Centroid Time Series
    Output: Processed Centroid Time Series
    1: To obtain Cα and Cδ , convert row and column centroid time series from the charge-coupled device frame to R.A. and decl.
    2: Compute ${\bar{C}}^{\mathrm{out}}$ as the average out-of-transit centroid position.
    3: Remove colored noise in Cα and Cδ using Algorithm 1.
    4: Multiply the whitened arrays by ${\bar{C}}^{\mathrm{out}}$ to obtain c(t) to recover the original scale. We perform this step because dividing by the spline fit in Algorithm 1 removes the scale.
    5: Compute the transit source location using Equation (5).
    6: Compute the overall centroid motion time series using Equation (6).
    7: Convert the overall centroid motion time series from units of degree to arcsecond.
    8: Run Algorithm 2 to obtain full-orbit and transit views.
    9: Normalize full-orbit and transit views by subtracting the median and dividing by the training set standard deviation.

    Download table as:  ASCIITypeset image

  • 5.  
    Other DV metrics: Scalar features either related to stellar parameters or representing results of other diagnostic tests are another type of data provided in DV reports that experts use to classify TCEs. Similar to Ansdell et al. (2018), we use stellar effective temperature (Teff), radius (R*), mass (M*), density (ρ*), surface gravity ($\mathrm{log}g$), and metallicity ([Fe /H]) obtained from the Kepler Stellar Properties Catalog Update for Q1–Q17 DR25 Transit Search (Mathur & Huber 2016) and Gaia DR2 (Brown et al. 2018), which revised stellar properties for Kepler targets in the KIC. Among several DV diagnostic metrics for Q1–Q17 DR25 (Twicken et al. 2018), we use six of them: optical ghost core aperture correlation statistic, optical ghost halo aperture correlation statistic, bootstrap-pfa, rolling band-fgt, fitted orbital period, and derived planet radius. These features are not provided in the other data inputs described above, or are lost through data preprocessing. Thus, we add them as inputs to our DNN model (referred to as DV Diagnostic Scalars in Figure 4) to complement the information that exists in other branches. Except for bootstrap-pfa, we normalize all scalar features using Algorithm 3. For bootstrap-pfa, we use Algorithm 3 to normalize log(bootstrap-pfa) because bootstrap-pfa values do not follow a normal distribution and standardizing the raw data does not yield values in a comparable range to the other inputs.

3.5.1. Data correction

ML algorithms are usually tolerant to some level of noise in the input/label data by resolving conflicting information (especially when a sizable amount of data is available). However, if the percentage of such noisy cases is high, the model is not able to learn the correct relationship between the input data and labels. Given that the size of our annotated data is limited and most diagnostic tests are only useful for a small subset of data, it is important that a data-driven approach receives the critical information in a correct format so that it can learn accurate concepts/patterns. Because of this, we make two corrections to the data. First, we adjust the secondary diagnostic information of a subset of TCEs whose secondary event metrics do not meet our needs. When two TCEs are detected for an EB system, we would like the secondary event to refer to the other TCE of that system. However, the pipeline removes the transits or eclipses of previously detected TCEs associated with a target star. For example, consider KIC 1995732, which has two TCEs with similar period, 77.36 days, and that are both part of the same EB system. As you can see from the full-orbit view of the first TCE of this system in Figure 7(a), there are two transits for this system. The secondary of the first TCE, depicted in Figure 7(b), points to the second TCE. However, the pipeline removes the transits associated with the first TCE to detect the second TCE of this system, which results in a failed secondary test for the second TCE (Figure 7(c)). To correct the secondary information of such systems, we search the earlier TCEs of each given TCE to make sure the secondary is mutually assigned, i.e., if a TCE represents the secondary of another TCE, the secondary of that TCE should represent the first TCE as well. There were a total of 2089 such TCEs in the Q1–Q17 DR25 TCE catalog. After correcting the secondary event for such TCEs, we also corrected the geometric albedo and planet effective temperature comparison statistics, weak secondary maximum MES, and transit depth. These corrections are important because without the correct secondary event assignment the model does not have enough information to classify these TCEs correctly. Note that the correcting procedure we explain above might fail if (1) the phase of the weak secondary or the epoch information of the two TCEs of the EB is not accurate, e.g., TCE KIC 10186945.2, and (2) if the period of the first detected TCE is double that of the second detected TCE, e.g., TCE KIC 5103942.2. The secondary event of such EBs are difficult for a data-driven model to classify correctly if other diagnostic tests are good.

Second, the period values listed in the Q1–Q17 DR25 TCE catalog differ from the period reported in the Cumulative KOI catalog for some TCEs, as seen in Figure 8. The periods reported for most of these TCEs in the Q1–Q17 DR25 TCE catalog are double or half that of the correct periods reported in the Cumulative KOI catalog. For EBs, this does not create a problem, as there are two tests created for these different scenarios: either the secondary test or odd & even test, depending on whether the period is half or double that of the true period. For exoplanets, however, an incorrectly reported period can create an artificial secondary event or a failed odd & even test when the period is double or half that of the true period. Examples of such scenarios are Kepler-1604 b and Kepler-458 b, which are described in Section 2.1 (Figures 2(a) and (b)). Therefore, we corrected the period of confirmed planets (CPs) when the values differ between the Q1–Q17 DR25 TCE and Cumulative KOI catalogs. There were a total of five such TCEs: 3239945.1, 6933567.1, 8561063.3, 8416523.1, and 9663113.2.

Figure 7.

Figure 7. KIC 1995732—an example of an EB with two TCEs having inconsistent secondary assignments.

Standard image High-resolution image
Figure 8.

Figure 8. Period in the Q1–Q17 DR25 TCE catalog vs. Cumulative KOI catalog.

Standard image High-resolution image

3.6. Data-driven Approach

We use a data-driven approach to optimize two sets of parameters: the hyperparameters and model parameters.

  • 1.  
    Optimizing hyperparameters: We use the BOHB algorithm (Falkner et al. 2018) to optimize hyperparameters related to (1) the architecture, i.e., the number of convolutional blocks for full-orbit and transit-view branches, the number of convolutional layers for each convolutional block, the filter size in convolutional layers, the number of FC layers, the number of neurons in FC layers (one for the FC layer after the convolutional branch and one for the FC layers in the FC block), the dropout rate, the pool size, and the kernel and pool strides; and (2) the training of the architecture: choice between Adam (Kingma & Ba 2014) and Stochastic Gradient Descent (SGD) optimizers, 26 and the learning rate (step size in gradient descent). The optimized architecture found using the BOHB algorithm is shown in Figure 9. Note that we use a shared set of hyperparameters for the full-orbit branches (flux and centroid) and another set of shared hyperparameters for the transit-view branches (flux, centroid, secondary, and odd–even) to reduce the hyperparameter search space. The full-orbit view branches have three convolutional blocks, and the transit-view branches have two convolutional blocks, each with only one convolutional layer. The FC block has four FC layers, each with 128 neurons, followed by a dropout layer with dropout rate = 0.03. The FC layer after the convolutional branch has only four neurons, which means that only a few features for each diagnostic test are used for classification.
  • 2.  
    Optimizing model parameters: After optimizing the hyperparameters and fixing the DNN architecture, the connection weights of the DNN are learned in a data-driven approach using the Adam optimizer with a learning rate = 6.73e-05, using the Keras (Chollet 2015) deep learning API on top of TensorFlow (Abadi et al.2015).

Figure 9.

Figure 9. Optimized DNN architecture using hyperparameter optimizer. Convolutional layers: Conv 〈 kernel size 〉-〈 number of filters 〉; maxpooling layers: Maxpooling 〈 pool size 〉-〈 stride size 〉; FC layer: FC 〈 number of neurons 〉; dropout layer: Dropout 〈 dropout rate 〉.

Standard image High-resolution image

4. Experimental Setup

4.1. Data Set

Previous deep learning studies for classifying Kepler's TCEs (Ansdell et al. 2018; Shallue & Vanderburg 2018) have used Autovetter training labels for the Q1–Q17 DR24 TCE catalog (Seader et al. 2015). As seen in Figure 10, the data distribution has changed significantly from the Q1–Q17 DR24 TCE catalog to the Q1–Q17 DR25 TCE catalog, and there are more TCEs with longer periods in Q1–Q17 DR25 than in Q1–Q17 DR24 because the statistical bootstrap was not used as a veto in the Transiting Planet Search (TPS) component of the Kepler pipeline for the former (Twicken et al. 2016). To obtain a more reliable and updated training set, similar to Armstrong et al. (2021), we utilized the Kepler Q1–Q17 DR25 TCE catalog and generated a working data set as follows: First, we removed all rogue TCEs from the list. Rogue TCEs are three-transit TCEs that were generated by a bug in the Kepler pipeline (Coughlin et al. 2016). For the PC category, we used the TCEs that are listed as CPs in the Cumulative KOI catalog. 27 For the AFP class, we used the TCEs in the Q1–Q17 DR25 list that are certified as FPs (CFP) in the Kepler Certified False Positive table (Bryson et al. 2015). According to Bryson et al. (2015), the CFP disposition in the CFP table is very accurate, and a change of disposition for this set is expected to be rare. For NTPs, while Autovetter uses a simpler test to obtain non-transit-like signals, Robovetter performs a more comprehensive set of tests. Thus, for the NTP class, we combine TCEs vetted as nontransits by Robovetter (any TCE from the Q1–Q17 DR25 TCE catalog that is not in the Cumulative KOI catalog) and TCEs reported as certified false alarms (CFAs) in the CFP table.

Figure 10.

Figure 10. Histograms of orbital period for the Q1–Q17 DR25 and Q1–Q17 DR24 TCE catalogs.

Standard image High-resolution image

Table 2 compares the data sets used in Ansdell et al. (2018), Shallue & Vanderburg (2018), Armstrong et al. (2021), and this paper. Note that there are a lot more PCs and AFPs in Shallue & Vanderburg (2018) and Ansdell et al. (2018) compared to our data set. This is because we only used CPs from the Cumulative KOI catalog and excluded PCs that are not yet confirmed in order to have a more reliable (less noisy) labeled set, similar to Armstrong et al. (2021). However, unlike Armstrong et al. (2021), which only used a subset of NTPs to create a more balanced data set, we included all NTPs. Similar to the previous ML works, we use a binary classifier to classify TCEs into PC and non-PC classes. Therefore, the classifier is indifferent to a training set with mixed labels between AFPs and NTPs. 28 Given that our training set has a lower rate of PCs (8.5% vs. 22.8% in Shallue & Vanderburg 2018 and Ansdell et al. 2018, and 33.3% in Armstrong et al. 2021, which makes it more imbalanced), building an accurate classifier that can correctly retrieve PCs is more difficult.

Table 2. Data Sets Used in This Paper Compared to Ansdell et al. (2018), Shallue & Vanderburg (2018), and Armstrong et al. (2021)

ClassOur Data SetArmstrong et al. (2021)Shallue & Vanderburg (2018) and Ansdell et al. (2018)
PC229122743600
AFP353831009596
NTP2477929592541
Total30609833315737

Download table as:  ASCIITypeset image

Similar to Ansdell et al. (2018), Shallue & Vanderburg (2018), and Armstrong et al. (2021), we take an 80%, 10%, and 10% split for training, validation, and test sets, respectively. However, in order to study the behavior of different models when the training/test split changes, we perform a 10-fold cross-validation (CV), i.e., we split the data into 10 folds, and each time we take one fold for test and the other nine folds for training/validation (eight folds for training and one fold for validation). Also, unlike the existing studies that split TCEs into training, validation, and test sets, we split TCEs by their respective target stars. TCEs detected for the same target star have the same light curve and stellar parameters. By splitting the data set at the TCE level, one shares data between the training, validation, and test sets and produces a biased and unrealistic test set that the model is not supposed to be exposed to during training and validation.

Splitting a data set based on target stars is more realistic given that the model needs to classify the TCEs of new target stars in the future. We do not want to train the model with test data so that we can more accurately evaluate the model. This new setup creates a more difficult ML task but ensures that the model can be used across missions (e.g., a model trained using the Kepler data can be employed to classify TESS data).

4.2. Baseline Classifiers

We compare the following classifiers:

  • 1.  
    Robovetter: We used the Robovetter scores for the Kepler Q1–Q17 DR25 TCE catalog (Coughlin et al. 2016). That version of Robovetter utilized all known CPs and CFPs to fine-tune its rules. To correctly understand its performance, however, it must be fine-tuned using only the training set and then evaluated on the test set. Given that it has used the data in the test set, the performance metrics we report on the test set here are an overestimate of the true performance. The correct numbers would be lower if the test set were hidden from Robovetter during parameter setting.
  • 2.  
    AstroNet: We used the AstroNet code available on github (Shallue & Vanderburg 2018) and preprocessed and trained the model using the same setup and DNN architecture as provided in Shallue & Vanderburg (2018) on the data set used in this paper. We also optimized the architecture for their model using our own data and hyperparameter optimizer BOHB. This led to slightly worse results, which was expected because the authors used more resources to optimize their architecture. Thus, we report only the results using their optimized architecture. Following Shallue & Vanderburg (2018), we trained 10 models for each of the 10 folds and report the average score for each fold in a 10-fold CV.
  • 3.  
    ExoNet: The original code of ExoNet is not available. Thus, we implemented it on top of the code available for AstroNet based on the details (including the DNN general architecture) provided in Ansdell et al. (2018). We optimized the details of the architecture for ExoNet using our own data and BOHB. This resulted in a better performance, which was expected, since Ansdell et al. (2018) did not optimize their architecture for the new set of inputs (flux and centroid motion time series data, and stellar parameters) and just used the one optimized by Shallue & Vanderburg (2018). Therefore, we report the results of the optimized architecture using BOHB. Following Ansdell et al. (2018), we train 10 different models and report the average score for each fold in a 10-fold CV.
  • 4.  
    Random forest classifier (RFC) in Armstrong et al. (2021): We used the scores provided by the authors. Note that Armstrong et al. (2021) used the Q1–Q17 DR25 TCE data set and similar labeling to us (CPs, CFPs/CFAs, and NTPs) and provided the scores for all TCEs using CV. However, Armstrong et al. (2021) split the data based on the TCE instead of the target star used in this work.
  • 5.  
    Gaussian process classifier (GPC) in Armstrong et al. (2021): We used the scores provided by the authors, similar to RFC above. We report GPC because it is the only classifier in Armstrong et al. (2021) that receives light curves directly as input.
  • 6.  
    ExoMiner: This is the ExoMiner classifier shown in Figure 9. All hyperparameters, including the DNN architecture, were optimized using our hyperparameter optimizer. Similar to AstroNet and ExoNet, we trained ExoMiner 10 times and report the average score for each fold in a 10-fold CV.
  • 7.  
    ExoMiner-TCE: This is the ExoMiner classifier shown in Figure 9 when trained on a data set split into training, validation, and test sets based on TCEs instead of target stars. All hyperparameters are optimized using our hyperparameter optimizer. We trained ExoMiner 10 times and report the average score for each fold in a 10-fold CV. The motivation for this baseline is to have a fair comparison with GPC and RFC, whose results are based on splitting the data at the TCE level.

4.3. Evaluation Metrics

To study and compare the performance of the above models, we use different classification and ranking metrics listed below:

  • 1.  
    Accuracy: This is the fraction of correctly classified TCEs in the test set. Given that our data set (and generally any realistic transit data set) is very imbalanced, with about 7.5% PCs, this metric is not very informative; To see this, note that a classifier that classifies all transits into the FP class has an accuracy of 92.5%. However, accuracy provides some insights when studied in conjunction with other metrics.
  • 2.  
    Precision: Also called positive predictive value in ML, this is the fraction of TCEs classified as PC that are indeed PC, i.e.,
    Note that as we make the data set on which we measure precision more imbalanced, the precision decreases for a fixed model. As an example, suppose we reduce the size of the positive class to $\tfrac{1}{k}$ of the original size (k > 1); the new precision would become
    Equation (7)
  • 3.  
    Recall: Also called the true positive rate, this is the fraction of PCs correctly classified as PC, i.e.,
  • 4.  
    Precision-Recall (PR) curve: The PR curve summarizes the trade-off between precision and recall by varying the threshold used to convert the classifier's score to a label. PR AUC is the total area under the PR curve. An ideal classifier would have an AUC of 1.
  • 5.  
    Receiver Operating Characteristic (ROC) curve: The ROC curve summarizes the trade-off between the true positive rate (recall) and FP rate (fall-out) when varying the threshold used to convert the classifier score to label. ROC AUC is the total area under the ROC.
  • 6.  
    Recall@p: This is the value of recall for precision = p. We are specially interested at p = 0.99 to study the behavior of the model for planet validation.
  • 7.  
    Precision@k (P@k): P@k is the fraction of the top k returned TCEs that are true PCs. P@k is a commonly used metric in recommendation and information retrieval systems in which the top k retrieved items matter the most. For the classification of TCEs, this metric is valuable because it provides a sensible view of interpreting the performance of the classification model in ranking the TCEs. Overall, P@k for a model shows how reliable that model is for recommending a PC for follow-up study or for validating new exoplanets. For example, if P@100 is 0.95, it means that there are only five non-PCs in the top 100 TCEs ranked by the model. More formally, P@k is defined as
    Equation (8)
    Note that if a model gives the same scores to the top m > k, for all such m and k, we have
    Equation (9)
    This is because the model does not provide any particular ordering for the top such m TCEs. To understand this, consider a model that gives a score of 1.0 for 100 TCEs out of which 50 are PCs. If we are only interested in knowing how many of the top 10 TCEs are PC, i.e., P@10, we have a 50/100 chance of success on average.

A few notes about the above metrics.

First, note that the ML classifiers we compare in this paper generate a score between 0 and 1 for each TCE. The higher the score generated for a TCE, the more confident the model is that the TCE is indeed a PC. In order to classify a TCE, one must use a threshold to convert the generated score into a label. The values of accuracy, precision, and recall are a function of this threshold, which can be set to obtain a desired precision/recall trade-off. Unless noted otherwise, we use the standard value of 0.5 for this threshold. When we report the results later in Table 4 to investigate the performance of different models, we put precision and recall in the same column (denoted Precision & Recall) because we can always change the threshold to trade off one versus the other. The idea is to have a model whose performance is better overall.

Second, note that the ROC AUC and PR AUC are not threshold dependent and are thus good metrics for evaluating the performance of different classifiers. The difference between ROC AUC and PR AUC is that the former measures the performance of the classifier in general and the latter measures how good the classifier is with an emphasis on the class of interest, in our case PC. In other words, the PR curve is more useful for needle-in-a-haystack-type problems where the positive class is rare and interesting.

Third, note that precision and recall are called reliability and completeness in exoplanet research terminology (Coughlin 2017).

Finally, note that P@k is indifferent to the specific scores of ranked items and only cares about the rank and relevancy. For example, the score for the item ranked at position l (lk) could be very low for P@k. P@k just measures how relevant the top items are independent of how large their scores are.

5. Performance Evaluation

The results of the classifiers trained in this work are reported in Table 3. Using these scores reported in this table, we study the performance of different classifiers in this section.

Table 3. Scores of Different Classifiers Trained in This Work

ColumnDescription
target_idKIC ID
tce_plnt_numTCE planet number
tce_periodTCE period
mesTCE MES
original_labelNTP, AFP, or PC
binary_label0 for NTP or AFP, 1 for PC
astronetScore assigned by AstroNet
exonetScore assigned by ExoNet
exominerScore assigned by ExoMiner
exominer_tceScore assigned by ExoMiner-TCE

Note. This table describes the available columns. The full table is available online.

Only a portion of this table is shown here to demonstrate its form and content. A machine-readable version of the full table is available.

Download table as:  DataTypeset image

5.1. Comparison on Kepler Q1–Q17 DR25 Data

Table 4 shows the performance of different classifiers using 10-fold CV. For each evaluation metric, we report its value over the Kepler Q1–Q17 DR25 data set after combining the scores for all test sets (10 nonoverlapping folds). We also report the mean and 99% confidence interval using Student's t-test over 10 folds in parentheses. Given that for Robovetter, RFC, and GPC the data splits for training, test, and validation sets are different from the ones we used here, the means and confidence intervals reported for these methods are approximations. However, because these two estimates (for the complete data set and for the average of the 10 folds) do not differ significantly for other classifiers, we can assume that the approximation is reasonable.

Table 4. Performance of Different Classifiers Using 10-fold CV

Model/MetricPrecision & RecallAccuracyPR AUCROC AUCP@100P@1000P@2200 a
     (P@50)(P@100)(P@200)
Robovetter b 0.951 & 0.9750.9940.9580.9940.966 c 0.966 c 0.966 c
 (0.950 ± 0.015) & (0.975 ± 0.000)(0.994 ± 0.001)(0.958 ± 0.019)(0.994 ± 0.003)(0.967 ± 0.019)(0.967 ± 0.019)(0.969 ± 0.017)
AstroNet 0.861 & 0.8850.9810.9250.9930.9500.9670.882
 (0.862 ± 0.022) & (0.885 ± 0.000)(0.981 ± 0.003)(0.927 ± 0.014)(0.993 ± 0.003)(0.976 ± 0.013)(0.967 ± 0.016)(0.906 ± 0.024)
ExoNet 0.925 & 0.8640.9850.9560.9950.9900.9890.916
 (0.925 ± 0.008) & (0.865 ± 0.000)(0.985 ± 0.002)(0.956 ± 0.007)(0.995 ± 0.001)(0.996 ± 0.009)(0.990 ± 0.008)(0.936 ± 0.016)
GPC b 0.921 & 0.9640.9910.9820.998 1.000 0.997 0.954
 (0.921 ± 0.011) & (0.964 ± 0.000)(0.991 ± 0.002)(0.982 ± 0.007)(0.998 ± 0.001)(0.996 ± 0.009)(0.997 ± 0.005)(0.972 ± 0.013)
RFC b 0.929 & 0.9550.9910.9790.9980.991 0.991 0.954
 (0.930 ± 0.015) & (0.956 ± 0.000)(0.991 ± 0.002)(0.979 ± 0.008)(0.998 ± 0.001)(0.992 ± 0.011)(0.992 ± 0.011)(0.970 ± 0.011)
ExoMiner 0.968 & 0.974 0.996 0.995 1.000 1.000 0.999 0.985
 (0.968 ± 0.011) & (0.974 ± 0.000)(0.996 ± 0.001)(0.995 ± 0.002)(1.000 ± 0.000)(1.000 ± 0.000)(1.000 ± 0.000)(0.990 ± 0.009)
ExoMiner-TCE 0.965 & 0.977 0.996 0.995 1.000 1.000 0.998 0.987
 (0.964 ± 0.011) & (0.978 ± 0.000)(0.996 ± 0.001)(0.996 ± 0.002)(1.000 ± 0.000)(1.000 ± 0.000)(1.000 ± 0.000)(0.991 ± 0.006)

Notes. The numbers are bold when the corresponding model outperforms other classifiers for an evaluation metric. For each metric, we report (1) on the first row the value computed over the full data set for each classifier and (2) on the second row and in parentheses the 99% confidence intervals using Student's t-test around the mean of the 10 folds in the CV for each classifier. Note that the values of P@k are computed for different values of k for these two options because the number of PCs for each fold is much less than that in the full data set.

a There are a total of 2291 confirmed exoplanets in the data set. b The confidence intervals for Robovetter, GPC, and RFC are only approximations. This is because Robovetter does not use CV, and GPC and RFC use different folds in CV. c Robovetter assigns a score of 1.0 to 1312 TCEs, of which there are a total of 1268 PCs. Thus, P@100 = P@1000 = 0.966. The value for P@2200 is also 0.966 by chance.

Download table as:  ASCIITypeset image

Overall, the performance of ExoMiner is better than all baselines, in terms of classification and ranking metrics, particularly when looking at existing deep learning models, i.e., AstroNet and ExoNet.

There are a few points to note here.

First, one should keep in mind that there is some level of label noise in the data set. Thus, any performance comparison needs to be done with caution. We would like to note that the label noise in the NTP and AFP classes is minimal. The major portion of the label noise is in the exoplanet (PC) category. Because the label noise in the positive class does not affect the performance of a model negatively in terms of metrics such as precision and P@k, it is hard to attribute the superior performance of ExoMinerto the label noise.

Second, note that there is little difference between the performance of ExoMiner and ExoMiner-TCE. As we see in the reliability analysis (Section 6), the size of the training set is large enough to cover different data scenarios even when we split by target stars. Thus, splitting by TCE or target star does not change the results significantly. Nonetheless, splitting by target star is the correct approach because of potential data leakage between training and test sets, as mentioned before.

Third, note how different classifiers compare in terms of P@k. In order to understand how many PCs exist in the top TCEs ranked by each model in the full data set, simply multiply P@k by k. For example, ExoMiner returns 999 PCs in the top 1000 ranked TCEs. These numbers are 991 and 966 for RFC and Robovetter, respectively. A model that can rank transit signals by putting PCs in the top of the list is highly desirable for occurrence rate calculations and follow-up studies to confirm new exoplanets.

Finally, note that ExoMiner has lower variability (smaller confidence intervals) compared to the baselines. This implies that ExoMiner is more stable with regard to changes in the training and test sets.

To get a better sense of the performance of different methods, we plotted their precision-recall curves in Figure 11(a). It is clear that ExoMiner outperforms the baselines significantly in all points. When validating new exoplanets, high values of precision are desirable, and ExoMiner has high recall values at high precision values. As a matter of fact, at a fixed precision value of 0.99, ExoMiner is able to retrieve 93.6% of all exoplanets in the test set. This number is 76.3% for the second-best model, GPC. We provide the recall values of different methods for different precision values in Table 5.

Figure 11.

Figure 11. Precision-Recall (PR), Precision@k, Score@k, and Score Distribution of different models.

Standard image High-resolution image

Table 5. Recall Value at a Desirable Precision Value (Recall@p)

MethodRecall@0.99Recall@0.975Recall@0.95
Robovetter 0.0 a 0.0 a 0.976
 0.364 ± 0.4840.433 ± 0.4850.974 ± 0.000
AstroNet 0.001 b 0.3330.643
 0.198 ± 0.1740.343 ± 0.1430.642 ± 0.012
ExoNet 0.5060.6190.803
 0.479 ± 0.1900.643 ± 0.090.806 ± 0.001
GPC 0.7630.8250.943
 0.716 ± 0.1560.848 ± 0.1010.94 ± 0.001
RFC 0.6770.8070.937
 0.563 ± 0.3140.776 ± 0.1150.935 ± 0.001
ExoMiner 0.936 0.966 0.986
 0.939 ± 0.0300.964 ± 0.0090.988 ± 0.000
ExoMiner-TCE 0.949 0.960 0.990
 0.95 ± 0.0200.961 ± 0.0150.991 ± 0.000

Notes. The numbers are bold when the corresponding classifier outperforms other classifiers. The numbers in parentheses show the 99% confidence interval around the mean of 10-fold CV using Student's t-test. The confidence intervals for Robovetter, GPC, and RFC are only approximations.

a Robovetter assigned a score of 1.0 to 1312 TCEs, from which there are 44 non-PCs. So no threshold can attain a precision of 0.99 or 0.975. b There are FPs with very high AstroNet scores.

Download table as:  ASCIITypeset image

We compare how different models perform in terms of ranking the top k TCEs in Figure 11(b). Note that ExoMiner is a better ranker compared to existing classifiers. To provide a clear comparison, there is no FP in the top 975 and 984 TCEs when ranked by ExoMiner and ExoMiner-TCE, respectively. This value is 0, 3, 26, 379, and 0 for Robovetter, AstroNet, ExoNet, GPC, and RFC, respectively. The reason that this is zero for Robovetter and RFC is because these models assign a score of 1.0 to many TCEs including FPs (check Equation (9)), which prevents sorting the top TCEs.

To further understand how the scores change at different rank positions and how the scores of different classifiers are distributed, we plot the scores of different models for the TCE at position k in Figure 11(c) and the score histogram in Figure 11(d). Note that there is not much difference between the behavior of different classifiers in terms of the score distribution. The better performance of ExoMiner is due to its better score assignment. However, ExoMiner is generally more confident about the label of TCEs. Generally speaking, ML models that use a more comprehensive list of diagnostic tests are more confident about their label too. There are 1844, 24, 58, 1399, 1694, 1507, and 1781 TCEs with score >0.99 for Robovetter, AstroNet, ExoNet, GPC, RFC, ExoMiner, and ExoMiner-TCE, respectively (top right panel in Figure 11(d)). ML classifiers such as AstroNet and ExoNet that do not use different diagnostic tests required to correctly distinguish exoplanets from FPs are less confident than ML classifiers such as GPC, RFC, and ExoMiner that utilize a more comprehensive list of diagnostic tests. As we will show in Section 6.1, the confidence of the model is also dependent on the size and the level of label noise in the training set.

5.2. Comparison on Kepler Planets Confirmed by Radial Velocity

Given that some of the Kepler-confirmed exoplanets could have the wrong disposition, in this section we report the performance of difference classifiers on a subset of TCEs that only includes those Kepler exoplanets confirmed using radial velocity. We obtained the list of Kepler exoplanets that are confirmed using radial velocity from the Planetary System table on the NASA Exoplanet Archive. 29 There are a total of 310 such confirmed exoplanets in the Q1–Q17 DR25 data set. Table 6 reports the performance of different classifiers on this data set that includes all NTPs and AFPs, but only PCs that are exoplanets confirmed using radial velocity. We report a single value for the full subset here because the results of averaging over 10 folds are very similar. Note that the recall value is equivalent to the accuracy of a classifier on the positive class; in our case, it counts the percentage of exoplanets confirmed by radial velocity that the model classifies as PC. As can be seen, ExoMiner consistently performs better than existing classifiers on this data set. Also, compared to the performance on the full data set, the precision for all classifiers decreases significantly. This is because there are significantly fewer confirmed planets in this data set (almost 1/7 of the original data set). As we discussed in Section 4.3, precision decreases for more imbalanced data sets (check Equation (7)).

Table 6. Performance of Different Classifiers Using 10-fold CV on the Data Set That Only Includes Those Kepler Exoplanets Confirmed by Radial Velocity

Model/MetricPrecision & RecallAccuracyPR AUCROC AUC
Robovetter 0.721 & 0.9680.9960.8000.992
AstroNet 0.438 & 0.8190.9870.6220.989
ExoNet 0.618 & 0.8390.9930.8160.994
GPC 0.601 & 0.9190.9930.8760.997
RFC 0.630 & 0.9130.9930.8250.998
ExoMiner 0.800 & 0.9550.9970.9520.999
ExoMiner-TCE0.785 & 0.9680.9970.9611.000

Download table as:  ASCIITypeset image

5.3. Comparison to Santerne et al. (2016)

Similar to Morton et al. (2016) and Armstrong et al. (2021), we report the distribution of scores for ExoMiner on those Kepler candidates for which Santerne et al. (2016) provided dispositions. Table 7 reports the mean and median scores. The behavior of ExoMiner on these disposition categories is very similar to existing classifiers. ExoMiner generates high scores for both planet and BD categories and low scores for EB and CEB categories. However, note that the scores for BD are well below the validation threshold of 0.99.

Table 7. Scores of ExoMiner for Santerne et al. (2016) Dispositioned KOIs

LabelNumberMeanMedian
Planet440.7730.861
EB480.1380.086
Contaminating EB (CEB)150.1750.04
Brown dwarf (BD)30.860.84
Undetermined180.7540.896

Download table as:  ASCIITypeset image

The mean and median score values of each category do not provide much detailed information about the performance of the model or score distribution of the model. Given that there are only a total of 110 dispositioned KOIs in Santerne et al. (2016) and it is balanced (44 planets and 56 FPs), we report the performance of different classifiers in terms of accuracy when different threshold values are used to label a KOI as a planet. Note that accuracy is more informative for this small balanced data set. 30 Figure 12 reports the results of this experiment. As you can see, ExoMiner and ExoMiner-TCE outperform all classifiers, except Robovetter, on this set of independently dispositioned KOIs. The fact that Robovetter performs well on this data set could be because Robovetter was fine-tuned on all known exoplanets and FPs at the time (including this data set), whereas the performance of all other comparing baselines are measured on a held-out test data set. Note that ExoMiner-TCE performs better than ExoMiner. Even though this is not a sizable data set, splitting by TCE seems to make the classification easier.

Figure 12.

Figure 12. Accuracy and number of detected planets for different classifiers on Santerne dispositioned KOIs (Santerne et al. 2016) when a threshold T is set for classification, i.e., we classify a signal as PC when the classifier score > T.

Standard image High-resolution image

Also, note that for higher threshold values we observed that all methods perform poorly on the planet category, as seen in Figure 12(b). This is because the planets in this set of KOIs are in the large planet region, for which classification is more difficult. Thus, the classifiers are less confident on this data set.

5.4. Comparison to vespa

Figure 13 compares the scores provided by ExoMiner with those provided by vespa for the KOIs. Note that we reported the scores for unused KOIs, i.e., those that are not part of our evaluation data set (combined training, validation, and test sets) using blue star markers. We will discuss this later in Section 9. Overall, ExoMiner and vespa agree on 76% of cases when a threshold of 0.5 is used. This number is 73% for GPC, as reported in Armstrong et al. (2021). vespa fpp scores are generally not accurate when they are in disagreement with ExoMiner scores. Similar behavior was observed in Armstrong et al. (2021). Figure 13(b) shows the score comparison for the corner cases. The numbers above each panel show the number of cases for each category. As can be seen, there are a lot of FPs (520 FPs) with low vespa fpp values (upper left corner of Figure 13(b)), which is the range where validation of exoplanets occurs. To correct this, Morton et al. (2016) proposed multiple vetoing criteria in addition to low vespa fpp for validation.

Figure 13.

Figure 13. Comparison with vespa (Morton et al. 2016). Confirmed exoplanets are shown as red circles, AFPs as yellow triangles, and unused KOIs as blue stars. Note that ExoMiner is often correct when ExoMiner and vespa disagree (upper left and lower right corners).

Standard image High-resolution image

For a validation threshold of 0.99, ExoMiner gives scores >0.99 to 19 dispositioned KOIs (strongly classifying them as exoplanets), while vespa gives fpp scores >0.99 (strongly classifying them as FPs). These are all confirmed exoplanets. On the other corner, ExoMiner gives scores < 0.01 to 363 dispositioned KOIs (strongly believe they are FPs) for cases that vespa fpp < 0.01 (strongly believe they are planets). Except one KOI in this list, i.e., Kepler-452 b, they are also FP. Based on this analysis, ExoMiner is generally more accurate than vespa in the validation threshold region.

6. Stability and Reliability of ExoMiner

In this section, we perform reliability analysis by (1) studying the sensitivity of ExoMiner to training set size and label noise and (2) examining some examples of misclassified exoplanets and FPs.

6.1. Training Set Sensitivity

We perform two tests in order to understand the sensitivity of ExoMiner with regard to imperfect training set scenarios, namely, training set size and label noise.

First, we would like to test the hypothesis that the size of the training set for this problem might not be large enough in order to cover different data scenarios, i.e., the existence of underrepresented scenarios in the training set. As we keep reducing the size of the training set, we start removing some of the more uncommon scenarios in the data set at some point, preventing the model from learning from these specific observations, and leading to a decrease in performance for the fixed test set. We study this using the following two settings:

  • 1.  
    Full training set: In this first setting, we reduce the size of the full training set as follows: (1) we split 80% of our data into training, 10% into validation, and 10% into test sets; and (2) we randomly select q% of training and validation sets to train a new model and check its performance on the held-out fixed test set generated in step 1. Figure 14(a) shows the results of this experiment where we report precision, recall, PR AUC, and average scores for each category (PC, AFP, and NTP) for q% varying between 1% and 100% over five random runs for each q% to curb stochastic effects due to the random selection of a subset of TCEs. As expected, the performance of ExoMiner decreases in terms of all evaluation metrics as the size of the training set decreases. However, it is very robust even up to 5%. The main effect is the decrease in the recall values and the narrowing of the differences between scores of PCs and FPs, as shown in Figure 14(b). To study the effect on the validation of new exoplanets when a threshold = 0.99 is used, we plotted in Figure 14(c) the precision, recall, and total number of cases for this threshold on the test set. For q = 5%, no TCEs in the test set receive scores >0.99, meaning that for a small enough training set there are no validated TCEs in this set. By increasing the size of the training set, the total number of TCEs with score >0.99 grows up to q = 20% and then does not change. ExoMiner becomes more confident of the labels it assigns to the test set TCEs for a larger training set size. Interestingly, at q = 10%, there is a large variance over the five runs of experiments in terms of both the number of TCEs with scores >0.99 and recall value. The higher variance for q = 10% means that 10% of the training set is not large enough for the model to classify PCs with high confidence. Depending on the amount of information provided by 10% of the training set to classify different scenarios in the test set, the total number of TCEs with scores >0.99 in the test set can vary largely.
  • 2.  
    Only AFPs: In this setting, we only reduce the number of AFPs in the training and validation sets in step 2 of the above procedure. This experiment tests the behavior of the ExoMiner for the hypothesis that some AFP scenarios might not be well represented in the training set compared to the variation that exists in the PC category. Figure 15 shows the performance when the percentage of AFPs is decreased down to 0%. As mentioned before, because the relative rate of planets versus AFPs increases when lowering the percentage of AFPs in the training set, the precision decreases and recall increases in the test set. However, even at 1% of AFPs, precision is not significantly affected. At the extreme case of 0% of AFPs, the precision decreases to around 86.5%, which is still reasonably good. The good performance observed at 0% of AFPs might be due to the presence of AFPs in the NTP category. A subset of AFPs in the NTP (non-KOI) category corresponds to the secondary eclipses detected for EBs, as we discussed in Section 4.3. Removing those 2089 AFPs in the NTP category leads to another 2% reduction of precision (precision = 0.842). This shows that there is an inherent label noise between these two categories and that there are still more AFPs in the NTP category that allow the learner to understand the difference between PCs and AFPs. Furthermore, these results seem to show that the label noise between PC and non-PC categories is minimal. If the PC category was contaminated significantly with FPs, we would expect to see a significant decrease in performance in the test set as the number of AFPs in the training set is reduced. More interestingly, precision at a validation threshold of 0.99 is not much affected even at as low as 1% of AFPs, as shown in Figure 15(c). Also, note that the average scores of AFPs in the test set increase with lower numbers of AFPs in the training set, but overall they are far from the classification threshold of 0.5.

Figure 14.

Figure 14. Randomly selecting q% TCEs from the original training set to train the model. The plots show the mean value ± standard deviation of different test set metrics over the five runs for each value of q. The right y-axis in panel (c) shows the total number of test set TCEs with scores >0.99.

Standard image High-resolution image
Figure 15.

Figure 15. Randomly keeping only q% of AFPs from the original training set to train the model. The plots show the mean value ± standard deviation of different test set metrics over the five runs for each value of q. The right y-axis in panel (c) shows the total number of TCEs in the test set with scores >0.99.

Standard image High-resolution image

Lower average scores (lower confidence of the model) for smaller training set size can be explained using bias/variance trade-off that we discussed in Section 3.1. When the model does not see enough examples, it can easily overfit to the training set owing to small sample size. To prevent overfitting, the learner automatically stops optimizing the model parameters on the training set when the performance on the validation set starts degrading. This early stopping results in a less confident model that has lower PC scores and higher FP scores. At the extreme 1% (or 0% for the second experiment), the average PC scores become very similar to average FP scores on the test set and the model performs poorly. To understand this behavior in terms of bias/variance trade-off, note that the variance of model increases with smaller training set, i.e., there is more variation of the model performance when the size of the training set is reduced. Thus, more regularization, including early stopping, is required in order to reduce variance. This will increase the bias. One can also look at this from a Bayesian perspective; the model aims to label an unseen (test set) transit signal x using the evidence provided by training set ${ \mathcal D }$. As we increase the size of ${ \mathcal D }$, the model becomes more confident about the label of x because the model has seen more similar cases.

Based on these results, ExoMiner shows robustness to the degree of representation of different subpopulations of PCs and FPs in the training set. Even in cases where those groups are misrepresented, ExoMiner is able to learn useful representations to distinguish between PCs and non-PCs and attain a reasonable performance.

Second, we would like to test the effect of label noise on the performance of ExoMiner. Similar to the previous experiment, we split the data into 80% training, 10% validation, and 10% test sets. We then artificially inject label noise into both training and validation sets by switching the labels of TCEs, train the model using this noisy data, and report its performance on the original held-out fixed test set. For each noise level setting, five runs are performed to curb stochastic effects due to the random selection of a subset of TCEs whose label is flipped. We run the following two experiments:

  • 1.  
    KOI label noise: Given that distinguishing PCs from AFPs is a more difficult task, we only flip the label of q% of AFPs to PCs and q% of PCs to AFPs in this experiment (no added label noise in NTP category). Figure 16(a) shows the performance of ExoMiner for varying values of q from 1% to 50%. Even at q = 40%, ExoMiner is able to perform well on the test set. This result is consistent with the previous research (e.g., Rolnick et al. 2017). However, ExoMiner's confidence on its assigned class label for PCs decreases with increasing label noise, as shown in Figure 16(b). As expected, the distribution of scores for PCs shifts toward lower values, while the opposite happens for AFPs. At q = 50%, in which 50% of PCs and AFPs have their labels flipped, ExoMiner performs poorly and the average PC and AFP scores get very close to each other, even though the average PC score is still higher than the average AFP score. The reason that ExoMiner can still have higher scores for PCs is that there are still FPs in the NTP category whose labels are not flipped. Using these FPs, the model is able to understand similar cases and perform better than random guessing. Similar to the experiment for training set size, we plot the performance of ExoMiner for validation threshold = 0.99. For KOI label noise as small as 4%, no KOI 31 gets a score >0.99 and there are only about 30 TCEs passing the validation score at KOI label noise 2%. The total number of validated TCEs for q = 0% is around 160.
  • 2.  
    TCE label noise: In this experiment, we flip the labels of q% of PCs to non-PCs and $\tfrac{q}{2} \% $ of AFPs and $\tfrac{q}{2} \% $ of NTPs to PCs in both training and validation sets. This leads to q% label noise in PC class and q% label noise in FP class (AFPs+NTPs). Figure 17(a) shows the performance of ExoMiner for this experiment for varying values of q from 1% to 50%. Because the total number of TCEs in the NTP class is much more than the other two classes, the amount of label noise is effectively larger in this experiment compared to the previous setting. Nonetheless, ExoMiner performs better compared to the previous setting. This is because AFPs are more difficult to classify than NTPs. Similar to the previous setup, the model is highly robust up to 40% label noise, but confidence in its label decreases more rapidly when compared to KOI label noise (Figure 17(b). The reason for this behavior is that a large number of NTPs are labeled as PCs. We plot in Figure 17(c) the effect of this label noise on the validation threshold = 0.99, which shows similar behavior to the KOI label noise.

Figure 16.

Figure 16. Randomly switching labels of q% KOIs in the training set to train the model. The plots show the mean value ± standard deviation of different test set metrics over the five runs for each value of q. The right y-axis in panel (c) shows the total number of TCEs in the test set with scores >0.99.

Standard image High-resolution image
Figure 17.

Figure 17. Randomly switching labels of q% TCEs in the training set to train the model. The plots show the mean value ± standard deviation of different test set metrics over the five runs for each value of q. The right y-axis in panel (c) shows the total number of TCEs in the test set with scores >0.99.

Standard image High-resolution image

Note that the results of label noise experiments can also be justified from the bias/variance trade-off, similar to the training set size experiment.

The above experiments imply that if the size of the training set is not large enough or if there is a considerable amount of label noise in the training set, ExoMiner is not able to classify TCEs with high scores. This is a very important property when we use ExoMiner to validate new exoplanets.

6.2. Analysis of Misclassified Cases

In this section, we examine TCEs that were misclassified in order to better understand the behavior of ExoMiner.

There are a total of 60 CPs in all 10 folds (six PCs on average per fold) misclassified by ExoMiner. Most of these misclassified CPs are cases for which one diagnostic test fails, as we detail below.

  • 1.  
    Centroid test: One subset of CPs are misclassified by ExoMiner solely because there is some degree of centroid shift, based on the DV summary report. Note that the centroid motion test may fail owing to aperture crowding, imperfect background removal, or saturated stars (Twicken et al. 2018) that are unknown to the model at the time of training or classification. Thus, the model does not have access to such information to correctly classify TCEs that have a clear but unreliable centroid shift. Note that Kepler magnitude (Kp ) can be used to recognize the saturated target stars for which the centroid shift test is invalid. An example of a highly saturated target star that hosts misclassified planets is Kepler-444. ExoMiner does not currently use Kp as an input feature; however, Kp will be added in the next version of ExoMiner to improve the model's performance on saturated target stars.
  • 2.  
    Odd & even test: There are CPs that have a considerable odd and even transit depth difference when looking at the absolute depth values of odd and even views. However, to make sure that the transit depth difference between odd and even views is statistically significant and not due to chance, one must use a statistical test that considers the variability of the depth value of different transits in the odd and even views and also the total number of transits for each view; this can be described by a confidence interval for the sample mean of the transit depth for both views. Given that the information related to the number of transits per view and the variability of the depth for each view is not available to ExoMiner, the model is not able to understand the reliability of the depth difference to correctly classify such cases. Examples of misclassified CPs in this category are KICs 8891684.1 and 10028792.2. Adding information such as the number of transits and a measure of the uncertainty for the transit depth estimate of each view to the odd & even branch is the next step for improving the performance of ExoMiner.
  • 3.  
    Secondary test: There are CPs that have a significant secondary transit event. These can be further categorized into three groups:
    • (a)  
      Some of these TCEs are certified as FPs in the CFP list (Bryson et al. 2015). Examples of such TCEs are KICs 11517719.1 and 7532973.1.
    • (b)  
      Some other TCEs in this category, e.g., KICs 10904857.1, 3323887.1, and 11773022.1, have a clear secondary transit event with statistically significant geometric albedo and planet effective temperature statistics. Thus, the information provided in the DV summary report leads to an FP classification for these TCEs by ExoMiner. TCE KIC 11773022.1 (Kepler-51 d) failed the secondary test because the weak secondary detector triggered on transits of another planet in the system. For KIC 10904857.1, the odd and even transit depth comparison and the centroid motion tests are also flagged by DV. Thus, the non-PC classification of this TCE by ExoMiner is well justified. FPWG comment in the CFP list for this TCE reads as follows: "Clearly there is a secondary, but is it due to a planet? 250 ppm. Albedo is right near 1.0 given current stellar parameters. Logg is likely underestimated, so star is likely bigger, pushing it further into possible planet territory."
    • (c)  
      A third group of TCEs in this category, e.g., hot Jupiters such as TCE KIC 4570949.1, have a clear secondary transit event, and their geometric albedo and planet effective temperature comparison statistics are not significant enough to yield an FP disposition. It seems that ExoMiner failed to capture the relationship between the secondary transit events and the albedo and planet effective temperature comparison statistics. It is not clear why ExoMiner is not able to understand such cases. We will study ways to improve this behavior of ExoMiner with respect to the secondary test in our future work.

There are a total of 74 misclassified FPs (7.4 FPs per fold on average) out of 28,318 AFPs+NTPs in the full data set that can be categorized as follows:

  • 1.  
    Ephemeris Match Contamination: There are a total of 16 TCEs with their "Ephemeris Match Indicates Contamination" flag on. This flag means that a transit signal shares the same period and epoch as another object and is the result of flux contamination in the aperture or electronic crosstalk. For 11 out of these 16 TCEs, the only information used to vet them as FPs is the "Ephemeris Match Indicates Contamination" flag, which is not available to ExoMiner. Examples of TCEs in this category are KICs 9849884.1, 10294509.1, and 11251058.1.
  • 2.  
    Centroid test: A good portion of FP cases have some centroid shift that ExoMiner fails to capture. It is not very clear why ExoMiner fails to correctly classify these TCEs. One possible explanation could be that there are CPs with centroid shift (as discussed earlier with regard to misclassified CPs) in the training set that prevent the model from capturing the correct patterns in the data. Examples of TCEs in this category are KICs 6964159.1 and 5992270.1.
  • 3.  
    V-shape transits: For some TCEs, the only useful information available in the DV reports for correctly classifying them is the V-shape form of the transit signal. Given that there are also exoplanets with V-shape transit signatures with high impact parameters, it is not clear how ExoMiner should learn to classify such FP cases. Examples of TCE cases in this category are KICs 6696462.1 and 7661065.1.
  • 4.  
    Odd & even test: There are FPs with statistically significant odd and even transit depth difference that ExoMiner fails to classify correctly. Similar to the centroid motion test, the model is provided with conflicting information during training, i.e., there are CPs with a clear odd and even transit depth difference if one does not take into account the uncertainties for the transit depth estimates. We believe that the model would be able to correctly classify cases like this if we quantified the uncertainty regarding the depth values as additional features (as discussed earlier in this section). An example of a TCE in this category is KIC 8937762.1.

7. Explainability of ExoMiner

In this section, we report the development of an explainability technique for analyzing the classifications made by ExoMiner. In the past few decades, advances in computing power and improvements in ML methods have allowed for the development of highly accurate data-driven models; however, increasing a model's performance often comes at the cost of increasing its complexity. Thus, explainability techniques are of broad and current interest in the field of ML because models that are deployed for any given application should also have human-interpretable outputs to aid in the decision-making process.

In the case of exoplanet detection, a robust model will not only be able to accurately distinguish between PC and non-PC TCEs but can also provide an explanation for its classifications. For example, an FP classification could be complemented with an explanation that the model identified the TCE as an EB based on its transit view of odd & even test. Our goal is to develop a post hoc explainability technique for labeling FP classes without explicitly training the model on these labels. Instead of searching for more granular insights about which specific region in a time series is most important to the model's classifications as was studied by Shallue & Vanderburg (2018), we propose the development of a branch-occlusion sensitivity technique to identify, more broadly, which diagnostic test has the greatest contribution to the model's classifications of each TCE. The multibranch DNN design of ExoMiner, inspired by the DV report, is ideally suited for performing such explainability tests.

In our preliminary study of the explainability of ExoMiner, we devise a simple branch-occlusion explainability test that yields FP flags by identifying the branch of the model that has the most significant contribution to the predicted score of each TCE. To determine which branch of the model makes the greatest contribution to a TCE's score, we mask the inputs of each time series branch by setting the input data to that branch to zero. The DNN generates six new classification scores from sequentially masking each of the six branches (full-orbit flux, transit-view flux, full-orbit centroid, transit-view centroid, transit-view odd & even, transit-view secondary). Note that we excluded stellar parameters and DV diagnostic parameters from this preliminary explainability study because these scalar parameters only provide insight when combined with other branches, a combinatorial effect we did not consider in this paper. For each TCE, we identify the branch whose exclusion from the model changes the model's disposition score the most in a positive or negative way. Thus, the input data to this maximally contributing branch contain the information that most influenced the model to label the TCE as an FP or PC, depending on the class label predicted by the model.

To evaluate the effectiveness of this branch-occlusion explainability framework, we split the data into 80% training, 10% validation, and 10% test set. We train a model using these training and validation sets and applied it on the test set using our explainability test. This provided us with the following insights about ExoMiner:

Classification of PCs: In this analysis, we were interested in understanding how ExoMiner classifies TCEs as PCs. After examining the contribution of different branches, we learned that the full-orbit flux view has a significant positive contribution for PC classifications, meaning that without this flux information most PCs cannot be correctly classified. This makes sense because if all diagnostic tests (the centroid tests, odd & even test, secondary test, and the shape of the transit in transit view of flux data) are unremarkable, their contributions to the disposition score are minimal. We are able to come up with a simple rule for explaining how ExoMiner classifies most of the PCs: if the full-orbit branch is the maximally contributing branch and its contribution toward the PC category is greater than 0.1, ExoMiner likely classifies that TCE as a PC. To be more precise, 90.83% of TCEs whose maximally contributing ExoMiner branch is the full-orbit flux view, with a contribution greater than 0.1, are indeed PCs in the test set (similar statistics in other sets). On the other hand, 99.12% of PCs in the test set satisfy this rule, i.e., for 99.12% of PCs, the full-orbit flux view is the maximally contributing branch with a contribution greater than 0.1 toward the PC class. In other words, this simple rule for explaining how ExoMiner classifies TCEs as PCs has a precision of 0.991 and a recall of 0.908.

Classification of AFPs: In this analysis, we were interested in understanding whether ExoMiner utilizes the correct type of diagnostic test in order to vet an AFP as non-PC. Based on domain knowledge of the diagnostic information contained in each branch's input data, we map the different FP classes to the branches whose inputs would be most crucial for manually diagnosing the respective FP type. Specifically, TCEs with their greatest contribution from the transit-view secondary eclipse branch or the transit-view odd & even flux branch are labeled as EB FPs, and TCEs with their greatest contribution from the full-orbit centroid branch or transit-view centroid branch are labeled as background object FPs. Given that we did not have access to gold standard labels for these subcategories, we utilized the flags, i.e., "Eclipsing Binary High Level Flag" and "Offset High Level Flag," provided in the CFP list (Bryson et al. 2015) as gold standard labels. Using these noisy labels, we found that the FP labels generated by applying the branch-occlusion test to ExoMiner match those provided by the CFP list 78.92% of the time. We believe that when we perform a more careful explainability study of ExoMiner, we will improve our ability to flag the different TCE FP types.

Example TCEs: Table 8 shows a few TCEs for which our explainability analysis provides further insights and demonstrates how useful explainability can be in practice for understanding how the model works and finding ways to improve it. The positive/negative contribution of a branch implies that removing that branch increases/decreases the score of that TCE, which makes that TCE more/less likely to be a PC.

Table 8. Contribution of Different Branches in Classifying Interesting TCEs

TCE KICPeriod (days)Original Label ExoMiner Score B1 B2 B3 B4 B5 B6
6543893.1 a 10.30PC0.498−0.440−0.400−0.006 0.270 0.1240.120
9141355.1 b 469.621PC0.4250.188 0.325 −0.034−0.027 0.101 0.016
11517719.1 c 2.50PC0.353−0.340−0.1680.0010.0340.003 0.245
9996632.1 a 8.44AFP0.017−0.014−0.0120.030 0.971 0.0040.030
8937762.1 b 7.53AFP0.7300.606 0.298 −0.2230.061 0.256 −0.034
6500206.1 c 13.375AFP0.154−0.135−0.023−0.079−0.0470.033 0.838
6666233.2 d 1.0248NTP0.066 0.578 −0.027−0.001−0.0100.0140.023
10186945.2 d 0.794NTP0.095 0.549 −0.019−0.019−0.0080.0070.005
8052016.4 e 33.1842NTP0.004−0.003 0.165 −0.002−0.0000.0750.008
8043714.2 f 3.407NTP0.013−0.006−0.0000.0040.000 0.696 −0.003
9705459.2 f 1.243NTP0.0000.0000.0000.0000.000 0.635 0.000
5446285.2 g 416.653NTP0.003−0.003−0.0030.0000.0000.002 0.803

Notes. Note that the contribution is calculated by subtracting ExoMiner's original score without occluding any branches from the score when a given branch is occluded. Thus, a negative value implies that the branch is contributing toward a PC label, and a positive value implies that the branch is contributing toward a non-PC label. B1: full-orbit flux; B2: transit-view flux; B3: full-orbit centroid; B4: transit-view centroid; B5: transit-view odd–even; B6: transit-view secondary.

a The centroid branch is one of the main contributors to the final score. b The odd and even transit depth difference is one of the main contributors to the final score. c The secondary branch is one of the main contributors to the final score. d The secondary transit of this TCE should refer to the first TCE. It is not corrected because the phase information for the secondary of the first TCE is inaccurate. e The transit view of flux reveals that this is not a real transit. f After removing the primary, the pipeline folded the secondaries onto the data gaps left behind by the primary event. Thus, the period for the second TCE in each case was one-half that of the primary that resulted in an odd and even depth difference for the second TCE. g Even though this is an NTP, there is a clear secondary event helping the model to classify this as a non-PC.

Download table as:  ASCIITypeset image

For the PC category, the full-orbit flux branch is almost always (99.1% of the time) the maximally contributing branch toward a PC classification (negative values). However, when one or more other branches have a positive nonnegligible contribution, that leads to a misclassification. We provide three examples in Table 8 for which other diagnostic test branches, besides the full-orbit branch, are the maximally contributing branch. TCEs KIC 6543893.1, KIC 9141355.1, and KIC 11517719.1 are misclassified owing to centroid shift, odd–even depth difference, and existence of a weak secondary event, respectively. For each of these TCEs, the related branch is correctly used to reduce the score of ExoMiner. Thus, ExoMiner clearly utilized the correct diagnostic test for each classification even though it misclassified those TCEs.

To correctly classify an AFP, however, the contribution of one of the diagnostic tests should be large enough to compensate for the contribution of the full-orbit flux branch. When that fails, ExoMiner is not able to classify the TCE correctly. An example of an AFP that is misclassified by ExoMiner is TCE KIC 8937762.1, for which the contribution of the transit-view odd & even branch is not large enough to compensate for the negative contribution value of the full-orbit and transit-view flux branches. TCEs KIC 9996632.1 and KIC 6500206, for which the transit-view centroid and weak secondary branches, respectively, are the main contributing branches, are correctly classified by ExoMiner.

The most interesting insights provided by this explainability analysis are those related to the NTP class. For two TCEs, i.e., KIC 6666233.2 and KIC 10186945.2, which are labeled as NTP, the branch contributing maximally toward a non-PC classification (i.e., positive value) is the full-orbit view. Both of these TCEs are the secondary eclipse of an EB; thus, their correct label is AFP. Our preprocessing failed to correct the secondary event for these TCEs because the phase generated by the pipeline for the secondary transit of the primary event (i.e., TCEs KIC 6666233.1 and KIC 10186945.1) was inaccurate (Figure 18). Thus, there was no transit in the secondary test for ExoMiner to capture and utilize to generate a correct classification. However, because we do not remove other TCEs when processing our flux data for each TCE, the secondary transit (i.e., transit corresponding to TCEs KIC 6666233.1 and KIC 10186945.1) was visible in the full-orbit view of KIC 6666233.2 and KIC 10186945.2, and the model could use them in order to classify these TCEs correctly.

Figure 18.

Figure 18. The phase value pointing to the secondary event is inaccurate, which prevents us from correcting the secondary of the second TCE of that EB system (see Section 3.6).

Standard image High-resolution image

Another interesting NTP is TCE KIC 8052016.4. The branch contributing the most to ExoMiner's score for this TCE is the transit-view flux branch. A close look at the transit view of this TCE (Figure 19) reveals that the model captured the unusual transit shape of this TCE and labeled it correctly. Other interesting scenarios are provided by TCEs KIC 8043714.2 and KIC 9705459.2, which are actually part of an EB but are labeled as NTP. In both cases, the primary events were completely removed by the pipeline and the second TCE was triggered by the remaining secondaries. However, the pipeline folded the secondaries onto the data gaps left behind by the primary event, so the period for the second TCE in each case was one-half that of the primary. Because of this, the depths of the odd and even views are different for the subsequent TCE since the transits in one view are residuals of the original TCE while transits in the other view represent another transit. The last example is TCE KIC 5446285.2, which is correctly classified using the secondary event branch. The target star associated with this TCE is Kepler-88, which is a Transit Timing Variation (TTV) system and hosts multiple exoplanets. The TTVs for Kepler-88 b are sufficiently large ( ± 0.5 days) that the pipeline was not able to lock on to the nominal orbital period. At the detected period, other transits of the planet are actually identified as the secondary for TCE KIC 5446285.2, leading to its correct classification.

Figure 19.

Figure 19. KIC TCE 8052016.4 is an NTP and is correctly classified based on its unusual transit-view flux shape. Note that the blue circles are the real flux data, and the red solid line is the transit fit generated by the pipeline.

Standard image High-resolution image

The results of this preliminary study indicate that identifying the maximally contributing branch to each TCE's predicted score holds a high degree of explainability power for an extremely simple explainability technique. To develop higher-fidelity explainability tests that can be integrated into the ExoMiner classification pipeline, several improvements to this technique will be explored in future work. The model architecture diagram shown in Figure 4 shows that the stellar parameter scalars branch and the DV diagnostic scalars branch exist at a different hierarchical level than the convolutional branches. Rather than setting the inputs of each convolutional branch to zero, if we instead set each branch's contribution to the FC block to zero, the FP flags we obtain may be more accurate. In particular, this is important for the transit-view odd & even flux branch, where an input of all zeros has a different meaning owing to the subtraction layer. Achieving more accurate flags of EBs through the transit-view odd & even flux branch may be possible by implementing noise sensitivity mapping, as discussed in Section 3.3. More efficient gradient-based methods like Grad-CAM (Selvaraju et al. 2019) will also be explored as components of a more robust explainability testing for ExoMiner. Refining these explainability tests not only will provide more information about the model classifications but also will provide insight into how the model can be improved in future iterations. As ExoMiner is applied to TESS data, these explainability tests will be helpful in providing additional confidence in the dispositions and pointing out the limitations of the model for classifying new signals.

8. ExoMiner for Exoplanet Validation

8.1. Classification with Rejection

For applications where the cost of misclassification is high (e.g., validating new exoplanets), we need ways to reduce the error rate of the classifier. This is a relatively old concept in statistics (Chow 1970) and a standard one in ML (Nadeem et al.2010; Cortes et al. 2016). It is called classification with rejection. One classification with rejection approach widely used in ML is based on failing to classify examples when the model is not very confident of its label. The idea is that if the classifier cannot classify a case with high confidence, then it is not sure about its label. When the classifier predicts the label of an example with a very high score, it means that the model has a good understanding of the feature space in which that example lies (think about it in terms of p(yX)). As we reported in Section 6.1, by reducing the size of the training set or adding label noise, ExoMiner becomes less confident about its disposition. This verifies that by setting a confidence-based rejection option, we control the risk of misclassification. To study confidence-based rejection, we plot in Figure 20 the precision of different classifiers when different threshold values for rejection are used to reduce the rate of FPs. As can be seen, all methods perform reasonably well for large thresholds. Note that the precision of ExoMiner at T = 0.99 is around 99.7%, much higher than the required precision value of 99% for that threshold. The precision of ExoMiner even for T = 0.9 is around 99%.

Figure 20.

Figure 20. Precision of different classifiers when a threshold T is used to reduce the FP rate.

Standard image High-resolution image

It is important that one uses a model-specific threshold T for classification with rejection because the precision does not necessarily increase with larger thresholds, e.g., check AstroNet and ExoNet for T > 0.97. However, given that T = 0.99 has been accepted to use for validating new exoplanets (Morton et al. 2016; Armstrong et al. 2021) and ExoMiner performs very well for that threshold, in the remainder of this paper we focus on using this threshold for different analyses (in addition to the standard classification threshold of 0.5).

8.2. Classifier Calibration

In order to make a decision based on the output of a classifier (e.g., to validate new exoplanets), we should be able to interpret the scores generated by that classifier. If a classifier generates probability scores, we can estimate the error rate for different probability scores, i.e., for a given probability score value of s, the classifier correctly classifies 100 × s% of examples. Such classifiers are called calibrated. Most discriminative models for machine classification, however, do not generate calibrated scores even when their generated scores are in the [0, 1] range. Such classifiers require calibration before their scores can be used for decision-making, especially when a decision must be taken for any scenario (e.g., consider a self-driving car making a decision to stop or keep moving). Because of this, Armstrong et al. (2021) calibrated their classifiers prior to validating new exoplanets using isotonic regression (Niculescu-Mizil & Caruana 2005), one of the two most common calibration methods (Platt 1999, pp. 61–74; Niculescu-Mizil & Caruana 2005).

To study the calibration of a classifier, people use calibration plots (reliability curves) that show the error rate of the classifier for different probability score values. A classifier is perfectly calibrated if the error rate is equal to the score it generates. Figure 21 shows the calibration plots of uncalibrated ExoMiner, vespa 1-fpp, and two calibrated classifiers 32 from Armstrong et al. (2021), GPC and RFC, for the full data set (using CV). Note that vespa fpp is only available for KOIs, so its reliability plot only uses a subset of the data used for the other three classifiers. The results and conclusion, with regard to the level of calibration, do not change if we use only KOIs for all models. As can be seen, ExoMiner appears well calibrated already. To measure the level of calibration of these classifiers, we use Expected Calibration Error (ECE) and Maximum Calibration Error (MCE) (Guo et al. 2017). ECE and MCE are, respectively, the weighted average and maximum of the absolute difference between the classifier's average accuracy and score for each score bin. Table 9 shows the level of calibration for different classifiers used for validation of new exoplanets. As can be seen, the out-of-box ExoMiner is well calibrated and can be used without further calibration. It has been previously shown that the scores generated by certain classifiers such as NNs are well calibrated (Niculescu-Mizil & Caruana 2005). Nonetheless, we tried both isotonic regression (Niculescu-Mizil & Caruana 2005) and Platt scaling (Platt 1999, pp. 61–74) to further improve the calibration of ExoMiner. These calibration techniques did not work well because the bins in the middle of the reliability plots do not have enough points per fold, leading to poor post-processing calibration. To overcome this problem, Armstrong et al. (2021) generated synthetic examples by interpolating between members of each class. Generating such synthetic data sets for diagnostic test time series is not easy and does not guarantee better results. As such, and because the level of calibration of ExoMiner is already high, we use the scores generated by the classifier directly for planet validation.

Figure 21.

Figure 21. Reliability plots of ExoMiner, GPC, RFC, and vespa.

Standard image High-resolution image

Table 9. Level of Calibration for Different Methods

ClassifierECEMCE
vespa (1-fpp)0.2460.68
GPC 0.0130.249
RFC 0.0090.145
ExoMiner 0.0010.067

Note. Lower values are better.

Download table as:  ASCIITypeset image

We would like to emphasize that the calibration of methods for scores >0.99 is of particular interest for validating new exoplanets. As can be seen in the zoomed-in area of Figure 21, only ExoMiner is an underestimator classifier. A classifier is an underestimator when the fraction of positive examples for a given probability score is higher than the score assignment, i.e., the mean predicted value assigned by such a classifier is lower than its precision. This is a useful property for validating exoplanets because a classifier that is an underestimator in the high score range (close to 1.0) makes fewer errors when validating new exoplanets. GPC, RFC, and vespa are all overestimator classifiers in that range, meaning that there is more than 1% error when validating exoplanets using a threshold value of 0.99. Note that these results are consistent with Figure 20 and the classification with rejection discussion.

8.3. Utilizing Prior Information

As we mentioned in Section 6.2, ExoMiner does not perform well under some specific data scenarios. In order to address that, we utilize prior probabilities for different scenarios as used in Morton & Johnson (2011), Morton (2012), and Armstrong et al. (2021). More specifically, Armstrong et al. (2021) utilized prior probabilities for different scenarios (EB, hierachical EB, BEB, hierarchical transiting planet, background transiting planet, known other stellar source, nonastrophysical/systematic, planet) and combined them with their calibrated classifier scores to improve the performance of the classifier. We take the same approach to combine the prior probability calculated in Armstrong et al. (2021) with the scores of ExoMiner, with a somewhat different justification, as follows. There are two sets of probability scores computed from two different sources of information: (1) p(y = 1∣M, X, x*) that learner M (in our case ExoMiner) generates for signal x* using the information from the training set X; and (2) P(y = 1∣I), where I represents the target star's individual characteristics (stellar type, local environment, spacecraft status at the transit time, etc.). In order to combine these two scores, assuming that (1) I and (2) x*, M, and X are conditionally independent (i.e., p(x*, M, X, Iy) = p(x*, M, Xy) × p(Iy)), we can use Bayes's theorem to get the following equation, which is an identical equation to the one reported in Armstrong et al. (2021):

Equation (10)

The results of applying this technique to different classifiers are reported in Table 10, with the performance metrics reported in Table 11. As can be seen, the performance of all classifiers improved by applying the prior information I. However, the use of this prior information is more helpful to those classifiers, e.g., AstroNet and ExoNet, that do not use the comprehensive diagnostic test data provided by the Kepler pipeline. The performance of the classifiers that use a more comprehensive set of diagnostic tests, such as ExoMiner and GPC, is not improved as much. Interestingly, the application of the prior information I to these classifiers often makes the classifiers more confident about the labels. This is because if either p(y = 1∣M, X, x*) or p(y = 1∣I) is close to 1, p(y = 1∣M, X, x*, I) becomes close to 1. 33 Thus, there will be more TCEs with scores >0.99. This is why the recall values for all classifiers at threshold 0.99 increase. ExoMiner is a conservative classifier, and applying prior information I leads to a large recall improvement at the threshold 0.99.

Table 10. Scores of Different Classifiers Trained in This Work after Applying Priors

ColumnDescription
target_idKIC ID
tce_plnt_numTCE planet number
tce_periodTCE period
mesTCE MES
original_labelNTP, AFP, or PC
binary_label0 for NTP or AFP, 1 for PC
astronet_prior AstroNet score with priors
exonet_prior ExoNet score with priors
exominer_prior ExoMiner score with priors
exominer_tce_prior ExoMiner-TCE score with priors

Note. This table describes the available columns. The full table is available online.

Only a portion of this table is shown here to demonstrate its form and content. A machine-readable version of the full table is available.

Download table as:  DataTypeset image

Table 11. The Effect of Prior on the Precision & Recall Values of Different Models

 Threshold = 0.5Threshold = 0.99
Model/MetricOriginalUsing PriorsOriginalUsing Priors
AstroNet 0.861 & 0.8850.920 & 0.9200.958 & 0.0101.000 & 0.102
ExoNet 0.925 & 0.8640.933 & 0.9020.983 & 0.0250.993 & 0.123
GPC 0.921 & 0.9640.910 & 0.9770.991 & 0.6050.993 & 0.663
RFC 0.929 & 0.9550.942 & 0.9640.979 & 0.7240.986 & 0.791
ExoMiner 0.968 & 0.9740.971 & 0.9780.997 & 0.6560.997 & 0.752
ExoMiner-TCE0.965 & 0.9770.968 & 0.9830.997 & 0.7750.996 & 0.823

Note. The boost in performance due to prior diminishes as the model uses a more complete set of diagnostic tests.

Download table as:  ASCIITypeset image

We also checked the performance of ExoMiner updated scores using the scenario priors for changes in training set size and label noise, as we have done in Section 6.1. Even though we are not providing the plots here for conciseness, using the scenario priors improves the performance of ExoMiner, more significantly when the size of the training set is small or there is more label noise. Given that using the scenario priors improves the performance of ExoMiner, we use these updated ExoMiner scores for the validation of new exoplanets.

8.4. Dependence on Candidate Parameters

In this section, we study ExoMiner dependence on candidate parameters using the Cumulative KOI catalog. Given that we use the updated ExoMiner scores using scenario priors, as described in Section 8.3, for validation of new exoplanets, we will focus here on these updated scores. Table 12 shows ExoMiner scores for different values of KOI radius and multiplicity. As can be seen, ExoMiner generally assigns higher scores to KOIs in multiple systems compared to single systems even though it does not have access to the information related to the number of detected transits for each target star. Moreover, ExoMiner assigned lower scores to larger KOIs.

Table 12.  ExoMiner Scores by KOI Radius and Multiplicity

SelectionCountMeanMedian
All64640.3790.002
KOI in multiple systems13130.8570.996
KOI singles51510.2570.000
Rp < 2R 20440.4020.000
2R < Rp < 4R 19020.6610.993
4R < Rp < 10R 6820.3510.001
10R < Rp < 15R 2630.3360.011
Rp > 15R 14870.0280.000

Download table as:  ASCIITypeset image

To study the performance with regard to candidate parameters above the validation threshold, we plotted the precision and recall of ExoMiner for scores >0.99 for different pairs of candidate parameters in Figure 22. ExoMiner is reliable in different areas of candidate parameters for planet validation. At the validation threshold of 0.99, ExoMiner is very conservative in labeling KOIs as planets, so it has high precision and low recall. Also, note that there are regions of parameter space for which there are not enough KOIs. Because there are not enough examples in those area in the training set, the model is not confident about the scores even if it assigns the cases correctly to the class. This shows a similar effect to the training set size experiment we have done in Section 6.1.

Figure 22.

Figure 22. Precision and recall for different parameters when threshold 0.99 is used for validation.

Standard image High-resolution image

In the low-MES region (MES < 10.5), there is only one FP, KIC TCE 10187017.6, out of a total of 13 TCEs that have scores >0.99. Interestingly, this is the sixth TCE of a system for which the first five TCEs are all confirmed exoplanets. Thus, we believe this could be planet number six with a period that fits between planet e and planet f of KOI-82. This object is marked as a centroid offset FP in the certified FP table, but the TCERT report does not show clear evidence for this offset, and upon review this object's original vetter for the FP working group agrees that the background FP classification is an error due to not accounting for saturation (Steve Bryson, personal communication).

Figure 23 shows the distribution of scores for planet radius, MES, and galactic latitude of KOI versus KOI orbital period. As can be seen, short- and long-period KOIs, KOIs in the low-MES region, and KOIs at lower galactic latitudes are all assigned low scores.

Figure 23.

Figure 23. Mean score of KOIs for different parameters. The number in each bin shows the total number of KOIs in that bin. There are no KOIs in white bins.

Standard image High-resolution image

9. Planet Validation

9.1. Setup

Similar to existing planet validation approaches, we use the rejection threshold of >0.99 to validate new exoplanets. To make the validation process more reliable, we need to identify where ExoMiner does not perform well in order to design vetoing criteria to improve the validation process even further.

As we detailed in Section 6.2, ExoMiner classifies some FPs into the PC category for different reasons: lack of enough information such as that given by the contamination flag, or poor interpretation of diagnostic tests such as the centroid test. As we discussed in Section 8.3, utilizing scenario priors addressed some of these misclassifications and improved the performance. Nonetheless, there are still cases that are not resolved by the application of scenario priors, even though ExoMiner is making less than one error out of every 100 TCEs that exceed the rejection threshold of 0.99. To be precise, ExoMiner classifies 1727 TCEs with score >0.99, of which there are five FPs, 34 leading to a precision of 0.997, better than the required precision of 0.99 for the threshold of 0.99. To further improve the precision of ExoMiner in validating new exoplanets, we ensure that the KOI is dispositioned as a "candidate" in the Cumulative KOI catalog (similar to Morton et al. 2016) and that no disposition flag is on in the Cumulative KOI catalog. These vetoes remove those five FPs from the validated list. Regarding the low-MES region, even though the model performs well in this region 35 (see Figure 22(c)), we take the more conservative approach and do not validate any KOI with MES < 10.5.

Unlike Armstrong et al. (2021), we do not recommend using vespa fpp scores to veto the validated exoplanets for multiple reasons:

  • 1.  
    vespa is poorly calibrated, as shown in Figure 21. As an overestimator classifier, its scores cannot be used for validation because the error for any rejection threshold is higher than that threshold, as seen in Figures 21 and 24. As an example, vespa gives scores below 0.01 to 2409 TCEs, of which 446 TCEs are FPs, leading to a precision of 0.815, which is much lower than the required value of 0.99 for the applied threshold. As we mentioned in Section 9, Morton et al. (2016) used other criteria in order to improve the vetting. This led to a highly accurate approach for exoplanet validation in terms of precision, but very low recall.
  • 2.  
    Inspired by Armstrong et al. (2021), we use the scenario priors I as complementary sources of information. These scenario priors are also used in calculating vespa fpp scores. Thus, we are using part of the vespa model. By only using scenario priors, we ignore the likelihood used to calculate vespa fpp. We discussed in Section 2.2 the problems with the generative approach and why a discriminative approach is preferred for classification problems.
  • 3.  
    The vespa model is highly dependent on the accuracy of stellar parameters. Given the uncertainty about the values of stellar parameters, vespa fpp is subject to continuous change.
  • 4.  
    As discussed in Section 5.4, when there is a disagreement between ExoMiner and vespa scores for validating exoplanets at a threshold of 0.99, ExoMiner is almost always correct. Thus, the advantage of vetoing based on vespa fpp is not clear.

Figure 24.

Figure 24. Distribution of CPs and CFPs as a function of vespa fpp score. There are 596 CFPs out of a total of 2719 CPs+CFPs with vespa fpp ≤ 0.05. This implies a misclassification rate of 21.9% if we use vespa fpp ≤ 0.05 to validate new exoplanets. Not shown in this picture: there are 403 CFPs with vespa fpp ≤ 0.01 out of 2372 CPs+CFPs (a misclassification rate of 17.0%). These results demonstrate that vespa fpp is not effective for screening against AFPs at the 0.01 confidence level in the absence of other constraints (see Section 9).

Standard image High-resolution image

We also do not suggest using an outlier detection test to veto validated exoplanets for two main reasons:

  • 1.  
    Unsupervised outlier detection is a subjective task where the results vary from one algorithm to another or by changing the hyperparameters of the same algorithm, i.e., one can always come up with an outlier detection algorithm such that some of these validated exoplanets are outliers.
  • 2.  
    By using the rejection threshold, we already identified TCEs for which the model has not seen similar cases, as discussed in Section 8.1.

Nonetheless, as reported in Section 9.2, all validated exoplanets in this work pass the outlier tests of Armstrong et al. (2021).

9.2. Validated Exoplanets Using ExoMiner

To validate new exoplanets, we apply 10 ExoMiner models trained using 10-fold CV to those KOIs that are not part of our working set, i.e., KOIs that are not CPs, CFPs, or CFAs. We will designate this subset by unlabeled KOI set. There are a total of 1922 such KOIs in the Cumulative KOI catalog from the Q1–Q17 DR25 run. We use the average score of 10 models to classify these KOIs. Table 13 shows the scores provided by ExoMiner before and after applying priors for the unlabeled KOI set. Ephemerides and other relevant parameters are also presented. Given that a majority of TCEs in the training set are NTPs, and the KOIs from which we are validating do not include any NTPs, the rate of FPs in the training set is much higher than that of the unlabeled KOI set. Thus, the probability scores generated by ExoMiner are more conservative in labeling these KOIs as PC and validating them. ExoMiner labels 1245 36 KOIs as PC (scores > 0.5) out of 1922 KOIs. Other existing classifiers also provide similar results. For example, for posterior GPC (Armstrong et al. 2021), this number is 1126 KOIs.

Table 13. Original ExoMiner Score and ExoMiner Score after Applying Priors for the Unlabeled KOIs

ColumnDescription
target_idKIC ID
tce_plnt_numTCE planet number
tce_periodTCE period
mesTCE MES
exomineroriginal ExoMiner score
exominer_prior ExoMiner score with priors

Note. This table describes the available columns. The full table is available online.

Only a portion of this table is shown here to demonstrate its form and content. A machine-readable version of the full table is available.

Download table as:  DataTypeset image

Out of these 1245 positively labeled KOIs by ExoMiner, eight KOIs are dispositioned as FP or have an FP flag on in the Cumulative KOI catalog. Given that most of these positively labeled KOIs are already dispositioned as "candidate" in the Cumulative KOI catalog, this number should not be surprising. One possible justification for why there are so many candidates left in the unlabeled set could be because vetting FPs is easier than confirming PCs. A TCE is certified as an FP if one of the diagnostic tests indicates an FP. However, it is not enough to confirm an exoplanet if none of the diagnostic metrics are inconsistent with a true transit signal; confirming new exoplanets requires follow-up studies or statistical/ML validation.

Out of 1922 unlabeled KOIs, ExoMiner labels 368 KOIs as PC with validation score >0.99, of which 46 are among the previously validated exoplanets by Armstrong et al. (2021) and one among the previously validated ones by Jontof-Hutter et al. (2021), leaving 321 new ones. Interestingly, out of a total of 943 unlabeled KOIs with MES < 10.5, ExoMiner gives a score >0.99 to 20 KOIs, even though 839 of them have a "candidate" disposition in the Cumulative KOI catalog. This implies that ExoMiner is more conservative in the low-MES region, as discussed in Section 8.4. Removing these 20 KOIs with MES < 10.5, we validate 301 new exoplanets. Out of these 301 newly validated exoplanets, 42 KOIs are in multiplanet systems. We provide the list of newly validated exoplanets in Table 14. None of these newly validated exoplanets are found to be outliers based on the two algorithms in Armstrong et al. (2021). We also report those KOIs with ExoMiner scores >0.99 in the low-MES region in Table 15, even though we are not validating them. Nine out of 20 such KOIs are in multiplanet systems.

Table 14. List of Newly Validated Exoplanets with ExoMiner Score >0.99 and MES >10.5

NumberTCE KICKOI NamePeriod (days)Radius (Re)MES ExoMiner Prior vespa ScoreKepler Name
111852982.1K00247.0113.81502.283741.49000.99890.0076Kepler-1712b
27700622.1K00315.0135.58102.424873.59000.99810.0000Kepler-1716b
37135852.1K00875.014.22103.8940134.50000.99810.0009Kepler-1729b
46803202.1K00177.0121.06081.923043.05000.99590.0005Kepler-1711b
52444412.1K00103.0114.91113.1974108.00000.99800.0001Kepler-1710b
67941200.1K00092.0165.70463.124281.89000.99770.2000Kepler-1709b
78087812.1K04343.0127.21081.239213.39000.99310.0005Kepler-1952b
87749773.1K02848.0113.78731.486716.21000.99830.0006Kepler-1882b
911599038.1K01437.017.01731.824212.55000.99810.0001Kepler-1751b
1010587105.3K00339.0335.86622.795018.22000.99810.0006Kepler-529d
113645438.2K04385.017.29801.942112.20000.99880.0004Kepler-1600c
124665571.2K02393.014.60301.459414.98000.99800.0002Kepler-1834b
137364176.1K00373.01135.18802.568562.99000.99800.8100Kepler-1718b
149823487.1K01489.0116.00462.226827.32000.99860.0000Kepler-1753b
155374854.1K00645.0223.78322.296028.80000.99570.0035Kepler-633 c
168410415.1K02291.0144.29922.273519.37000.99800.1200Kepler-1817b
173937519.2K00221.025.89661.219412.11000.99860.0017Kepler-495 c
1810134152.1K02056.0139.31372.319025.30000.99840.0280Kepler-1787b
199150827.1K01408.0114.53411.613025.62000.99860.0010Kepler-1749b
203641726.1K00804.019.02933.173247.37000.99750.0000Kepler-1727b
2112121570.1K02290.0191.50101.967918.66000.99250.0003Kepler-1816b
2211046025.1K01646.013.72020.903814.02000.99720.0007Kepler-1759b
239837661.3K02715.035.72092.628415.06000.99740.1800Kepler-1321d
2411821363.1K01494.018.19592.440017.20000.99800.0009Kepler-1754b
2510074466.1K01917.0135.01192.611627.80000.99770.0730Kepler-1776b
2610386922.2K00289.0126.62954.209676.47000.99740.0000Kepler-511 c
277584650.1K02631.0144.99762.263115.96000.99680.0008Kepler-1863b
287670943.1K00269.0118.01161.716738.88000.99620.0005Kepler-1713b
295601258.1K02191.018.84791.545018.35000.99790.0650Kepler-1805b
307826659.1K02686.01211.03403.149559.00000.99590.0001Kepler-1868b
317386827.1K01704.0110.41892.495136.84000.99870.0250Kepler-1765b
329307509.1K02096.019.76301.794321.39000.99830.0260Kepler-1794b
336183511.1K02542.010.72731.216214.74000.99670.0054Kepler-1855b
347877978.1K02760.0156.57322.211118.77000.99180.0029Kepler-1874b
355436013.1K02471.0125.88601.489113.98000.99570.0053Kepler-1841b
3611297236.1K01857.0188.64404.055438.36000.99110.0000Kepler-1771b
3712469800.1K02543.011.30201.734519.04000.99460.0044Kepler-1856b
389836959.1K01715.01105.01502.279025.76000.99700.0045Kepler-1766b
393940418.1K00810.014.78302.955057.64000.99860.0370Kepler-1728b
405773121.1K04002.010.52421.514818.35000.99450.0160Kepler-1932b
415020319.1K00635.0116.71992.812551.87000.99780.0170Kepler-1725b
425978361.1K00558.019.17862.867244.10000.99840.0280Kepler-1722b
433109930.1K01112.0137.81033.876628.38000.99790.0130Kepler-1736b
449475552.2K02694.026.56611.844019.83000.99870.3000Kepler-1315c
4510793172.2K02871.025.36370.977011.98000.99690.0190Kepler-1693c
.........
.........
.........

Note. The full list is available online.

Only a portion of this table is shown here to demonstrate its form and content. A machine-readable version of the full table is available.

Download table as:  DataTypeset image

Table 15. List of KOIs with ExoMiner Score >0.99 and MES <10.5

NumberTCE KICKOI NamePeriod (days)Radius (Re)MES ExoMiner Score vespa Score
16871071.4K02220.047.66481.62889.75900.99690.0005
210793576.1K03451.0110.90761.69389.29500.99710.0005
312204137.1K04590.019.54062.08329.92200.99580.0850
49896018.3K02579.0310.30151.62639.27200.99270.0002
57289317.2K02450.027.19291.08979.67000.99180.0008
69011877.1K05597.017.77710.964610.17000.99400.0003
79457948.1K04694.0119.25271.68849.53300.99610.0015
84919550.1K04766.015.56681.131710.31000.99800.0380
94164922.2K03864.0218.25741.038910.21000.99620.0007
102860656.1K04857.014.11501.85549.07100.99440.0001
119489524.4K02029.044.78850.74289.91800.99660.0110
127098255.1K04762.0111.56731.40189.78800.99190.0790
137692093.1K03337.0111.16621.563110.16000.99330.0012
147100673.5K04032.057.23520.96629.03700.99030.2000
156131130.1K03307.0110.50711.40879.46100.99180.0250
165542466.3K01590.034.74671.51759.56300.99270.0071
175903749.1K03029.0118.97632.488010.21000.99310.0150
186144039.1K04602.012.85860.892210.20000.99160.0170
199002538.2K03196.026.88300.775710.11000.99160.4700
208874090.2K01404.0218.90631.756110.13000.99410.0014

Download table as:  ASCIITypeset image

Figure 25 displays a scatter plot of the planetary radius versus orbital period for the previously confirmed and validated exoplanets from the Kepler, K2, and TESS missions, along with the 301 new Kepler exoplanets validated by ExoMiner. The distribution of the ExoMiner-validated exoplanets is consistent with that of the Kepler sample, with periods ranging from ∼0.5 to over 280 days, and with planetary radii as small as 0.6 R to as large as 9.5 R. Figure 26 shows a scatter plot of the planetary radius versus the energy received by each planet in the Kepler, K2, TESS, and ExoMiner samples. As with the previous parametric plot, the distribution of the radius versus received energy of the ExoMiner sample generally follows that of the Kepler sample.

Figure 25.

Figure 25. Planet radius vs. orbital period for confirmed transiting planets and ExoMiner-validated planets with a score greater than 0.99. The CPs are indicated by the discovery source, with Kepler indicated by plus signs, K2 by diamonds, and TESS by asterisks. ExoMiner-validated planets are indicated by a magenta square. For reference, Earth is indicated by a black circle.

Standard image High-resolution image
Figure 26.

Figure 26. Planet radius vs. energy received by a planet for confirmed transiting planets and ExoMiner-validated planets with a score greater than 0.99. Symbols are the same as in Figure 25.

Standard image High-resolution image

10. Transfer Learning to TESS Data

Training a machine classifier to vet TESS data is challenging because there are a limited number of TCEs with gold standard dispositions available for the TESS mission. Existing works try to overcome the challenge of limited data by using either synthetic data (Osborn et al. 2020) or self-dispositions based on high-level heuristics (Yu et al. 2019) in order to train a machine classifier to vet TESS TCEs. However, it is not easy to ensure that simulated signals, including transiting planets, astrophysical signals, and systematic noise and artifacts, are sufficiently similar to the flight data; when used for training, these synthetic data may harm the performance of the model. The self-disposition approach is also noisy and could lead to an ineffective model that is not able to vet TCEs accurately. Instead, we argue that if the architecture of a machine classifier is general enough for the classification of TCEs and the data are preprocessed appropriately, a trained classifier should be able to vet TCEs from different missions comparably well. The processes for manually vetting transit signals from TESS and Kepler are very similar, i.e., independent of the source of the TCEs, vetting a TCE requires evaluating the same diagnostic plots and metrics. If a TCE passes these tests, it is vetted as a PC no matter where the data come from. Thus, we can transfer (Ng 2016) a model learned from Kepler data to vet TESS TCEs.

We conducted a preliminary analysis of ExoMiner's performance at classifying transit signals obtained from the ongoing TESS mission. The Science Operations Center (SOC) pipeline, which was created to process the data from the Kepler mission, was ported and tuned for TESS data in the Science Processing Operations Center (SPOC; Jenkins et al. 2016). The SPOC pipeline is currently used to generate TCEs that can be further designated as TOIs (Guerrero et al. 2021). Even though the population of target stars observed by TESS 37 and the characteristics of the systematic noise are different, the nature of the signals captured by the two satellites is similar. In this first analysis, the same preprocessing steps were applied to the TESS data to generate the input data to feed into the model trained exclusively on data from the Kepler mission.

One specific feature of TESS data is that the pipeline is run for each individual sector, but there are also multisector runs that are conducted using data from multiple sectors. The single-sector runs produce TCEs in the approximately 28-day-long observation window, so the detected TCEs have orbital periods shorter than this time interval. 38 Since there is overlap between sectors, especially at the ecliptic poles, multisector runs are able to produce TCEs with longer orbital periods when transits are observed in multiple sectors. Currently, the data for the first 36 sectors are publicly available, 39 as well as the TCEs detected by the SPOC pipeline. 40 A total of approximately 42,500 TCEs were detected, with an average of ∼1200 TCEs per observed sector. Spanning from Sector 1 to Sector 26, nine multisector runs were conducted, yielding a total of ∼29,500 TCEs. From these runs, experts vetted the detected TCEs and combined data from multiple sources to arrive at a catalog of TOIs 41 (Guerrero et al. 2021).

The TOI catalog made publicly available on 2020 December 4, which covered Sectors 1–30, was used to generate the data on which we evaluated ExoMiner's performance. After preparing the TOI catalog (removing TOIs with missing ephemerides or observation sectors), 1167 SPOC TOIs were preprocessed using our pipeline to generate the data inputs to ExoMiner. Given that the TOI catalog does not include the phase of the secondary event, ephemeris matching between the TOIs and the detected TCEs was performed such that our preprocessing pipeline could generate the transit-view secondary eclipse time series. In this preliminary analysis of TESS data, we trained a new model on Kepler Q1–Q17 DR25 TCE data using only the time series and stellar parameters as features. We call this model ExoMiner-Basic. The classifications produced by the model were compared against the TESS Follow-up Observing Program Working Group (TFOPWG; Collins et al. 2018) dispositions, which are based on additional observations and are the current gold standard dispositions. Our analysis focuses on the TOIs dispositioned as known planets (KPs), CPs, FPs, and FAs. We did not include the TOIs with PC disposition because this disposition is not conclusive.

Table 16 shows the confusion matrix of applying ExoMiner-Basic to TESS data for the standard classification threshold of 0.5. ExoMiner-Basic performs reasonably well on the TESS data, resulting in a precision value of 0.88 and a recall value of 0.73. We expect the performance of the model to increase if the model has access to the DV diagnostics that the full ExoMiner model uses for classifying Kepler data. However, an analysis of the misclassified KP and CP TOIs gave insight into some issues that must be addressed in the future, as shown in Figures 27, 28, and 29.

Figure 27.

Figure 27. Odd and even transit views generated for TOI 822.01. The blue circles show the average bin value, and the black circles show the cadences in the phase-folded time series.

Standard image High-resolution image

Table 16. Confusion Matrix of TESS Classification Using a Basic Version of ExoMiner

ClassPCNon-PCTotal
KPs18521206
CPs325789
FPs276794
FAs21618
Total246161407

Download table as:  ASCIITypeset image

There are cases of TOIs with transits located inside the downlink window for TESS. Data are downlinked to Earth at the perigee, which happens approximately every 13.7 days and lasts about a day. During that time, new data are not collected, which prevents the observation of some transits. For example, TOI 822.01 (WASP-132b) was observed in Sector 11 and has an orbital period of approximately 7.13 days, but the corresponding TCE (TIC 127530399.1) is reported as having double the period. Given that one transit was located in the downlink window, TPS did not fold the missing transit onto the observed transits, incorrectly determining the period of the TOI, which is known from previous observations. Since we are using the true orbital period from the TOI catalog, the preprocessing pipeline generates one of the odd and even transit views to be flat (Figure 27), rendering this diagnostic test unreliable and leading the model to incorrectly classify this TOI as an FP, even though it is a KP. As we discussed earlier, this problem can be addressed by using additional input features that take into account the number of observed transits and their variability for the odd and even views.

Another issue with TOIs that only have a small number of observed transits is that averaging the phase-folded time series is not enough to filter the stochastic noise in the time series; this becomes even more critical when S/N is low. We expect to see more of these cases in TESS data considering that for significant areas of the sky the total observation time for many TCEs is much shorter than for Kepler, which had a fixed field of view that was observed for 4 yr. For example, TOI 1774.01 was observed in Sector 21 and only has two observed transits, as shown in Figure 28, which creates flux views with significant levels of noise.

Figure 28.

Figure 28. Full-orbit flux time series for TCE TIC 4897275.1 (associated with TOI 1774.01) obtained from the DV summary generated by the SPOC pipeline for Sector 21.

Standard image High-resolution image

A further issue arises when incorrect epoch values are reported in the TOI catalog because it causes the transit to be off-centered in the view. Since ExoMiner expects the transit to be located in the center of the view, the model classifies such TOIs as FPs. For example, the transit for TOI 185.01 is shifted to the right in our view, as shown in Figure 29. A possible approach that can be used to make the model less sensitive to the location of the transit source in the views is to provide the model examples in which the transit is slightly shifted from the center of the view during training. Using such a data augmentation technique may make the model more robust to inaccuracies in the orbital period and epoch estimates.

Figure 29.

Figure 29. Inaccurate epoch information leads to off-centered transit views.

Standard image High-resolution image

11. Discussion

We introduced ExoMiner, a new deep learning classifier that outperforms existing machine classifiers in terms of various classification and ranking metrics. ExoMiner is a robust and explainable deep learning model that allows us to reliably and accurately build new catalogs of PCs. The key to ExoMiner's performance is its architecture that mimics the process by which domain experts vet transit signals by examining multiple types of diagnostic tests in forms of scalar values and time series data. This architecture also enabled us to design a preliminary branch-occlusion explainability framework that provides interpretability in terms of which diagnostic tests ExoMiner utilizes in order to vet a signal. Moreover, since the general concept behind vetting transit signals is the same for both Kepler and TESS data, and ExoMiner utilizes the same diagnostic metrics as expert vetters do, we expect an adapted version of this model to perform well on TESS data. Our preliminary results on TESS data verify this hypothesis. Using ExoMiner, we also demonstrate that there are hundreds of new exoplanets hidden in the 1922 KOIs that require further follow-up analysis. Out of these, 301 new exoplanets are validated with confidence using ExoMiner.

11.1. Caveats

There are caveats related to the ML approach taken in this work, which include the following:

  • 1.  
    As we reported in Section 6.2, ExoMiner sometimes fails to adequately utilize diagnostic tests, e.g., centroid test, or does not have the data required for some FP scenarios, e.g., flux contamination. This leads to misclassification of transit signals. To provide the extent to which this causes problems, some statistics are in order. Out of a total of 8053 KOIs, there are 1779 KOIs that have the centroid test flag on. ExoMiner misclassifies 23 (0.29%) of these, of which only one had a score >0.99 (0.01%). After applying the priors, the number of misclassified cases reduces to 13 (0.16%), out of which two (0.025%) have scores >0.99. Out of 1087 KOIs that have flux contamination, ExoMiner misclassifies 16 (0.2%) and gives a score >0.99 to only one (0.01%). After applying the priors, only one KOI with flux contamination is misclassified by ExoMiner with a score of 0.84. Overall, applying the scenario priors alleviates this problem; however, as we have suggested in the validation process discussed in Section 9.1, the classification performed by ExoMiner should be vetoed using FP flags provided in the Cumulative KOI catalog in order to reduce the risk of FPs being validated as exoplanets.
  • 2.  
    The assumption made in Equation (10) that x and I are independent might not be valid. This is because the information used in I, e.g., crowded field, might be present in the flux or centroid data (x). When this assumption does not hold, we are counting evidences for or against supporting an exoplanet classification twice (one from I and another from x). Thus, the application of the priors could lead to underestimating or overestimating the true probability. However, as we reported in Table 10, ExoMiner and other existing classifiers benefit from the scenario priors overall in terms of precision and recall. Therefore, the application of priors is beneficial.
  • 3.  
    The performance comparison between different classifiers should also be taken with caution owing to the existence of label noise. However, the relative performance of different classifiers is reasonably sound, as we showed in Sections 5.2 and 5.3 using additional independent data sets.
  • 4.  
    In Section 6.1, we only studied the effect of random label noise and underrepresented scenarios in the training set. However, the true nature of label noise or underrepresented scenarios could be systematic, limiting the reliability of our analysis. The presence of systematic label noise or underrepresented data scenarios could potentially lead to misclassification of FP scenarios for which there is significant label noise or lack of enough representative samples.

There are also astrophysical caveats, which include the following:

  • 1.  
    Given that there is no viable way to distinguish between BD FPs and real exoplanets using only Kepler photometry, ExoMiner and all similar photometry-based data-driven models are not able to correctly classify BD FPs. However, BDs are very rare in Kepler data: Santerne et al. (2016) found only three BDs in a follow-up study of 129 giant PCs from Kepler. The authors also calculated a BD occurrence rate of 0.29% for periods less than 400 days, 15 times lower than what they find for giant planets. Considering the geometric probability of alignment, which we can assume is ∼5% (conservatively), the chance of seeing a transiting BD is ∼0.05 × 0.29 ≈ 0.01%. Moreover, note that there are only a handful of validated large planets in our list, i.e., 17/301 with 4 R < Rp < 10 R and only 7/301 with 5 R < Rp < 10 R. Thus, given that transiting BDs can often be identified via radial velocity measurements, as well as the low occurrence rate of such systems, the BD FPs should be of little concern.
  • 2.  
    There could exist blended EBs that do not exhibit secondary eclipses but might be exposed by high-resolution imaging and photometric observations with better spatial resolution. There could still exist relevant patterns in the photometry data for such objects that a machine can utilize to correctly classify them, even though those are not easy for human vetters to find and use. However, we did not provide any specific studies of the performance of deep learning on such objects in this work. Thus, we assume that ExoMiner is not able to detect such FPs.
  • 3.  
    Due to the different stellar samples, pixel scales, and blending issues, TESS will have a different distribution of FPs, and this will affect classifier results. Thus, special care needs to be taken transferring to TESS where validation is concerned.

11.2. Future Direction

The analysis performed in this paper will allow us to pinpoint a number of future directions for improving and building on this work. First, we will change the ExoMiner architecture to include new important features, including (1) the number of transits and the uncertainty about the depth of the odd and even views to ensure that the model's use of the odd & even branch is more effective, and (2) the Kepler magnitude to allow the model to recognize saturated stars for which the centroid test is invalid. Second, we will perform a full explainability study by utilizing more advanced explainability techniques and design a framework that can provide experts with insight about ExoMiner's labels and scores. Third, we will be doing more in-depth analysis of the effect of label noise and underrepresented scenarios and provide a list of possible inaccurate labels in the Q1–Q17 DR25 data set. Finally, we will soon publish the results of utilizing an adapted version of ExoMiner to classify TESS transit signals by extending the set of features in a similar way to the one used for Kepler data and addressing some of the issues mentioned previously that affect TESS data.

H.V., M.M., L.W., and N.W. are suported through NASA NAMS contract NNA16BD14C. J.S., D.C., and J.T. are supported through NASA Cooperative Agreement 80NSSC21M0079. We would like to thank multiple people who directly or indirectly contributed to this work. We are very grateful to Porsche M. Parker and Haley E. Feck from USRA and Krisstina Wilmoth from NASA ARC, who have helped us tirelessly to recruit interns through the NASA ARC I2 intern program and NASA ARC Office of STEM Engagement. Without these amazing interns, this work would not have been possible. We are grateful to Patrick Maynard, who has recently joined our team as an intern and generated new insights into TESS data. We also appreciate David J. Armstrong for providing us with the detailed results of their ML models (Armstrong et al. 2021) and Megan Ansdell, who answered our questions regarding the ExoNet code (Ansdell et al. 2018). Our discussion with David J. Armstrong on the caveats of using ML for this problem helped us tremendously in improving this work. This paper includes data collected by the Kepler and TESS missions and obtained from the MAST data archive at the Space Telescope Science Institute (STScI). Funding for the Kepler mission is provided by the NASA Science Mission Directorate. Funding for the TESS mission is provided by the NASA Explorer Program. STScI is operated by the Association of Universities for Research in Astronomy, Inc., under NASA contract NAS 526555. We acknowledge the use of public TESS data from pipelines at the TESS Science Office and at the TESS Science Processing Operations Center. Resources supporting this work were provided by the NASA High-End Computing (HEC) Program through the NASA Advanced Supercomputing (NAS) Division at Ames Research Center for the production of the SOC Kepler and the SPOC TESS data products.

Footnotes

  • 11  

    Kepler observed an additional 86,663 stars for some portion of its 4 yr mission (Twicken et al. 2016).

  • 12  
  • 13  

    Early vetting efforts by the Kepler and TESS teams involved a two-step process wherein a team of vetters reviewed the DV reports and other diagnostics of each TCE and then voted to elevate potential PCs to KOI/TOI status for prioritization for follow-up observations.

  • 14  

    Citizen scientist projects to identify transit signatures in large transit photometry data sets were initiated by the Planet Hunters Zooniverse project for the Kepler mission (see, e.g., Schwamb et al. 2013) and continue in the TESS era (Eisner et al. 2020).

  • 15  

    (15 quarters × 90 days + 2 quarters × 30 days) × 2 cadences/hour × 24 hour/day = 67,680 cadences.

  • 16  

    Two TCEs can be generated for an exoplanet when its orbital period is less than the minimum orbital period in the Kepler pipeline transit search (0.5 days).

  • 17  

    The period can be one-half of the true value when there are less than three observed transit events for a given exoplanet. A minimum of three transits is required to define a TCE in the Kepler pipeline transit search.

  • 18  
  • 19  

    When using a ReLU function, some neurons get stuck on the negative side because the slope of ReLU is zero on that side. A neuron stuck on the negative side is called a dying neuron.

  • 20  

    This is the essence of any regularization technique.

  • 21  

    The out-of-transit centroid position for the DV OOT centroid offsets is determined by a pixel response function (PRF) fit to the time-averaged, background-corrected pixel values in the vicinity of the observed transit(s) for the given quarter (or sector).

  • 22  

    Similar to Ansdell et al. (2018), Shallue & Vanderburg (2018), and Armstrong et al. (2021).

  • 23  
  • 24  

    The bin width of 0.16 is obtained in Shallue & Vanderburg (2018) through hyperparameter optimization.

  • 25  
  • 26  

    Adam and SGD are two widely used stochastic gradient descent algorithms for optimizing DNNs.

  • 27  

    The cumulative KOI catalog was downloaded in 2020 February, so it does not include the newest CPs, including those validated by Armstrong et al.(2021).

  • 28  

    Our specific way of annotation, described in the previous paragraph, may confuse some NTPs with AFPs. For example, the second TCE belonging to an EB usually does not become a KOI, but technically it is an AFP.

  • 29  
  • 30  

    Because the data set is small, other metrics such as precision are very unstable for many thresholds. Precision can be 0.5 at one threshold when there are only two KOIs above threshold and become 1.0 for the next threshold when there is only one KOI.

  • 31  

    To be precise, there are 0.2 KOIs, i.e., only one KOI out of five runs.

  • 32  

    Note that GPC is calibrated by design and RFC was calibrated in post-processing.

  • 33  

    This is because if either of these probabilities is close to 1.0, numerator and denominator in Equation (10) become equal.

  • 34  

    These numbers are for the adjusted ExoMiner scores using scenario priors. The total number of TCEs with original scores >0.99 is 1507, of which there are four FPs.

  • 35  

    There was only one FP in the low-MES region whose correct disposition might be a planet.

  • 36  

    This number is 1186 for original ExoMiner scores without applying scenario priors.

  • 37  

    TESS target stars are on average 30–100 times brighter than those surveyed by the Kepler mission.

  • 38  

    A constraint is defined such that TESS TCEs must have at least two observed transits.

  • 39  
  • 40  
  • 41  
Please wait… references are loading.
10.3847/1538-4357/ac4399