This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy.
Brought to you by:
Article

On the Automated and Objective Detection of Emission Lines in Faint-Object Spectroscopy

, , and

© 2014. The Astronomical Society of the Pacific. All rights reserved. Printed in U.S.A.
, , Citation Sungryong Hong et al 2014 PASP 126 1048 DOI 10.1086/679285

1538-3873/126/945/1048

Abstract

Modern spectroscopic surveys produce large spectroscopic databases, generally with sizes well beyond the scope of manual investigation. The need arises, therefore, for an automated line detection method with objective indicators for detection significance. In this paper, we present an automated and objective method for emission line detection in spectroscopic surveys and apply this technique to observed spectra from a Lyα emitter survey at z ∼ 2.7, obtained with the Hectospec spectrograph on the MMT Observatory (MMTO). The basic idea is to generate on-source (signal plus noise) and off-source (noise only) mock observations using Monte Carlo simulations, and calculate completeness and reliability values, (C,R), for each simulated signal. By comparing the detections from real data with the Monte Carlo results, we assign the completeness and reliability values to each real detection. From 1574 spectra, we obtain 881 raw detections and, by removing low reliability detections, we finalize 652 detections from an automated pipeline. Most of high completeness and reliability detections, (C,R) ≈ (1.0,1.0), are robust detections when visually inspected; the low C and R detections are also marginal on visual inspection. This method of detecting faint sources is dependent on the accuracy of the sky subtraction.

Export citation and abstract BibTeX RIS

1. Introduction

Modern spectroscopic surveys produce large and relatively uniform spectroscopic databases (e.g., AGES, BOSS, BigBOSS, and SEGUE; Kochanek et al. 2012; Dawson et al. 2013; Schlegel et al. 2009; Yanny et al. 2009). Though many preprocessing reduction pipelines exist, the reduced spectra traditionally have been analyzed visually. Since the advent of the Sloan Digital Sky Survey (SDSS; York et al. 2000; Ahn et al. 2012), we have now entered an era of immense spectroscopic databases that defy manual inspection and necessitate automation. For example, Bolton et al. (2012) presented automated pipelines to classify galaxies and measure their redshifts from BOSS spectra, and Lee et al. (2008) did the same to measure stellar properties such as radial velocity, effective temperature, and metallicity from SEGUE spectra.

In this paper, we present an objective and automated method for the detection of line emission in spectroscopic data. This method is designed for the detection of single emission lines (e.g., arising from Lyα emission) within a certain optical spectral range; in our application, 4000–5000 Å. Since the continuum is usually undetected in the observed spectra of most Lyα emitting galaxies, our method mainly focuses on discerning real astronomical signal from coincident noise features by providing quantitative measurements of detection significance. Although the main application of our method is to search for faint emission lines on a zero-mean (continuum subtracted) background, we can extend our method to more generic spectra including radio and infrared spectroscopy, if the continuum can be properly removed.

The outline of our paper is as follows. In § 2, we describe the basic algorithm and present results from Monte Carlo (MC) simulations. In § 3, we apply our technique to 1574 sample spectra obtained using the Hectospec instrument on the MMT. We summarize our findings in § 4.

2. Method Description

The problem we address in this paper is how to automatically extract faint emission features (called "detections") and quantify their significance. For this purpose, we employ commonly used indicators of detection significance: "reliability" and "completeness." "Reliability" is the probability that a given detection is real (i.e., not resulting from noise). Completeness is the probability that a source of a given flux is detected, given the noise model and the detection method. A completeness equal to 1.0 implies that the feature is detected consistently in all observations of similar depth. In what follows, we describe the mathematical framework used in this paper and then calculate the reliability and completeness using mock observations from Monte Carlo (MC) simulations.

2.1. Mathematical Formulation

In this section, we present our mathematical framework and definitions of completeness and reliability.

2.1.1. Mathematical Definitions

A signal detection algorithm must quantitatively describe three parts: (1) the intrinsic signals, (2) the measured features, and (3) the detection criteria.

We can parametrize signals using a set of mathematical quantities. We call the set of parameters defining the signal as the "source vector," , which spans a "signal space." Signals are represented as points in this signal space. For example, Gaussian signals can be parametrized by three numbers, i.e, the centroid, width, and height.

We next define the quantities that characterize the detection of signals in observations. In our method, we use a threshold line on spectra (e.g., a flux or signal-to-noise threshold) and measure the width of features lying above the threshold line. In this case, the height of the threshold, ΘD, and the width of features selected by the threshold, ΔD, will be used to define a "detection" as shown in Figure 1. We call the set of parameters used to define the detection the "detection vector," denoted by .

Fig. 1. 

Fig. 1.  A mock spectrum ranging 4250–4750 Å with an underlying Gaussian signal, consisting of random noise with a Gaussian-profile emission line at λ = 4500 Å. We call the three parameters representing the underlying Gaussian profile the "source vector." The dashed horizontal line represents the threshold line with the height of ΘD. We define "detection" if the width of the feature above the threshold line is larger than ΔD. Changing the detection vector ξ = (ΘDD) affects the number, completeness, and reliability of detections.

Once source and detection vectors are defined, we can run a test for each signal, , to determine whether it is "detected" or "undetected." We denote this test using a "test function," , which gives Boolean results:

In this mathematical framework, the test function is a mapping from signal space to Boolean values, 0 or 1. The test function has a functional dependence on the detection vector, ; i.e., different can produce different test results. This means that by varying we can optimize the detection efficiency, reliability, and completeness.

In our analysis, we use a Gaussian profile to represent each source signal. Hence, the source vector can be written as for , where s(λ) is a source profile.

Now we take into account the "noise" contribution in our mathematical framework. A given detection vector produces one test result on a given single signal. To study the effects of noise, we describe the source as an MC ensemble

where is the i-th MC realization of . We can adopt various noise models to produce MC ensembles for . In our analysis, we choose simple Gaussian noise, assuming that each pixel is independent, and use a search window of 4000–5000 Å. The sky is nearly constant within this wavelength range, with the exception of two relatively weak telluric emission lines, i.e., Hg i 4047 Å and 4358 Å.

Rewriting the test result, , of in a simpler form,

we can summarize the presented mathematical representations as

  • 1.  
    : detection vector of parametrized detection criteria,
  • 2.  
    : source vector of parametrized source signal,
  • 3.  
    : an MC realization (simulated line profile) of source vector, ,
  • 4.  
    : a test result from an MC realization, .

2.1.2. Completeness and Reliability

Now we derive completeness and reliability from the above formulation. A series of N MC realizations for yields N test results, {τ12,...,τN}, for a given . Hence, the detection probability (or completeness) can be defined as the ensemble average of these test results

where the "MC" bracket represents the Monte Carlo average.

To derive reliability, we need to quantify the probability of false detections, i.e., detections resulting from noise. If the source signal is a null, , all detections are false and the result of contiguous (and perhaps correlated) noise spikes. Therefore, the false detection probability is the detection rate of a null signal; i.e.,

This false detection rate is an intrinsic limitation of any detection criteria. A conservative detection vector can reduce this rate but this also suppresses the detection of weak but real source signals. Detection vectors must be chosen to maximize completeness, while minimizing false detections.

Simply defined, reliability is the complement of the false detection probability, . However, since is constant (i.e., an MC averaged value), this would imply that different S/N (signal-to-noise ratio) detections such as 10σ or 100σ detections in fact have the same reliability value, which would be misleading. Reliability is therefore more appropriately defined as the complement of the false detection probability density.

We assume a source vector, , its MC ensemble, , and its test results, . According to the test results, we divide each ensemble into two disjoint "detected" and "undetected" subsets denoted by and . These two subsets satisfy

where we use braces to denote the sets collecting their MC instances and n{A} represents the number of elements for the set A. Equations (7)(9) are simply a restatement of the definition of the test function; the completeness of equation (5) is now rewritten as equation (9).

We use the term, , to represent one simulated MC line profile for . We can estimate the corresponding parameters for each by fitting the profile with a Gaussian model. We refer to this estimation process as "reprojection."

Thus, for a given source vector, , we have its MC ensemble of simulated profiles. By reprojection, we obtain its reprojected MC ensemble, , in signal space. From this distribution of , we can obtain its corresponding probability density function, , for , where is an independent variable for the density function in signal space. All the three processes, MC realization, reprojection, and probability density function generation can be depicted as

We denote the reprojection process by "↮" due to incompleteness of the line fitting process. Generally, the fitting procedure depends on a root-finding algorithm. There are inevitable fitting anomalies such as converging to a wrong solution or failing to converge within the root finding trials. Therefore, and do not necessarily represent a one-to-one mapping of each other. However, when the fitting process shows fair correspondence between and , we can represent by its density function, , as a practical estimate in signal space. In this argument, the detected subset, , can be described by a probability density function in signal space with an acceptable fitting performance:

We call the Detection Probability Density Function (DPDF) for . We note that, by the reprojection process and the definition of DPDF, all observed detections and quantities derived from MC simulations are now plotted in signal space; in our implementation, the parameter space for a Gaussian profile.

From equation (5), the normalization of DPDF for is . But, for convenience, we rescale the normalization condition as

where its original normalization is replaced by the conditional probability, .

The sourceless DPDF, , is the "false detection probability density" and its normalization is the "false detection probability" in equation (6),

Therefore, the distribution of is the zone of false detections in signal space. To derive the final mathematical expression for reliability, we assume an observed spectrum, σo, and its reprojection, . As a trivial case of , any observed detections would be deemed fully reliable due to the perfect decoupling of true and false detections. Therefore, at least, we can assign reliability = 1 to all observed detections of .

In the case where , we can have a couple of definitions for reliability. One of the simplest definitions would be to reject all of those partially reliable detections,

This definition can be powerful when focusing on detecting strong emission lines. However, it is clear that we need a better definition if we hope to detect weaker emission lines as well. The quick-and-dirty extension can be to use the profile shape of ,

where fmax is the maximum value of . In this definition, we can assign fractional reliability to the detections of . Expanding this to the full probability density function in order to obtain the correct reliability fraction, we choose a "contour" in the signal space to separate "reliable" from "unreliable" detections. For instance, we define this contour to be C(μ) such that:

where we have the boundary conditions of C(0) = 0 and C(1) = fmax. If we choose μ = 0.99, then any observed detection falling outside this region has a less than 1% chance of being a false detection; its "reliability" is therefore ≥99%. Therefore, we can define the reliability at as,

In practice, we do not have to calculate for all in signal space. We only need a reliability threshold for the adopted detection vector and its corresponding contour line such as C(0.05) or C(0.01) for 95% or 99% reliability.

Before we move on to the implementation, we discuss two points related to our mathematical framework. The first is about another possible variant of reliability,

Generally the false detection probability is very low, such as . If the rejection rate of noise signals is very high, then the reliability of a detection may be underestimated by the definition in equation (17). A weighted approach (provided by eq. [19]) may have some advantages in this case. However, we choose to ignore any detection that is indistinguishable from noise, and hence conservatively adopt the definition in equation (17) for the present study.

The second is about a more general interpretation of the DPDF. Since we have focused on the test of detection, we have only needed the sourceless DPDF and derived equation (18). However, we can derive the DPDF, , for a general source vector, . For an observed detection, , we can calculate as well as . As explained before, the latter, , provides the reliability of the detection. Similarly, but more interestingly, the former, , provides the probability that the observed feature, , is originated from the underlying source, . Therefore, the on-source DPDFs provide all the possible candidates of the underlying sources for a given observed detection with their own probabilities, though we do not have to use this extra information.

By following the basic descriptions above, the automated pipelines we will implement can extract detections from spectra and assign the completeness and reliability values to each detection. By restricting the reliability or completeness, we can finalize our detections. This is the basic frame for our automated detection method. The following sections will present the specific steps to build the detection pipelines.

2.2. Implementation

2.2.1. Standardization and Generation of the Pseudospectrum

Generally astronomical spectra are imaged on Charge-coupled Device (CCD) cameras. One-dimensional spectra are extracted from the two-dimensional images and typically represented by three data arrays of wavelength, spectrum profile, and inverse variance with pixel indices. Their pixel sampling rate, S/N, and flux calibration vary according to weather conditions, instruments, and optics. Therefore, we need steps to standardize the observation-dependent spectra for use with a generic detection algorithm.

In our detection method, we need two requirements of standardization: (1) uniform wavelength sampling and (2) normalization of variance to unity. The first is necessary to interchange between wavelength scale and pixel scale. The pixel scale is an easier and more standard unit to work on data-processing problems, while scientific requirements and constraints are written in terms of wavelength. The conversion between the two scales is consistent and trivial when the sampling rate is uniform. Therefore, for convenience and consistency, uniform wavelength sampling is required in our method. In practice, if pixel sampling rates vary within ± 10% in a target spectral range, the uniform resampling is not necessary.

The second is to normalize the noise level to unity. Since the noise level is a complex result of observational effects (e.g., due to instruments, optics, detectors, weather, etc.), we cannot model all possible situations in our MC simulations. Therefore, we normalize the observed spectra by their standard deviations to equalize the noise to unity. That is, if we denote the original data set of wavelength, spectrum, standard deviation of noise by (λ(x),s(x),σ(x)), then the normalized data set will be (λ(x),s(x)/σ(x),1). We call this normalized spectrum, s(x)/σ(x), the "pseudospectrum." Our Monte Carlo analysis is built for these pseudospectra. The pseudospectrum is not in physical units (e.g., flux density), but is effectively a dimensionless signal-to-noise ratio. The profile shape in the pseudospectrum is also deformed during the normalization process. This deformation is relatively small if the noises in neighboring pixels are similar. In practice, the shapes of features in the pseudospectra are similar to those in the original spectra as long as the noise varies smoothly under them, as is the case in spectral regions that do not include strong telluric lines.

To describe the meaning of a pseudospectrum, we start by summarizing the quantities in mathematical terms. In the spectral range of (λ00 + Λ), the flux, pseudoflux (integration of the pseudospectrum), and S/N can be written as:

where λ(x = 0) = λ0, λ(x = N) = λ0 + Λ, and σrms is a root-mean-square of σ(x); generally, Λ is a couple of times the FWHM of a typical emission line. For the trivial case of uniform noise, σ(x) = σ0, the mean and rms values are the same, , where is the mean of σ(x). If we extend this trivial case to "quasi-uniform" cases of , the pseudoflux can be approximated as

Unless our target line emission falls in close proximity to a bright sky line, equation (20) generally holds. The pseudospectrum is (roughly) a scaled measure of the S/N per pixel.

2.2.2. Threshold, Detection Mask, and Detected Width

We now choose a detection threshold and create a Boolean detection mask from the pseudospectrum, where "1" corresponds to pixel values falling above the threshold, and "0" corresponds to values below it. Figure 2 shows two threshold lines, ΘD = 1.5 (middle panels) and 2.0 (bottom panels) and their detection masks, where λ is on a uniformly resampled pixel scale as explained in the previous section. The top-left panel shows a spectrum composed of Gaussian random noise with σ = 1 (gray line) and a Gaussian emission feature of width 3 and height 5 centered at pixel λ = 200 (black line). The top-right panel shows their combined signal as a pseudospectrum. Using the detection mask, we identify clustered blocks of detected pixels and measure their sizes (widths). We call this block size the "detected width," δ.

Fig. 2. 

Fig. 2.  The left-top panel shows a Gaussian random noise spectrum with σ = 1 (gray) and a Gaussian source signal with pseudoflux = 37.6 and FWHM = 7.06 (black; or fmax = 5 and σ = 3). The right-top panel shows the combined profile of the source and noise. The y-axis is in the pseudospectrum scale and the x-axis in the pixel scale. With the given threshold lines, ΘD = 1.5 (middle-left) and 2 (bottom-left), we can make the detection mask assigning "1" where the signal is larger than threshold (middle- and bottom-right panels). This detection mask results in detection blocks of different widths δ. For each random realization, we measure detected widths and stack them to the probability function, pλ(δ; ΘD), at each pixel position. For a given position, λ, and a detection threshold, ΘD, the sum of all stacked counts for each detected δ is exactly the same as the total number of random realizations; i.e., . The detected width at λ = 200 is larger than at other sourceless positions. This property is the key feature for our detection method.

For a given source vector , we obtain one detected width value from each random noise realization. From a series of MC realizations, we obtain an MC ensemble of the detected widths denoted by . These measured detected widths are stacked and reduced to probability function:

with the normalization of . We call this probability function as Detected Width Probability Function (DWPF).

The two ensembles of on-source and off-source (i.e., on and off the emission line feature) width measurements are the fundamental data sets used to define the reliability and completeness in our detection method. From the ensembles, we obtain the corresponding DWPFs, and . Figure 3 shows the two DWPFs, pλ=200(δ; ΘD) and pλ=300(δ; ΘD), measured at the two different positions, λ = 200 and 300, for the given source vector shown in Figure 2. By definition, the DWPF measured at λ = 200 is the on-source DWPF for ,

Fig. 3. 

Fig. 3.  The distribution of detected widths, δ, at λ = 200 and λ = 300 from 105 random realizations. The y-axis represents a normalized count; hence, the integration of each distribution is unity. The measured δ distribution at the on-source position of λ = 200 is different from the one at the off-source position of λ = 300.

We can obtain the off-source DWPF by rerunning the MC simulations for . However, since there is no contribution from the source signal at λ = 300 (or sufficiently away from the signal), we can measure the off-source DWPF simultaneously as

The clear difference between pλ=200(δ; ΘD) and pλ=300(δ; ΘD) shown in Figure 3 is the key feature to set our detection criteria.

2.3. Completeness

From equations (22) and (23), we can directly derive completeness from the definition of "detection." Figure 3 shows that the probability of δ≥4 for the off-source ensemble is quite small, while most of the on-source ensemble have δ≥4. From this key observation, we choose a detection width threshold, ΔD, and define a "detection" as a case where the measured width is larger than this threshold, δ≥ΔD. If we parametrize a Gaussian profile using its pseudoflux and full-width-half-maximum (FWHM), the source vector and detection vector of our method can be denoted by

We now write the detection probability and false detection probability defined in equations (5) and (6) as

Using equations (22) and (23), Figure 3 shows the detection and false detection probabilities, and FD = 1.5,ΔD = 4) = 0.00005, where the source vector, , is given on the top-left panel in Figure 2. Since is the completeness for , we finally have

For each set of detection criteria {ΘDD}, we can calculate the completeness values for all source vectors .

Figure 4 shows the completeness contours in signal space for various {ΘDD}, derived from 104 MC realizations. The blue dotted and dashed lines show 0.95 and 0.99 completeness contours. Strong signals and high pseudofluxes guarantee high completeness as we would expect. For the same pseudoflux, the profiles with smaller FWHMs are more likely to be detected, since their peaks are higher. The limit of the narrowest width is enforced by the threshold width, ΔD. All of these properties form the bow-shaped contours of completeness in the signal space.

Fig. 4. 

Fig. 4.  The completeness (blue contours) and false detection probabilities (green) from 104 random noise realizations for various detection parameters; (ΘDD). The y-axis of FWHMs is in the pixel scale. By adopting more conservative detection criteria, i.e., higher ΘD or higher ΔD, we can reduce the false detection probability but, at the same time, we lose more completeness; i.e., losing more chances to detect faint source signals. We need to consider this tradeoff when choosing detection parameters. To apply our Monte Carlo results to the MMT/Hectospec data presented in § 3, we choose ΘD = 1.5 and ΔD = 4 shown in Fig. 5.

As ΘD and ΔD increase, the 0.95 and 0.99 contour lines move to the right side reducing the coverage area, but ensuring smaller false detection probabilities. This tradeoff needs to be considered when choosing the final detection vector. The data we will analyze in § 3 from the MMT/Hectospec have the spectral resolution of FWHM = 6 Å and pixel sampling rate of 1.1 Å/pixel. Based on Figure 4, we choose {ΘDD} = {1.5,4} as the final detection vector. This corresponds to the detection of the core of an emission line with 4 pixels lying at ≥1.5σ, or roughly a total core flux of ≥3σ.

2.4. Reliability

As described in § 2.1, we need to investigate each MC realization to derive reliability. Since we have one detection width for one MC realization, we can use this one-to-one correspondence to represent each MC realization,

Then, the detected and undetected subsets can be represented by the two δ≥ΔD and δ < ΔD as

Finally, during reprojection, a Gaussian is fit to features with δ≥ΔD,

The top panel of Figure 5 shows three reprojection samples for

  • 1.  
    C: the signal of pseudoflux = 70 and FWHM = 8,
  • 2.  
    R1: false detections for ΘD = 1.5 and ΔD = 4,
  • 3.  
    R2: false detections adding +1σ to baseline; its effective threshold ΘD = 0.5,

with the completeness contours for ΘD = 1.5 and ΔD = 4

The completeness of the sample C (red points) is equal to 1. The most probable point (black cross) of C coincides with its original source signal of pseudoflux = 70 and FWHM = 8. This is a typical reprojection pattern for strong signal. The R1 sample shows the reprojections of false detections, i.e., zero source signal, for ΘD = 1.5 and ΔD = 4. This is the theoretical limit of false detections. Practically, however, baseline estimation is rarely perfect and we may need to allow for some additional systematic uncertainty. Generally both sky subtraction and continuum estimation affect this baseline uncertainty. Underestimating the baseline (i.e., undersubtracting the sky background or continuum from spectra) will artificially elevate the noise above the threshold and result in an excess number of false detections. Conversely, oversubtracting the background will cause real features to be missed, though the false detection rate may be underestimated. In order to estimate the impact of errors in baseline subtraction, we added a systematic 1σ offset to the baseline and measured the resulting detections and their widths. We refer to this sample, which has an effective detection threshold of ΘD = 0.5, as the "R2 sample." The R2 sample results in a false detection probability ≈300 times larger than that of the R1 sample. Since the observed false detections (§ 3.2.1) are much lower than the false detection counts of R2 sample, we can use the R2 sample as a generous upper limit to the false detection rate (i.e., a conservative limit of real detections) to cover the conditions of poor sky subtraction or continuum subtraction which increase the uncertainty of baseline estimates.

The bottom panel of Figure 5 shows the final reliability contours defined in equations (18) and (17) for the R2 sample. This panel is our key result, and is used to assign reliability and completeness to each detected feature. Since R2 is an upper limit, the R2 contours provide more conservative reliability (i.e., more generous false detections probability density). The actual distribution of false detections likely lies between R1 and R2.

Fig. 5. 

Fig. 5.  Top panel: The contour plot of completeness for ΘD = 1.5 and ΔD = 4 and the reprojections of the three samples, (1) C: the signal for pseudoflux = 70 and FWHM = 8, (2) R1: false detections for ΘD = 1.5 and ΔD = 4, and (3) R2: false detections resulting from adding +1σ to the baseline; this results in an effective threshold ΘD = 0.5. R2 is an upper bound for the false detections; in practice, the false detection distribution is likely to lie between the R1 and R2 distributions. Bottom panel: The gray contours show the R2 distribution. The region lying outside a given R2 contour (gray) is reliable at the contour levels on the bottom panel.

2.5. Incomplete Reprojections: Fitting Anomalies

As previously mentioned, most root finding or optimization routines suffer at times from converging to local minimum (finding a wrong solution) or never converging to any solution (failing to find solution). The reprojection process, therefore, is not complete due to these inevitable fitting anomalies. Here, we investigate these anomalies to quantify the reprojection process.

In our method, we use MPFIT packages (Markwardt 2009) to fit a Gaussian line profile to each detected feature in order to estimate the source parameters. For a given detected width, δ, we add a width, W, on both of the left and right sides of the detection to define a fitting range of δ + 2W. When we run MPFIT with default settings and W = 3, we find three kinds of fit anomalies: (1) E1: centroid is out of the fitting range, (2) E2: negative flux, and (3) E3: width is narrower than 1 pixel (which is not possible for real features given the oversampled data). Figure 6 shows these fitting results for the R1 and R2 samples. Instead of pseudoflux and FWHM, we use the height, Fmax, and Gaussian sigma width, σ, to show the anomalies more clearly. They are related as pseudoflux = 2.5066σFmax and FWHM = 2.3548σ. Since we search for emission lines within the detected width, none of the three anomalies are valid fits.

Fig. 6. 

Fig. 6.  The fitting results for R1 (left panel) and R2 (right panel) samples, demonstrating the incomplete reprojection process. E1 and E2 are artificially assigned to (-2,-2) and (-3,-3). E3 anomalies have detection widths smaller than 1 pixel size, which is too small given the pixel sampling of the data.

In order to minimize fitting anomalies, we investigated the impact of various constraints to the postdetection line-fitting algorithm and converged on the following constraints: (1) positive flux only, (2) W = 3, and (3) the centroid must be located in the fit range; applied in the fitting routines. Based on these constraints, we obtained near-complete reprojections; zero anomalies for the R1 sample and four E3 anomalies for the R2 sample from 106 MC realizations.

2.6. Number Statistics of False Detections

In this section, we discuss how the false detection statistics depend upon the size of the spectral range over which the detection search algorithm is run. We assume that Λ is a spectral range of interest and pN(δ; ΘD) is a noise DWPF as described by equation (23). Since the source signal term is zero, pN(δ; ΘD) only depends on ΘD. Figure 7 shows pN(δ; ΘD) versus ΘD. This pN(δ; ΘD) is the probability of a detection of width δ at a certain sourceless position. Therefore, over the range Λ, we can conjecture that the possible false detection counts are the statistic of binomial trials using the probability of pN(δ; ΘD). This is true if we take into account the redundant counts of δ by diving pN(δ) by δ. Therefore, the number of false detections in the spectral range, Λ, can be expressed as:

where B(N,p) represents a binomial distribution of N trials with a probability p.

Fig. 7. 

Fig. 7.  The noise DWPF derived from 105 MC realizations. The inset figure shows the noise DWPF in log scale for ΘD≥1.0. For ΘD = 1.5 and 0.5 and a minimum detection width of δ = 4, the binomial distribution sums (eq. [33]) suggest false detection rates of ≈0.0002% and 0.6%, respectively.

To verify our argument, we produced 105 MC realizations with Λ = 500 and 2000 and measured the number of noise detections and compared these with their binomial distributions. Figure 8 shows the false detection counts from MC realizations (diamonds) and the binomial distributions (lines) from equation (33) for ΘD = 0.5 and ΔD = 4. We use ΘD = 0.5 to obtain enough counts for comparison, since pN(δ; ΘD = 1.5) is too small. This good agreement is not a surprising result because pN(δ; ΘD) itself is derived from the same setups of the MC simulation. This binomial description for the number statistics of false detections is in a sense rephrasing the definition of the DWPF. We note that, if Λ ≈ δ, the binomial description fails because within the length of δ the probability is not independent. In general, however, the search window is much larger than the width of an emission line, Λ≫δ. Therefore, for most cases, we can easily estimate the false detection counts from the DWPF, without running additional MC simulations, we can easily estimate the false detection counts from DWPF.

Fig. 8. 

Fig. 8.  The number of false detections for the two spectral ranges; Λ = 500 (top) and 2000 (bottom), with the detection criteria ΘD = 0.5 and ΔD = 4. The diamonds represent the false detection counts from 105 MC realizations, and the solid lines represent the binomial distributions derived from the DWFP. This figure verifies that the number statistics of false detections follows the binomial distribution derivable from the DWPF: .

For example, as we show in § 3, when we search for LAEs in the range from 4000 to 5000 Å with 1 Å/pixel sampling, i.e., Λ = 1000, with the detection criteria, ΘD = 1.5 and ΔD = 4, the false detection probability, , is 1.78 × 10-5. Hence, the average false detection counts for each spectrum are 0.0178 ± 0.1334. Since the false detection counts are much lower than 1 for a single spectrum, this detection threshold is sufficient to suppress most false detections. Since our total number of sample spectra is near 1000, a dozen false detections can occur in our sample. However, since those false detections have low reliabilities, we can filter them out. For ΘD = 0.5 and ΔD = 4, , is 0.00624. In this case, we expect 6.24 ± 2.49 false detections for each spectrum. As described in the previous sections about the R1 and R2 samples, these expected counts can be used as a generous upper limit in the case of poor sky or continuum subtraction.

3. Application

We now apply our method to search for Lyα emission lines in a LAE follow-up spectroscopic survey carried out with the MMT/Hectospec. We previously identified LAE candidates from an imaging survey using Subaru/SupremeCam with the IA445 filter (Prescott et al. 2008). As spectroscopic followup, we have targeted 1574 LAE candidates using the MMT/Hectospec (A. Dey et al. 2015, in preparation). We apply our detection pipelines to those 1574 spectra and describe the results below.

3.1. Data

3.1.1. Sample Selection

We performed an intermediate-band survey using Subaru/SupremeCam with IA445 filter (Prescott et al. 2008) on the NOAO Deep Wide-Field Survey Boötes field (NDWFS; Jannuzi & Dey 1999). The central wavelength of IA445 is 4458 Å and its FWHM width 201 Å. The full details about the reductions of the IA445 image are presented in Prescott et al. (2008) and A. Dey et al. 2015, in preparation. Briefly, the total exposure time was 3 hr. The images were reduced using the SDFRED software (Yagi et al. 2002; Ouchi et al. 2004). We registered and resampled the IA445 image to match the BW and R imaging from NDWFS. We retrieved ≈38,600 sources from the IA445 image using SExtractor with the 5σ limit of 26.5 AB mag (Bertin & Arnouts 1996).

3.1.2. MMT/Hectospec Observations

We observed selected candidates using MMT/Hectospec to confirm the presence of Lyα and obtain spectroscopic redshifts of LAEs. Hectospec is a 300-fiber and 1° field-of-view multiobject spectrograph at the MMT Observatory (Fabricant et al. 2005). We used the program XFITFIBS to assign optical fibers to science and calibration targets in our field and observed 7 configurations. We used the 270 line/mm grating blazed at 5200 Å with a resolution of FWHM ≈ 6.2 Å.

We observed 1574 LAE candidates using seven configurations over six nights as summarized in Table 1. The seeing was >1.6 in windy weather on 2012 May 16 and 17. The effective exposure times for M2 and J1 configurations are, therefore, smaller than the presented ones and their data qualities are poor. On the other days, the weather conditions were acceptable and the seeing was 0.64–1.0''. The details about the MMT/Hectospec observations and reductions will be presented by S. Hong et al. (2015; in preparation). Briefly, we use the HSRED package, a modified version of the Sloan Digital Sky Survey (SDSS) pipeline, written in Interactive Data Language (IDL) by Richard Cool.4 A description of the HSRED also can be found in Kochanek et al. (2012). The basic reduction of subtracting bias, flat-fielding, calibrating arc, and sky subtraction are done by the HSRED. We apply our detection method to the output spectra from the HSRED pipelines.

3.2. Results

We split the application of our automated pipeline into two phases: (1) raw detections and (2) a customized selection from the raw detections. During the first phase, the pipeline assigns all detected features their own reliability and completeness values. In the second phase, we set a specific reliability threshold to identify "reliable" detections and maximize the exclusion of false positives.

3.2.1. Raw Detections

For illustration purposes, we start by focusing on a single pointing—the PM1 configuration. We observed 244 LAE candidates, of which 179 have raw detections in the search wavelength range 4000–5000 Å for the criteria of ΘD = 1.5 and ΔD = 4. The number of false detections for R1 (in § 2.4), representing the case of perfect baseline estimation, is 0.016 ± 0.13; for all 244 spectra, 3.9 ± 2.0. For R2 representing the case of poor baseline estimation, we expect 5.6 ± 2.4 false detections per spectrum, for all 244 spectra, this corresponds to 1370 ± 37. However, since the sample R2 assumes that the baseline is overestimated by +1σ through the whole searching spectral range, the estimate from R2 is an exaggerated upper-bound. When the real baseline errors are suppressed or enhanced through the spectral range larger than ± 1σ, the effective spectral length enhancing the +1σ baseline error should be some fraction of the total spectral range. Therefore, accounting for this effective fraction of the enhanced fluctuation, the practical estimates of false detections may be around 0–3 false detections on each spectrum.

The top panels in Figure 9 show the reprojections (Gaussian fits) of the raw detections with the reliability (cyan) and completeness (blue) contours taken from Figure 5. These top panels are the key outputs from our detection method. The bottom panels show the histograms of line centroids alongside the transmission curve of the IA445 filter (green lines). From the false detection statistics presented above, we generally expect a couple of detections on each spectrum. When there is more than one detection in a given spectrum, we use the term "second detections" for secondary features and "third detections" for tertiary features, ordered by pseudoflux. The red points and histogram represent the first detections (first column), the grey the second detections (second column), and the black the third detections (third column). The red dotted lines on the bottom panels represent the sky emission lines, Hg i 4047 Å and 4358 Å.

Fig. 9. 

Fig. 9.  Raw detections resulting from our analysis of the data from MMT/Hectospec configuration PM1. The top panels show the observed source vectors (Gaussian fits of the detections) for the first, second, and third detections in each spectrum, as defined in the text, with the reliability and completeness contours shown, as in Fig. 5. The bottom panels show the histograms of the measured centroids with the green lines representing the throughput of the IA445 filter. The red dotted lines are the sky emission lines, Hg i 4047 Å and 4358 Å.

We find two important results from Figure 9. First, the centroid histogram of the first detections (bottom-left panel) shows a good correlation with the IA445 filter transmission curve, while the other histograms of the second and third detections (bottom-middle and -right panels) do not. This implies that the target selection and our automated method work properly, suggesting that the first detections are likely to be real emission line detections. Conversely, the poor match for the second and third detections implies that they are more likely to be noise detections. The fact that their reprojections shown in the top-middle and -right panels fall within the low reliability contours also supports this argument. A slight excess in the second detections near 4500 Å implies that some of the second detections could be real Lyα emission lines. There are very few real detections among the third detections. Hence, we exclude all the third detections from the real emission candidates. The second result is that there are many junk detections caused by residuals associated with the improper subtraction of the Hg I 4047 Å and 4358 Å telluric emission lines in all of the first, second, and third detections. The second phase of the pipeline analysis is needed to deal with these junk (sky residual) detections.

To investigate the raw detections further, we categorize the 179 raw detections into three kinds, called D100, D110, and D111. D100 represents the spectra having only first detections, D110 having first and second detections, and D111 having three or more detections. Figure 10 shows the histograms of the three categories, D100, D110, and D111. The sum of each column results in the histograms of Figure 9. The numbers of D110 and D111 detections are 46 and 17. If we assume there is one real emission line in D110 and D111, we have 46 + 2 × 17 false detections. When comparing these 80 false positives with the numbers predicted from the R1 and R2 reliability contours (4 ± 2 and 1370 ± 37 respectively), we find that the R2 contour significantly overestimates the false detection rate, as expected.

Fig. 10. 

Fig. 10.  Histograms of the detected centroids for the three detection categories, D100, D110, and D111, in PM1. The sum of each column results in the histograms in Fig. 9. Since the expected number of false detections is 3.9 ± 2.0 for perfect sky and continuum subtraction, the 46 and 17 detections in D110 and D111 are likely the result of erroneous baseline fluctuations caused by the practical limitations of sky subtraction.

3.2.2. Postprocessing and Reliable Detections

In this section, we describe the second stage selection that results in reliable detections. The bottom panel of Figure 5 shows the R2 contours, completeness contours, and R1 distribution. Since the true reliabilities lie between R1 (conservative false detections) and R2 (generous false detections), we need to set a locus excluding R1 fully and R2 partially. We find that the contour of completeness = 0.5 seems to work well, but there is no correct choice. One can choose more generous or strict reliability depending on how the detections will be used. Alternatively, all detections may be used, if weighting by the reliability. In our method, we use this completeness = 0.5 line to remove low reliability detections.

The sky residuals are systematic errors resulting from imperfect sky subtraction of strong telluric emission lines and are not governed by gaussian statistics. As a result, our adopted false detection contours are not able to filter out the "detections" resulting from these features. We therefore introduce an additional filtering process in the post processing phase, where we remove any "detections" that lie within ± 5 Å of the Hg i 4047 Å and 4358 Å telluric emission lines. If the first detection is excluded by these windows, the second is chosen as the primary detection. If the second detection is also excluded by one of the filters (i.e., by lying inside the false detection contours or too close to a strong telluric emission line), then the third detection is chosen. This case is very rare: for the 1574 spectra, only three cases are found where the third detection is the primary non-rejected detection (§ 3.2.3).

We show three examples of LAEA5569, LAEB21116, and LAEA15047, to illustrate the process of identifying robust detections. In LAEA5569 (shown in Fig. 11), both the first and second detections fall at wavelengths affected by systematic errors due to the subtraction of strong sky lines. While the detections (just barely) lie within our chosen reliability criterion, they are excluded by their wavelength position. In LAEB21116 (Fig. 12), the emission line is well detected, highly reliable, and well separated from any region affected by sky subtraction systematics. In LAEA15047 (Fig. 13), the detected emission line is similar to that in LAEB21116 and is likely to be a reliable detection. However, the feature is less conspicuous in visual inspection, and as the detection measures show lower confidence values of (C,R) = (0.777,0.935). We refer the reader to the figure captions for detailed descriptions of the characteristic plots produced by our code.

Fig. 11. 

Fig. 11.  The black crosses in the top left panel shows the wavelength locations of all the lines detected for the entire set of LAE candidates; the location of the two emission line detections for LAEA5569 are represented by the blue asterisk. The vertical and horizontal lines represent the locations of the two telluric emission lines within the search wavelength range (Hg i 4047 Å and 4358 Å); both detections for LAEA5569 fall on the locations of these sky lines. The right panel shows the wavelength histogram for all detections in the sample. The bulk of detections lie within the boundaries of the narrow-band filter used to select LAE candidates (green line). The vertical dot-dashed line shows the location of the first detection for LAEA5569, which clearly lies outside the filter selection range. The middle row shows the distribution of LAEs (purple crosses) in a FWHM vs. Pseudoflux plane (left panel) and on the sky (right panel), with the LAEA5569 detection marked as a blue asterisk. The bottom left panel shows the pseudoflux spectrum in the vicinity of the detected feature; it clearly resembles a sky-line residual. The bottom right panel shows the spectrum of LAEA5569 over the full wavelength search range with the emission line detection highlighted and the location of the Hg i sky lines shown as dashed vertical lines. This is a typical example of a case that must be removed by appropriate rejection.

Fig. 12. 

Fig. 12.  The output of our automated line detection procedure for source LAEB21116 showing a typical reliable detection. An emission line is detected with high completeness and reliability (C,R) = (1,1) at λ = 4482 Å. We observe the differences in the top-left, top-right, and middle-left panels from Fig. 11 after the selection procedures. This is a typical example of a reliable detection.

Fig. 13. 

Fig. 13.  The output of our automated line detection procedure for LAEA15047. A single emission line is detected at 4428 Å, but the feature has lower completeness and reliability (C,R) = (0.777,0.935) than the source in Fig. 12.

3.2.3. Application to Large Samples

We have demonstrated how our method works for data from a single MMT/Hectospec configuration PM1. Now we apply our method to all 7 configurations of the 1574 LAE candidate spectra.

Figure 14 shows the first, second, and third detections from the 1574 spectra in the same format of Figure 9. There are 881 raw detections in the 1574 spectra, 616 of which are first detections and 196 of which are second detections. The raw detection rate of all configurations is 0.56, which is lower than 0.73 of PM1 alone. The drop in the success rate is partly due to weather (the observations in 2012 were taken in poor weather and typically had shorter exposure times) and partly because of changes in the candidate selection between runs (see A. Dey et al. 2015, in preparation for details). As in Figure 9, most of the second and third detections are sky residuals or noise detections that follow the low reliability contours.

Fig. 14. 

Fig. 14.  The first, second, and third detections from 1574 spectra in all configurations. The numbers for each detection category are 616 for D100, 196 for D110, and 69 for D111, totaling 881 raw detections out of the 1574 observed candidates. Like Fig. 9, most of the second and third detections are sky residuals or noise detections.

Figure 15 shows the final 652 detections from the 881 raw detections. The results shown in the top-left, top-right, and middle-right panels are the same with Figure 12. There is, however, one additional interesting feature revealed by the larger sample. In the top-left panel, we can find the three strips of clustered points. The two horizontal strips at 4047 Å and 4358 Å are produced by the sky residuals as second detections. This is of no interest. The real interesting feature is the ascending strip from 4350 Å to 4550 Å, where the centroids of the first and second detections are similar, λprimary ≈ λsecondary. This feature results from the detection of double-peaked Lyα emission lines. Although many of the Lyα emitters exhibit single-peaked emission lines (at least, at our spectral resolution of 6 Å), 21 detections exhibit double peaked lines, where both peaks are independently recorded as significant detections by our algorithm.

Fig. 15. 

Fig. 15.  The final results of 652 customized detections. In the top-left panel, we see three stripes of clustered points. The two horizontal stripes at 4047 Å and 4358 Å are secondary detections caused by sky residuals. The third feature is an ascending stripe from 4350 Å to 4550 Å, where the centroids of the first and second detections are similar, λprimary ≈ λsecondary. This feature results from the detection of double-peaked Lyα emission lines. The left-middle, left-bottom, and right-bottom panels show the RA–DEC–z distribution.

The bottom-left panel shows the R.A. and decl. positions, the middle-left the R.A. and redshift, and the bottom-right panel the decl. and redshift of the 652 detections. The redshifts are calculated from the observed centroids and the wavelength of Lyα emission. We can find clusters and filaments from the redshift distributions. The heavily populated LAEs near z ≈ 2.67 form a wall-like structure rather than a compact cluster.

Figure 16 shows 7 randomly chosen LAE detections along with their C and R values. The first column shows pseudoprofiles, the second original profiles, and the third original profiles in the wider spectral range from 4000–5000 Å. Figure 17 shows 32 additional detections with their C and R values in the same format as the third column of Figure 16. Overall the automated detections are consistent with visual inspection. The high significance detections of C ≈ 1.0 and R ≈ 1.0 are quite clear and robust signals, while lower C and R detections are indeed more marginal upon visual inspection. For marginal detections, we prefer our automated approach because it allows for greater consistency and the possibility of quantifying the reliability of a detection free from the subjectivity of visual inspection.

Fig. 16. 

Fig. 16.  The profiles of the final customized detections with C and R values. The left, middle, and right panels show, respectively, the pseudoprofile, the original emission line profile, and the original profile in the full wavelength search range. The detections of (C,R) ≈ (1.0,1.0) are consistent and clear from by-eye inspection. They are also robust to most fluctuations of baseline due to their high S/N. For detections of low reliability and completeness, the signal features are visibly more marginal.

Fig. 17. 

Fig. 17.  Full spectral search window (as in the third column of Fig. 16) for 32 additional detections. The two gray–dotted vertical lines indicate the Hg i 4047 Å and 4358 Å sky lines. LAEA4882, LAEA5082, and LAEA5602 show some baseline fluctuations, possibly due to poor sky subtraction or weak underlying stellar continuum. In these cases, some weak detections are disputable as enhanced false detections due to uncertain baselines.

4. Summary

We have presented an automated and quantitative method to detect a narrow emission line within a target spectral range. The key point of the method is to generate reliability and completeness contours using Monte Carlo simulations. By comparing the contours and the observed detections, we can assign reliability and completeness values for each detected feature and finalize the detection using customized selection criteria.

We have applied our method to MMT/Hectospec observations of a sample of candidate Lyα emitters. All high significance detections of (C,R) ≈ (1.0,1.0) are clear and robust upon visual inspection. For marginal (C,R) detections, we prefer our automated detection approach, as it is more objective than visual inspection. Though the application presented here is designed to detect single faint emission lines, we can in principle apply our method to more general problems, assuming that the continuum baseline can be properly subtracted. For most cases, this requires continuum fitting using model templates (e.g., Tremonti et al. 2004; Bolton et al. 2012; Lee et al. 2008). Since each kind of object has its own continuum model (synthesized stellar spectra for galaxies, quasar templates for quasars, stellar spectral types for individual stars, dust models for infrared emissions), the continuum subtraction is better left as a separate step performed prior to automated line selection. When combined with a proper method of continuum subtraction, our method can be applied beyond faint emitter detections.

We are grateful to an anonymous referee for comments that have improved this paper. Observations reported here were obtained at the MMT Observatory, a joint facility of the Smithsonian Institution and the University of Arizona. The telescope time was granted both by NOAO, through the Telescope System Instrumentation Program (TSIP; funded by NSF) and through Steward Observatory. This paper uses data products produced by the OIR Telescope Data Center, supported by the Smithsonian Astrophysical Observatory. We are grateful to both the NOAO and Steward Observatory time allocation committees for their continuing support of this program. The research activities of SH and AD are supported by NOAO, which is operated by the Association of Universities for Research in Astronomy under a cooperative agreement with the US National Science Foundation. AD's research is also supported in part by the Radcliffe Institute for Advanced Study at Harvard University. SH is supported by NASA grants HST GO 10890, GO 11195, and GO 13000. MKMP is supported by a Dark Cosmology Centre Postdoctoral Fellowship; The Dark Cosmology Centre is funded by the Danish National Research Foundation.

Footnotes

  • Please see http://www.mmto.org/rcool/hsred/ for more information.

Please wait… references are loading.
10.1086/679285