This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy.

THE HIGH CADENCE TRANSIENT SURVEY (HITS). I. SURVEY DESIGN AND SUPERNOVA SHOCK BREAKOUT CONSTRAINTS

, , , , , , , , , , , , , , , , , , , , , and

Published 2016 November 29 © 2016. The American Astronomical Society. All rights reserved.
, , Citation F. Förster et al 2016 ApJ 832 155 DOI 10.3847/0004-637X/832/2/155

0004-637X/832/2/155

ABSTRACT

We present the first results of the High Cadence Transient Survey (HiTS), a survey for which the objective is to detect and follow-up optical transients with characteristic timescales from hours to days, especially the earliest hours of supernova (SN) explosions. HiTS uses the Dark Energy Camera and a custom pipeline for image subtraction, candidate filtering and candidate visualization, which runs in real-time to be able to react rapidly to the new transients. We discuss the survey design, the technical challenges associated with the real-time analysis of these large volumes of data and our first results. In our 2013, 2014, and 2015 campaigns, we detected more than 120 young SN candidates, but we did not find a clear signature from the short-lived SN shock breakouts (SBOs) originating after the core collapse of red supergiant stars, which was the initial science aim of this survey. Using the empirical distribution of limiting magnitudes from our observational campaigns, we measured the expected recovery fraction of randomly injected SN light curves, which included SBO optical peaks produced with models from Tominaga et al. (2011) and Nakar & Sari (2010). From this analysis, we cannot rule out the models from Tominaga et al. (2011) under any reasonable distributions of progenitor masses, but we can marginally rule out the brighter and longer-lived SBO models from Nakar & Sari (2010) under our best-guess distribution of progenitor masses. Finally, we highlight the implications of this work for future massive data sets produced by astronomical observatories, such as LSST.

Export citation and abstract BibTeX RIS

1. INTRODUCTION

The advent of a new generation of large-field-of-view astronomical optical CCD cameras, which are already operational (e.g., iPTF, Law et al. 2009; SkyMapper, Keller et al. 2007; Pan-STARRS, Hodapp et al. 2004; OmegaCam, Kuijken et al. 2002; DECam, Flaugher et al. 2015; Hyper Suprime-Cam, Furusawa et al. 2010; KMTNET, Kim et al. 2011) and planned (e.g., ZTF, http://www.ptf.caltech.edu/ztf; and LSST, LSST Science Collaboration et al. 2009), is revolutionizing our understanding of the universe because of their surveying capabilities. Thanks to these instruments, large regions in the sky are being mapped, up to very large distances, and rare, short-lived optical transients are also being found as these large regions of the universe are observed with high cadence. The latter presents not only new opportunities for the study of astrophysical phenomena, but also new challenges from the point of view of the data analysis. Large volumes of data will have to be processed in real-time in order to trigger follow-up observations that would help disentangle the physical nature of the transients detected (see, e.g., Gal-Yam et al. 2014).

The High Cadence Transient Survey (HiTS) is a discovery survey that takes advantage of the large etendue, the product of collecting area and field of view, of the Dark Energy Camera (DECam) mounted on the 4 m Blanco telescope at the Cerro Tololo Interamerican Observatory (CTIO), the fast connectivity available between CTIO and the Center for Mathematical Modelling (CMM @ U. Chile), and the computing capabilities of the Chilean National Laboratory for High Performance Computing (NLHPC) that allows us to observe and analyze high cadence DECam data in real-time. Because DECam is the largest etendue project in the southern hemisphere until the arrival of the full LSST project, HiTS can be considered a precursor project for some of the challenges regarding the fast analysis of large data volumes, the high cadence observations of deep-drilling fields and, depending on the cadence, the exploration of the hour-timescale transient population.

HiTS aims to explore the population of transient or periodic objects with characteristic timescales between a few hours and days (c.f. Kasliwal et al. 2010) and apparent magnitudes down to about 24.5 mag. Its main science driver was to discover the elusive shock breakout (SBO) phase of extended red supergiant star (RSG) progenitors undergoing core collapse (Falk 1978; Gezari et al. 2008; Schawinski et al. 2008), but it also focuses on the study of young SN explosions in general. The latter includes shock-hit companion stars in multiple progenitor systems of Type Ia SNe explosions (see, e.g., Marietta et al. 2000; Kasen 2010; Bianco et al. 2011; Bloom et al. 2012; Goobar et al. 2014; Cao et al. 2015; Marion et al. 2016); the shock cooling phase of core collapse supernovae (SNe), which provide constraints on their progenitors size and circumstellar environments (Moriya et al. 2011; González-Gaitán et al. 2015; Arcavi et al. 2016); and the early light curves of Type Ia SNe, whose diversity could be driven by different radioactive profiles in their outermost layers (Piro & Nakar 2014).

The structure of this manuscript is as follows. In Section 2, we will describe some of the relevant physics of RSG SN explosions during their earliest observable phases. In Section 3, we will discuss how the survey was designed and the details of our observation strategy. In Section 4, we will show how the real-time analysis of the data was performed, including a brief description of a newly developed image subtraction, candidate classification, and visualization pipeline. In Section 5, we will discuss some of the first results, including a detailed discussion on the limiting magnitude of the survey and its implications for SBO model constraints. In Section 6, we summarize the main results from this paper, and in Section 7 we discuss the implications from this work. We note that in this manuscript we only present our conclusions about the presence or absence of RSG SBOs, leaving the discussion on other classes of transient events for subsequent publications.

2. CORE COLLAPSE SNe

2.1. Shock Breakout

The core collapse of a massive star leads to the formation of a compact object and a shock wave that can unbind its outer layers in a core collapse SN. The shock can traverse the entire progenitor at very high velocities, vshock ≈ 0.01 c, until it eventually emerges at the surface of the star in an SBO event. The shock's energy density will be dominated by radiation and its main source of opacity will be Compton scattering by free electrons (Weaver 1976). The shock thickness, ${d}_{\mathrm{shock}}$, can be estimated by equating the photon diffusion timescale, the typical time taken by photons to randomly walk out of the shock via electron scattering, and the advection timescale, the time taken by the shock to traverse one shock thickness. This results in a shock thickness of approximately ${d}_{\mathrm{shock}}\sim \tfrac{c\,{l}_{s}}{{v}_{\mathrm{shock}}}$, where ls is the electron scattering mean-free path and vshock is the shock speed, or an optical depth of ${\tau }_{\mathrm{shock}}\sim \tfrac{c}{{v}_{\mathrm{shock}}}$ (ignoring form factors that may be significant in some cases, see, e.g., Schawinski et al. 2008).

As the shock front approaches the stellar surface it will encounter a decreasing optical depth. When τ ≤ τshock the radiation dominated shock will leak its energy out of the stellar surface as a radiative precursor until the shock breaks out completely and becomes rarified. The SBO timescale, tshock, will be given by the time that it takes for the radiative precursor to leak out into the stellar surface, ${t}_{\mathrm{shock}}\sim \tfrac{{d}_{\mathrm{shock}}}{{v}_{\mathrm{shock}}}$, but for an external observer it will also be affected by the light crossing time, ${t}_{\mathrm{light}}\sim \tfrac{{R}_{\star }}{c}$, where R is the star's radius. Shock breakout timescales are expected to be typically of about an hour for RSGs, several minutes for blue supergiants (BSGs), and several seconds for stripped-envelope stars (Kistler et al. 2013, c.f. Schawinski et al. 2008).

The SBO properties will depend on many structure and composition parameters, but most strongly on the radius of the progenitor star (Calzavara & Matzner 2004). Stripped-envelope, BSG and RSG stars have very different radii: of the order of 5–10 R, 25–50 R, and 500–1000 R, respectively. Even for similar masses, the envelope structure can vary significantly depending on whether the envelope energy transport is radiative or convective. Because BSG stars have radiative envelopes and RSG stars, convective envelopes, they have very different effective polytropic indices (n = 3 in BSGs and n = 1.5 in RSGs), which also leads to different shock acceleration dynamics. Additionally, the presence of pre-SN winds can strongly change the shock propagation physics and the observable properties of the breakout event (Balberg & Loeb 2011; Moriya et al. 2011; Svirski et al. 2012; Svirski & Nakar 2014a).

The characteristic temperature of radiation during breakout depends mainly on the radius of the progenitor stars, being approximately proportional to the inverse of the progenitor radius squared (Calzavara & Matzner 2004), but the typical energy of the photons leaving the star will also depend on whether thermal equilibrium between the escaping radiation deep inside the star and the intervening gas is achieved and if so, the depth at which this occurs (Nakar & Sari 2010). Most photons leaving the star during SBO will be in X-ray and UV wavelengths, making the detection of these events very difficult from the ground. In fact, a few SBO candidates have been detected from space: three SNe II observed in the UV at early times with GALEX (Gezari et al. 2008; Schawinski et al. 2008) and a recent SN II observed in the optical using Kepler (Garnavich et al. 2016) have been associated with shocks breaking out of the envelopes of their RSG progenitors. Other SNe II appear to have shocks breaking into high density circumstellar material (CSM; Gezari et al. 2010, 2015; González-Gaitán et al. 2015; Garnavich et al. 2016; Khazov et al. 2016; Tanaka et al. 2016), which is supported by models of RSG winds (Mackey et al. 2014). A Swift X-ray transient has been associated with an stripped-envelope SNe Ibc, from what appears to be a Wolf Rayet star (Soderberg et al. 2008) and/or its surrounding wind (Svirski & Nakar 2014b). No direct detections of SBOs from BSG stars have been made, but there is indirect evidence for an SBO in SN 1987A (Lundqvist & Fransson 1996).

Although UV and X-ray detectors are better suited for SBO detections because of the typical temperatures encountered at the shock's front, Tominaga et al. (2011) suggested that optical detectors can also be used to find these events. In fact, their models suggest that it may be easier to detect SBOs in a systematic way with a new generation of large-field-of-view optical CCD cameras (see, e.g., Morokuma et al. 2014; Garnavich et al. 2016; Tominaga et al. 2016).

2.2. Shock Cooling, Plateau, and Radioactive Tail

After SBO, the outer layers of the star will enter an adiabatic expansion and cooling phase, the so-called shock cooling phase. Nakar & Sari (2010) have divided the adiabatic expansion evolution after breakout into two distinct phases: planar and spherical expansion. During planar expansion the dominant contribution to the luminosity comes from the breakout shell, which would evolve adiabatically due to expansion at almost constant radius and with a luminosity ∝t−4/3. During spherical expansion the radius cannot be considered constant and energy from the inner shells also contributes to the light curve, with a slower luminosity evolution ∝t−0.17 to t−0.35 for polytropic indices 1.5 ≤ n ≤ 3. For RSGs, the transition between planar and spherical expansion is expected at about 14 hr after explosion.

The evolution of the near-ultraviolet (NUV) and optical luminosity during adiabatic expansion can differ significantly from that of the total luminosity. While the total luminosity generally declines monotonically after SBO, the NUV and optical luminosity can decline and then rise again for several days up to maximum optical light (e.g., Schawinski et al. 2008; Tominaga et al. 2011).

The adiabatic approximation will be valid preferentially in the denser inner layers, as the outer layers become dominated by radiative diffusion with time. The radius where this transition occurs is known as the diffusion wave radius, which can be estimated by equating the star's radiation diffusion time with the time since explosion. In those stars, where radiative diffusion is negligible during this phase, with initial radii of R ≤ 100 R, the product of ${R}_{\star }\,{E}_{\mathrm{in}}^{0.91}{M}_{\mathrm{ej}}^{-0.40}$ can be constrained by the luminosity evolution, where R is the stellar radius, Ein is the explosion energy and Mej is the mass of the ejecta (Chevalier 1992).

As radiative diffusion becomes important, the ionized He and H envelope can recombine in a wave that sweeps over the star inwards in mass coordinates, dominating the luminosity evolution in the plateau phase. The time when this transition occurs depends on the envelope mass and its structure, but it will usually start a few days after explosion (Chevalier 1992). The luminosity during this phase can evolve relatively slowly for a few months (see, e.g., Arcavi et al. 2012; Anderson et al. 2014a), depending on the explosion energy, ejected mass, initial radius, and composition (see, e.g., Popov 1993; Kasen & Woosley 2009), and will be followed by an exponentially decaying radioactive tail phase of evolution explained by the presence of newly synthesized 56Ni. A schematic representation of the previous phases of evolution is shown in Figure 1.

Figure 1.

Figure 1. Schematic representation of the optical light curve of a red supergiant core collapse supernova in arbitrary units (c.f. Figure 1 in Garnavich et al. 2016). Note that the wind SBO optical flux depends on the properties of the circumstellar material and may not always be dominant at early times.

Standard image High-resolution image

3. SURVEY DESIGN

Designing a real-time survey to look for SN SBOs can be seen as an optimization problem where the objective function is the total number of events to be detected. The constraints are that the cadence should be similar to the typical timescale of SN SBOs (of the order of an hour), that the time between observations cannot be shorter than what can be processed in a steady state with the available computational resources, and for our purposes that the events cannot be located at distances that are too large in order to facilitate follow-up with other astronomical instruments.

Unambiguous SBO detections should ideally be made in observational triplets with a timescale comparable to the typical timescale of RSG SBOs, each composed of a non-detection, detection, and non-detection/confirmation; as long as the position of the candidate coincides with the position of a subsequent SN explosion to discard other variable sources. Triplets can be more efficiently obtained with a high cadence, full night strategy: five epochs in a night contain three triplets with the same cadence, but three epochs in a night contain only one triplet. However, requiring too many epochs per night per field with a cadence of the order of an hour can limit the area in the sky, which can be efficiently observed due to airmass constraints depending on the latitude of the observatory. The SBOs and subsequent SNe can both be observed during the high cadence phase if it spans several nights, which should ideally relax into a low-cadence follow-up phase for several months to extract physical parameters from the SN light-curve plateau evolution.

Most observational parameters are correlated and affect the objective function of the optimization problem. The faster the cadence, the smaller the survey area. Shorter individual exposure times for a fixed cadence result in a larger survey area, but in a shallower survey with a larger fraction of overhead time, affecting the total volume of the survey in a non-obvious way depending on the telescope being used. Multiwavelength high cadence observations result in longer overhead times and slower cadences in a given band for a fixed area.

In this section, we will focus on the optimization of the initial high cadence phase, introducing two figures of merit used for this purpose.

3.1. Survey Instrument and Cadence

The most relevant variable characterizing the discovery potential of a survey telescope is the etendue—the product of the telescope's collecting area and the camera's field of view. Similarly, the data analysis challenges of a real-time survey are best characterized by the data acquisition rate, given by the ratio between the camera's image size and the typical exposure time plus overheads. A summary of some large-field-of-view cameras mounted in large collecting area telescopes is shown in Table 1. The large etendue of DECam is what enabled HiTS, but its large pixel number combined with the required real-time analysis became its main technical challenges.

Table 1.  Selection of Large-field-of-view (FOV) Optical Astronomical Cameras

Camera Area Field of View Etendue Pixels
  (m2) (deg2) (m2 deg2) (Mpix)
Kepler 0.7 115 81.5 94.6a
HSCb 52.8 1.5 79.2 870
DECam 11.3 3.0 33.9 520
PanSTARRS-1c 2.5 7.0 17.5 1400
iPTFd 1.1 7.8 8.6 92
SkyMapper 1.4 5.7 8.2 256
KMTNete 2.0 4.0 8. 340
QUESTf 0.8 8.3 6.5 40.3
LSSTg 35.7 9.6 344.2 3200
ZTFh 1.1 47 51.7 576

Notes. The collecting area, FOV, etendue (product of area and FOV) and number of pixels of each survey telescope are given. All listed cameras are already operational, except for LSST and ZFT.

aLimited bandwidth requires pre-selection of pixels. bHyper Suprime Camera. cPanoramic Survey Telescope and Rapid Response System. dIntermediate Palomar Transient Facility. eKoren Microlensing Telescope Network. fQuasar Equatorial Survey Team. gLarge Synoptic Survey Telescope. hZwicky Telescope Facility.

Download table as:  ASCIITypeset image

As mentioned before, the most important science driver for HiTS was the detection of SBO events in the optical in a systematic fashion. In an initial stage, the strategy consisted of detecting SBO events in single band observations, giving preference to rapid monochromatic sampling over multiwavelength characterization, and attempting rapid follow-up of any of the detected events using our real-time analysis and Target of Opportunity follow-up capabilities. If the SBO detection is not achieved in real-time, the position of the associated rising SNe could be used to look for the SBO event in a post-processing, forced-photometry analysis (in which we fix both the position and shape of the point-spread function, PSF). Then, if the number of detected events was found to be significant, we would attempt the detection and spectral index characterization in a second phase of multiwavelength, high cadence, real-time analysis.

SBO characteristic timescales range from seconds to a few hours (Kistler et al. 2013), but we first focused on the longest timescales among SBO events, i.e., those originating from RSG stars and with characteristic timescales of an hour or more, in order not to compromise the survey effective volume. For this family of explosions, we use the RSG explosion models from Tominaga et al. (2011) and Nakar & Sari (2010), shown in Figures 2 and 3 for the same progenitor parameters, with zero-age main-sequence masses (MZAMS) of 13, 15, 20, and 25 M and explosion energies of 1 foe9 for the 13, 15, and 20 M models and 3 foe for the 25 M model, in order to mimic the mass-energy relation found by Hamuy (2003) for observed SNe II and with some theoretical support from Müller et al. (2016) for core collapse SNe in general. To our knowledge, the models from Tominaga et al. (2011) are the only set of optical SBO models available in the literature that use exact solutions for realistic progenitor models and with a range of progenitor properties. We also use the analytic solutions from Nakar & Sari (2010) with the same parameters for comparison throughout the analysis.

Figure 2.

Figure 2. Evolution of the difference between the apparent magnitude and the distance modulus, i.e., an effective absolute magnitude, in u band (a) and g band (b), as seen at redshift 0.1, from shock breakout (SBO) optical peak maxima. Based on models from Tominaga et al. (2011; continuous lines) and Nakar & Sari (2010; dashed lines) and the DECam filters. The legend shows the zero-age main-sequence masses (MZAMS), the radii just before explosion and the explosion energy. See the text for more details.

Standard image High-resolution image
Figure 3.

Figure 3. Apparent magnitude evolution in the u band (a) and g band (b) from explosion time as seen from different redshifts, based on different models from Tominaga et al. (2011) (continuous lines), Nakar & Sari (2010; dashed lines), the DECam filters and a standard Λ–CDM cosmology. The explosion times are taken from the models of Tominaga et al. (2011). The models from Nakar & Sari (2010) are aligned to match the position of the SBO optical peak maxima and are only shown at redshift 0.05. The color coding is the same as in Figure 2.

Standard image High-resolution image

We consider cadences between 1.3 and 2 hr (4, 5, or 6 epochs per field per night) in what follows, which can be considered to be on the long side of the SBO timescale distribution, but that would allow us to put constraints on the number of RSG SBO events that could be detected and characterized by future optical surveys with sub-hour cadences. For comparison, the Kepler ExtraGalactic survey (Olling et al. 2015) has a cadence of 30 minutes and the Dark Energy Survey time supernova survey (Dark Energy Survey Collaboration et al. 2016) has a typical cadence of about a week.

We focused on the detection of the initial peak seen in Figure 2, which we define as any detection with a signal-to-noise ratio (S/N) of at least 5 before 90% of the magnitude change between the optical peak maximum and subsequent minimum has occurred for the Tominaga et al. (2011) models (thick lines in Figure 2), or before the beginning of the spherical phase of evolution for the Nakar & Sari (2010) models (break in the dashed lines in Figure 3). To confirm that these restrictions would lead to recognizable SBOs we simulated 2400 light curves for each family of models with different extinctions, redshifts and progenitor properties using the empirical cadence and limiting magnitudes of the survey and found that we are typically 97% and 95% efficient in visually recognizing an SBO detection with the previous restrictions under the models of Tominaga et al. (2011) and Nakar & Sari (2010), respectively, with a purity of 99% for both families of models.

3.2. Targeted or Blind Survey Strategy

We considered whether to select the survey fields by their intrinsic properties or not, i.e., to perform a targeted or blind survey, respectively. Here, targeted refers to targeting clusters of galaxies instead of individual galaxies, given DECam's field of view. To choose the best strategy, we compared the outcomes of simulations of SBO events using the volumetric SN rate from Strolger et al. (2015) and the cluster SN rate from Graham et al. (2012) with typical galaxy cluster density profiles from Voigt & Fabian (2006). We found that targeting clusters of galaxies with DECam, we could obtain an estimated 10% increase in the number of events compared to an untargeted survey. Given that there is a relatively small number of known clusters at relatively low redshifts (Chon et al. 2013), of which only a small fraction remains visible for the entire night, we estimate that the time spent slewing the telescope from target to target would be more than 10% of the photon collecting time of a blind survey. Therefore, we decided to perform a blind search, targeting cluster, or supercluster fields only when they were adjacent to some of our blind fields, which would also allow us to obtain volumetric SN rates with fewer selection biases.

3.3. Field Configuration

Having selected a blind field strategy, we find a reference right ascension (R.A.) that makes our targets achieve the lowest possible airmasses during all epochs, which is determined by the time of the year when the observations were initially allocated. Then we find a reference declination (decl.) that minimizes the average total extinction. We use an approximately rectangular grid of closely separated fields in the sky around this position, always switching to adjacent fields to minimize slew times10 , and balancing the exposure times with the rectangle dimensions to achieve an approximately constant airmass during each epoch of our observations. For example, we can observe along an approximately constant R.A. arc until the fields move by an hour angle equal to DECam's angular diameter, ${{\rm{\Theta }}}_{\mathrm{DECam}}$, which means switching to a new set of fields after approximately nine minutes.11 Thus, the exposure time and number of approximately constant R.A. fields are related, as well as the exposure times plus overheads with the total number of fields and epochs, by the following approximate equations:

Equation (1)

Equation (2)

Equation (3)

where Nfields is the total number of fields to be visited, Tnight is the duration of the night, Nepochs is the number of epochs per field, Texp is the exposure time, Toverhead is the overhead time (maximum between slew and readout time), NR.A. is the number of fields in the R.A. direction in our grid of fields to observe, and Ndecl. is the number of fields in the decl. direction in our grid of fields to observe. The previous equations are only approximate and for a more realistic selection of the optimal combination of fields to observe, we simulated the exact evolution of airmass and extinction with time for different field configurations. Also note that because slew times are given by the largest R.A. or decl. difference in the Blanco equatorial mount, we try to slew the telescopes diagonally in a R.A.–decl. grid.

To estimate the decl. that would minimize the combined extinction from the Milky Way and the atmosphere, we used Milky Way extinction maps and the relation between atmospheric extinction and airmass. This requires having defined the exact number of fields to observe, which depend on the choices of the number of epochs per field per night, Nepochs, and the exposure time, Texp (see Equation (1)), which are given by the solution of the survey design optimization problem.

3.4. Figures of Merit and Simulations

In this analysis, we use two figures of merit: (1) the number of SNe that would be detected at least once within the initial optical peak caused by the shock breaking into the envelope of the RSG star and (2) the number of SNe that would be detected at least twice within the first day in rest-frame time after shock emergence. These two quantities are related, but their comparison will give a sense of the challenges associated with detecting the SBO optical peaks. Note that for a transient candidate to be considered stationary (as opposed to moving, e.g., asteroids) and to be detected in our online pipeline, we require a new event to be detected at least twice in image differences in the same position in the sky.

In order to simulate how many SNe would be detected with these restrictions, we use the previously described light curves computed by Tominaga et al. (2011) and Nakar & Sari (2010; see Figures 2 and 3), assuming a Λ–CDM cosmology with ΩΛ = 0.73, ΩM = 0.27, and H0 = 71 km s−1 Mpc−1 and the DECam total efficiency. For the core collapse SN rate, we use the cosmic star formation rate density (SFRD) of Madau & Dickinson (2014), assuming a conversion efficiency between stellar mass and SNe, η, of 0.0091 ${M}_{\odot }^{-1}$ as measured by Strolger et al. (2015). Note that this conversion efficiency does not consider a large fraction of core collapse SNe missed by optical surveys (Mattila et al. 2012), so it is a conservative value for our purposes. We also tried the SFR of Horiuchi et al. (2011), a parametrization of Hopkins & Beacom (2006), which resulted in approximately 20% more events. Within a sufficiently large time interval T, we simulate a sampling function, e.g., 4, 5, or 6 epochs per night during several consecutive nights, and simulate a large set of explosions to estimate the likelihood of the event being detected at least once during the initial peak of a few hours seen in Figure 2 and the likelihood of the event being detected at least twice during the first rest-frame day after shock emergence. We record the distribution of detection ages to estimate more precisely the likelihood of detecting an event with a given age.

We weight these results taking into account the initial mass function (IMF) and normalizing their sum to 0.524 in order to reproduce the observed fraction of SNe II among core collapse SN explosions found by Li et al. (2011). In particular, we use a model weight proportional to a Salpeter-like IMF integrated in the vicinity of the model MZAMS:

Equation (4)

where w(M) is the model weight associated with a zero-age main-sequence mass M and where aM and bM define the integration mass interval for a given model, chosen to be either at the low-mass limit of 8.5 M, equidistant between the masses of the available models, or at either 16.5 or 30 M at the high-mass limit. The low-mass limit follows the low-mass constraint from Smartt et al. (2009) and the high-mass limit uses the high-mass constraint from the same work, but also considers the possibility that RSG stars are surrounded by CSM before explosion (see, e.g., Walmswell & Eldridge 2012; Gezari et al. 2015; González-Gaitán et al. 2015; Garnavich et al. 2016; Khazov et al. 2016; Tanaka et al. 2016), possibly becoming hidden from progenitor searches and allowing their masses to extend up to 30 M, which is consistent with the 95% confidence interval of Walmswell & Eldridge (2012). These Salpeter-like IMF distributions will be simply referred to as M16.5 and M30 distributions for the upper mass limits of 16.5 and 30 M, respectively.

Additionally, we have calculated model weights that would be representative of the estimated masses of observed RSG stars in the Milky Way (Levesque et al. 2005) and the Magellanic clouds (Levesque et al. 2006), assuming that these RSG stars are at the base of the giant branch and counting the number of stars closer to the mass of a given model to compute the weigths. The model properties as well as the integral intervals for Salpeter-like IMF distributions and associated weights for all SN II IMF distributions are shown in Table 2.

Table 2.  Explosion Model Properties from Tominaga et al. (2011) and Model Weights for the M16.5 and M30 Distributions Described in Section 3.4, as Well as those Representative of RSG Stars in the Milky Way (RSGMW) and Magellanic Clouds (RSGMC)

MZAMS RpreSN Z Eexp ${a}_{M}^{{\rm{M}}16.5}$${b}_{M}^{{\rm{M}}16.5}$ ${w}^{{\rm{M}}16.5}$ ${a}_{M}^{{\rm{M}}30}$${b}_{M}^{{\rm{M}}30}$ wM30 ${w}^{{\mathrm{RSG}}_{\mathrm{MW}}}$ ${w}^{{\mathrm{RSG}}_{\mathrm{MC}}}$
13 M 564 R 0.02 1 foe 8.5–14.0 0.433 8.5–14.0 0.310 0.126 0
15 M 507 R 0.02 1 foe 14.0–16.5 0.091 14.0–17.5 0.086 0.154 0.091
20 M 795 R 0.02 1 foe 17.5–22.5 0.071 0.181 0.370
25 M 1200 R 0.02 3 foe 22.5–30.0 0.057 0.063 0.063

Note. The zero-age main-sequence (ZAMS), pre-sn radius, metallicity, and explosion energy are shown, as well as the integration intervals in M and model weights from Equation (4). Note that the model weight's sum is 0.524, the observed fraction of SNe II in Li et al. (2011).

Download table as:  ASCIITypeset image

Assuming that all SNe II follow one of these explosion models, we estimate the number of events per unit redshift per field multiplying the cosmic SFRD, SFR(z); the conversion efficiency between stars and SNe, η; the total time of our simulation, T; the comoving volume per unit redshift bin per unit solid angle, $\tfrac{{dV}}{{dzd}{\rm{\Omega }}}(z);$ the field of view of DECam, ${\rm{\Delta }}{\rm{\Omega }};$ and the detection probabilities in the first day, ${P}_{M}^{{\rm{d}}}(z)$, or in the initial peak, ${P}_{M}^{{\rm{p}}}(z)$, obtained for one explosion model of mass M:

Equation (5)

The detection probabilities are obtained simulating large sets of explosions in redshift bins of width 0.01 observed with the survey cadence and depth. We assume an exponential host galaxy extinction distribution AV with a characteristic scale of λV = 0.187 and RV = 4.5, as done in Strolger et al. (2015). We count the number of events satisfying the detection conditions and multiply this number by the efficiency with which we can visually recognize SBO light curves as described in Section 3.1.

For our survey design simulations, we assumed a median seeing at CTIO of 0farcs75 in r-band (Els et al. 2009), converted to full width at half maximum (FWHM) values for different frequencies using the relations found in the DECam exposure time calculator (ETC). Initially, we computed the survey depth at the zenith using the available DECam exposure time calculator (ETC) and then subtracted the excess extinction with respect to the zenith at a given airmass using the relations from Stone & Baldwin (1983). After our survey was completed, we found that this significantly underestimated the effect of airmass and modified the available ETCs to also take into account the effect of airmass on the FWHM, the atmospheric extinction and the sky emission. We also used a 0farcs63 DECam instrumental FWHM12 and used a realistic evolution of the sky brightness13 to compute 50% completeness magnitudes (m50).

The number of detected SNe per field up to a given redshift will then be the SN II IMF-weighted sum of the number of events per unit redshift per field for the different explosion models of mass M:

Equation (6)

Different combinations of numbers of fields and numbers of epochs per night per field are associated with different exposure times, which in turn are associated with different limiting magnitudes and numbers of expected detections per night. In order to choose the best compromise between these quantities, keeping a cadence between 1.3 and 2 hr, we computed the total number of SNe that would be detected during SBO or that would be detected twice during first rest-frame day for different numbers of fields assuming 4, 5, and 6 epochs per night per field in different bands. We will show the results of these simulations using empirically derived limiting magnitudes in Section 5.

For our 2013A pilot campaign, we chose to observe in the u band based on the large temperatures during the SBO predicted by our models. However, in 2014A and 2015A, we switched to the g band after finding that this band offered the best compromise between increased detector efficiency and reduced SBO emission at larger wavelengths. Using these simulations, we chose 40 fields and 4 epochs per night per field for both our pilot 2013A u band campaign and real-time 2014A g band campaign. In 2015A, we changed to 50 fields and 5 epochs per night per field in order to explore faster cadences and to reduce the typical limiting magnitude of the survey, which facilitates follow-up observations with other telescopes. We will refer to the 2013, 2014, and 2015 HiTS survey campaigns as 13A, 14A, and 15A, respectively. In total, we were allocated four full nights in 13A, five full nights in 14A, and six full nights in 15A, always near the new moon. In 15A, we obtained three additional half nights at later times for multiwavelength follow up. The final coordinates of the fields observed in 13A, 14A, and 15A are shown in Figure 4 and their exact values are included in the Appendix Section in Tables 57. Note that the high cadence phase of observations were performed near the new moon: from March 12th to March 15th in 2013A (4 epochs per field per night), from March 1st to March 5th in 2014A (4 epochs per field per night), and from February 17th to February 22nd in 2015A (5 epochs per field per night), which explains a variation in optimal central R.A. by about two hours from 2013A to 2015A. Also note that, in order to minimize atmospheric extinction, a decl. close to the latitude of CTIO (−30°) is favoured, but in order to minimize galactic extinction large galactic latitudes are favoured.

Figure 4.

Figure 4. Spatial distribution of the fields observed in the 2013A, 2014A, and 2015A HiTS campaigns in equatorial (R.A., decl.) and galactic (b, l) coordinates. See the text for more details.

Standard image High-resolution image

4. DATA ANALYSIS

The driving requirement for our data analysis strategy was the ability to run in real-time and in a steady state, i.e., every exposure should be fully processed in less than a given time interval, which should be comparable to the exposure time plus overheads and should be much shorter than the typical cadence of the survey. This would allow us to modify our observational strategy with DECam and to trigger rapid follow-up observations with other telescopes during the same night of discovery and to obtain crucial progenitor information only present during the first hours and days post-explosion (see, e.g., Gal-Yam et al. 2014).

Given the data ingestion and computation rates imposed by the real-time requirement, our data analysis pipeline had to be hosted in a distributed memory type cluster with sufficiently fast connectivity with the observatory, which is ideally offered by computational facilities located in Chile today. Then, to have full control over the pipeline's optimization, we chose to write our own image subtraction pipeline focusing on speed, trying to minimize input/output (I/O) operations using the cluster node's shared memory for temporary data storage, and avoiding redundant steps as much as possible.

4.1. Pipeline Outline

The data reduction pipeline consists of a series of sequential steps triggered in a distributed memory type cluster immediately after every observation, satisfying the following restrictions: (1) non-parallelizable steps cannot take longer than the exposure time plus overheads, e.g., data transfer, and (2) parallelizable steps cannot take longer than the exposure time plus overheads times the ratio between the total number of available cores and the product of the number of parallel tasks and the number of cores per task.

The general structure of the pipeline is the following. Data is first transferred to the distributed memory cluster to undergo an initial pre-processing phase with data stored in the different node's shared memory, including mosaic expansion, calibrations, cosmic-ray removal, and indexing of the processed data. Then, the image subtraction pipeline is triggered, again using the different node's shared memory, and including registration, convolution, difference, photometry, candidate classification, and indexing of the processed data. Finally, web visualization tools are updated with new candidates for human inspection.

The pipeline was written mostly in Python14 , using Sun Grid Engine (in 14A) and SLURM15 (in 15A) for job distribution within the NLHPC16 supercomputer. We use bash17 for many I/O operations, C for mosaic expansion (step 2 below), external software for pre-processing, cosmic-ray removal, and catalog generation (steps 3, 4, and 5 below), and Fortran 9018 subroutines parallelized via OpenMP19 and integrated with Python using F2PY20 (Peterson 2009) for projection, convolution, difference object detection, and photometry (steps 7, 9, and 10 below). For the registration, convolution, and difference object detection, we use a similar approach to existing methods (e.g., Alard & Lupton 1999; Becker 2015), though with important differences in the representation of the convolution kernel and with our own Fortran routines for most of the computationally demanding steps. The following sequential steps are performed continuously during the night as data is acquired in the telescope.

  • 1.  
    Data transfer: a bash script is left running on the telescope to upload raw DECam images from CTIO to the NLHPC (in 14A) or from La Serena to the NLHPC (in 15A) as soon as they arrive at the telescope's observer or La Serena pre-processing computer, respectively. The script looks for files that do not change in size after a lag of a few seconds and then transfers them to the NLHPC using rsync 21 in less than 10 s.
  • 2.  
    Image expansion: the standard pre-processing DECam pipeline expands mosaic data into their individual CCDs serially, which can add a significant lag to our real-time pipeline. Thus, raw DECam mosaic files were expanded into their individual CCD images using a parallel custom made program, pimcopy, written in C and based on the CFITSIO library (Pence 1999). This allows us to perform the image expansion in less than three seconds for one mosaic image, about 60 times faster than a serial expansion.
  • 3.  
    Pre-processing: with the expanded files, we run a pre-processing pipeline. In the 14A campaign, we used a custom pre-processing pipeline, which subtracted bias frames from the raw images, divided the resulting image by a flat-field frame, and computed inverse variance maps based on the Gaussian and Poisson noise components of the reduced images. The bias and flat-field frames were obtained for every CCD combining several bias and flat-field images observed during calibration time via median filtering and normalization. We also computed bad pixel masks for every CCD, applying thresholds over the individual flat-field images. However, we did not correct for CCD cross-talk effects, which we found later to be important because of the appearance of ghost images of bright stars. Therefore, in 15A, we used a modified version of the DECam community pipeline (DCP) for the pre-processing stage, including electronic bias calibration, cross-talk correction, saturation masking, bad pixel masking and interpolation, bias calibration, linearity correction, flat-field gain calibration, fringe pattern subtraction, bleed trail and edge bleed masking, and interpolation.
  • 4.  
    Cosmic-ray removal: we remove cosmic rays from the pre-reduced images using a modified version of the public code CRBLASTER (Mighell 2010), using four cores per CCD via MPI.22 This code removes cosmic rays using Laplacian filter information, as in van Dokkum (2001). It takes about 20 s per CCD to remove cosmic rays, using 4 cores per CCD or 248 cores in total.
  • 5.  
    Catalog generation and reference image: with the cosmic-ray-removed DECam images and previously obtained inverse variance maps, we use SExtractor (Bertin & Arnouts 1996) to derive source catalogs of the images in pixel coordinates. Given that we do not have previous reference images of the same area of the sky and that we are dealing with a variability of hours, we use the first image of the sequence that has a relatively low airmass and good atmospheric conditions as our reference image for the relative and absolute astrometric solutions, to define a projection grid, for image difference and for the relative and absolute magnitude calibrations of the image differences.
  • 6.  
    Astrometric solution: with the previously generated catalogs, we match the brightest sources between the reference image catalog and the new image catalogs in pixel coordinates. Given that we chose the same pointing between different epochs and that the 4 m Blanco telescope has an equatorial mount, there is only a small offset with almost no rotation between epochs. Then we solve for a quadratic transformation between the pixel coordinates of both images using least square minimization, which was found to give consistent results to using the astrometric solution to apply the projection and convolution. To obtain the astrometric solution of the reference image, where all the projections are made, we solve for the astrometric solution using the positions of known stars from the USNO catalog (Monet et al. 2003) around our observations.23
  • 7.  
    Image projection: based on the previously derived transformations between the reference and new image in pixel coordinates we project the new image into the pixel grid of the reference image using a Lanzcos windowed sinc kernel with a kernel size of two. This kernel has been found to be a good compromise in terms of reduction of aliasing, sharpness, and minimal ringing (Turkowski & Gabriel 1990). This takes about 10 s using four cores per CCD.
  • 8.  
    Kernel estimation: in order to match the image point-spread functions (PSFs) and scale of two images, we use pixel based kernels with a non-uniform pixel size distribution and an approximately circular shape. This is a different kernel model than that used in Becker (2015), which uses an analytic basis to describe the kernel accross the CCD. Our model is pixel based (see Masci et al. 2016), but with radially growing pixel sizes to reduce the number of free parameters per kernel from 625 to 81, which has a regularization effect over the values of the outermost pixels. An example of the previously determined convolution kernels for a given CCD is shown in Figure 5. DECam is composed of 62 CCDs, each one of 4k × 2k pixels. We divide every CCD in 18 (6 × 3) different regions with independently computed kernels. The size of the independent kernel regions was chosen by training kernels with only one bright star and then measuring the typical image subtraction residual change with distance for other test stars, such that the residual change at the kernel separation was comparable to the typical scatter of image subtraction residuals among different relatively bright stars. The kernels are derived via least square minimization of multiple pairs of stars taken from the same region of the CCD, selected for being isolated, bright, and within the linear response regime of the CCD. The kernels are defined so that the smaller full width at half maximum (FWHM) image is convolved into the larger FWHM image, in order to avoid effective deconvolutions.
  • 9.  
    Image convolution and difference: we use the previously derived kernels to convolve the corresponding image and region of the CCD where it was defined. This changes both the shape and scale of the PSF of the convolved image. We then subtract the science and reference images, where one of them would have been convolved, to obtain difference and difference variance images in units of the image that has not been convolved. This takes less than 30 s using four cores per CCD.
  • 10.  
    Image difference photometry: using the difference and difference variance images, we then perform an optimal photometry calculation (Naylor 1998) over the entire image assuming that the PSF is centered in every pixel. This is achieved in less than 10 s using four cores per CCD, producing a PSF integrated photometry image centered in every pixel and an associated PSF integrated photometry variance image, which are scaled up or down into the units of the reference image. We then select those pixels with S/Ns greater than five to obtain candidate image stamps, using the optimal photometry flux and variance, which are fed into a machine learning classifier. Computing S/Ns for measured fluxes centered in every pixel of the image allows us to perform normality tests over the S/N distribution for real-time quality control.
  • 11.  
    Candidate filtering: every candidate is used to create candidate image stamps for which the dimensionality is reduced with custom-designed features that are input for a random forest classifier (e.g., Breiman 2001; Bailey et al. 2007; Goldstein et al. 2015) to be classified as either real or bogus. Because the training set was based on the 13A campaign (in the u band) to be used in the 14A/15A campaigns (in the g band), a key principle during the feature engineering process was to use as many dimensionless and scaleless quantities as possible. In order to not be dominated by unknown moving objects, e.g., asteroids, we consider that true non-moving transients are those that appear twice within a distance consistent with the astrometric errors and when at least one of the difference fluxes is positive with respect to the reference frame. A repetition of the candidate rules out moving transients, e.g., asteroids (to be presented in a separate publication), which are dominant among the candidates we detect, but only if it has at least one positive difference flux with respect to the reference frame. This is because a moving transient in the reference image should be present in all difference images with a negative difference flux with respect to the reference.
  • 12.  
    Light-curve generation: once a candidate has passed all of the previous tests, a light curve is generated. By default, the light curve contains only those difference fluxes of candidates that were classified as true transients by the random forest classifier. Optionally, the light curve can contain the difference fluxes from all the available epochs, even when the candidate was not selected by its S/N or if it was not selected by the machine learning classifier. This can be computed revisiting the difference images and doing a forced photometry (fixing the PSF shape and position) using the optimal photometry method, which may take significantly longer times than all the previous steps because of the many more I/O operations required. For this reason, we revisit all the epochs only for a few visually selected candidates during the night, and for all the candidates classified as real non-moving transients during the day.
  • 13.  
    Webpage visualization: as soon as a candidate is classified as a true candidate, it will be linked to a website that displays in visual form the location of all the candidates within the CCD. Candidates that are repeated and that have at least one positive difference flux with respect to the reference frame are marked with different colors to aid the visual inspection (using a combination of PHP24 , JavaScript25 and Highcharts26 ). The positions of known stars, galaxies, moving objects, and variable stars are also shown with different colors. We mark those candidates that are found to be consistent with having periodic light curves, using a Lomb-Scargle periodogram analysis (Lomb 1976; Scargle 1982), to separate fast periodic stars from fast transients. All fields for a given campaign are displayed in a single map, with about three new candidates appearing at the location of the last field visited after every observation, giving higher visibility to those candidates that have not been inspected in order to aid with the process. The visual inspection of all fields and CCDs can be done in about one hour by a single person, but it can be easily accomplished in real-time with these visualization tools.
  • 14.  
    Follow up: if a candidate is designated as a true non-moving transient by the previous tests, it will be followed-up only after going through the visual inspection test. We then decide whether to change DECam's observation plan, to obtain more observations in the same or other bands, or to trigger photometric or spectroscopic observations with other telescopes. Although we had the capability for triggering spectroscopic follow-up in a time comparable to about twice the cadence, no clear SBO candidates were found. Instead, during the last night of the 14A/15A high cadence run, we changed our observation to include multiple wavelengths and/or triggered spectroscopic observations toward some of the rising SNe.

Figure 5.

Figure 5. Example kernel used for the PSF matching convolution in a section of a pair of images, in nonlinear scale to enhance the contrast at low pixel values (left) and in linear scale (right).

Standard image High-resolution image

In total, we processed more than 1012 pixels at a maximum rate of 4.5 Mpix s−1, or 40 Mbps, generating about 108 significant image subtraction candidates (with an S/N greater than five), of which about 106 were classified as not bogus. Less than 104 of these candidates were selected by the classifier twice in the same position in the sky, which were then visually inspected in a stream of about three new candidates per minute. The total lag time between shutter closing and end of candidate visualization for a given field was an approximately constant time of about five to six minutes taking into account all the previous steps. A visualization of the real-time pipeline steps is shown in Figure 6 for the 15A campaign, which had the shortest expsure times of the three available campaigns. In this figure, the yellow arrow labeled Image diff. pipeline contains steps 5–12. This figure shows why the visual inspection should be done as fast as the telescope slew plus exposure and that the DCP is normally run in parallel with the image difference pipeline from a previous exposure.

Figure 6.

Figure 6. Schematic representation of the data analysis pipeline from telescope slew to candidate visualization for the 2015 campaign. The top arrows indicate the pipeline steps described in the text and their approximate execution times. The bottom blocks represent the previous steps when the pipeline is executed in real-time during the fast cadence phase of the survey, resulting in an online processing with a lag of less than six minutes between shutter close and end of candidate visualization for every observation.

Standard image High-resolution image

5. RESULTS

5.1. Survey Depth

In this section, we discuss the empirically determined depth of the survey, its relation to airmass and other observational variables, and how it compares to our survey design assumptions.

The depth of the survey in each of these campaigns was measured as an efficiency (probability of detection) versus magnitude for every observed epoch separately. This can be used to better estimate the expected number of events that should have been detected in the entire series of observations of a given field. The probability of detection at different magnitudes for a given set of representative observations is computed by building deep image stacks, co-adding selected epochs of a series of observations of the same field, and then measuring the fraction of real objects that were detected in single epoch observations relative to those in the deep stacks for a given magnitude range. Assuming that in the deep stacks all the true sources near the limiting magnitude of the individual images are detected, we measure the efficiency of detections in single epoch observations as a function of magnitude, taking into account all available CCDs simultaneously. We then fit the observed fraction with the following function.

Equation (7)

where P(m) is the probability of detection of a stellar-like source at a given magnitude, m; erf is the error function27 ; m50 is the best-fitting 50% completeness magnitude; and Δm50 is a scale parameter approximately equal to half the width of the transition region to low efficiencies. We tried several analytic expressions, but we found Equation (7), using the error function, to better resemble the observed transition in efficiency. The best-fitting analytic approximation of the efficiencies for all the CCDs of one of the fields observed in 13A, 14A, and 15A is shown in Figure 7. The large spread in 50% completeness magnitudes is due to the airmass variations required in our observing strategy, shown in Figure 8, but also due to bad observing conditions in some epochs. Δm50 best-fitting values were typically 0.6 ± 0.1 mag.

Figure 7.

Figure 7. Best-fitting efficiency vs. magnitude relations for a set of observations in the 13A (a), 14A (b), and 15A (c) campaigns. The differences between observed relations are due to different airmasses and changing environmental conditions at which the observations were performed.

Standard image High-resolution image
Figure 8.

Figure 8. Airmass distribution of the 13A, 14A, and 15A HiTS campaigns. In both the 13A and 14A campaigns, we observed four epochs per night per field, two at lower airmasses and two at larger airmasses, while in the 15A campaign we observed five epochs per night per field (using shorter exposure times), three at lower airmasses and two at larger airmasses.

Standard image High-resolution image

In Figure 9, we compare the best-fitting 50% completeness magnitudes (m50) and the 5σ limiting magnitudes reported by the DECam ETCs versions 5 and 6 (ETC v5 and v6, respectively) modified to include the effect of airmass. Note that the ETC v5 is included because it was the latest available ETC at the time when the 15A semester observations were performed. We show the 5σ limiting magnitudes using the modified ETC v5 with the predicted FWHM and sky brightness (ETC v5 + pred. cond.), the modified ETC v5 with the observed FWHM and sky brightness (ETC v5 + obs. cond.), the modified ETC v6 with the predicted FWHM and sky brightness (ETC v6 + pred. cond.), the modified ETC v6 with the observed FWHM and sky brightness (ETC v6 + obs. cond.) as well as the empirical 50% completeness magnitude (Observed). Note that in the 15A campaign the fourth and fifth nights were strongly affected by cloud cover and resulted in only three epochs in total (observation numbers 14–16) and that the 15A campaign contained follow-up observations starting at observation number 22. In all plots, the measured 50% completeness magnitude matches better the 5σ limiting magnitude produced with the modified ETC v6 using the empirical FWHM. We found that there is a difference of approximately 0.6 mag between the ETC v5 and v6, but also that there were FWHM variations beyond what is expected from airmass variations, which are also required to explain some of the m50 variations.

Figure 9.

Figure 9. Best-fitting 50% completeness magnitudes (m50) for the 13A (a), 14A (b), and 15A (c) survey campaigns and the 5σ limiting magnitudes reported by the ETCs v5 and v6 modified to include the effect of airmass as a function of observation number. We use observation number instead of time to avoid having data points too clustered to accommodate the daily gaps. See the text for more details.

Standard image High-resolution image

As mentioned before, we initially used the relation between airmass and extinction from Stone & Baldwin (1983) to calculate an excess extinction with respect to the limiting magnitude at the zenith, which is given by the original ETC, and use this information to derive the limiting magnitude at a given airmass. However, this led to a significant overerestimation of the limiting magnitudes. Instead, in the modified version of the ETCs, we assume that the FWHM scales as $\cos {(z)}^{-3/5}$, where z is the zenithal angle, that the sky emission scales linearly with airmass and that the atmospheric extinction increases exponentially with airmass to derive the limiting magnitude at a given airmass. A Python version of our modified ETC can be found at https://github.com/fforster/HiTS-public.

The relation between m50, FWHM, and airmass is shown in Figure 10. The expected relation between m50 and FWHM for constant seeings of 0farcs75 (dotted–dashed gray line) and 1farcs0 (continuous gray line) in the r band are shown, scaled to the u or g bands and using a range of airmasses (between 1.0 and 1.6 for 2013A and between 1.1 and 1.9 for 14A and 15A) to derive the FWHMs, where we also assume an additional 0farcs63 DECam instrumental seeing. We use a fixed sky brightness of 22.6 mag arcsec−2 in the u band for 13A and 22 mag arcsec−2 in the g band for 14A and 15A, scaled linearly with airmass (in physical units) with the modified ETC v6. With the modified ETCs, we expect that for the typical range of airmasses the limiting magnitudes can vary as much as 1.1 mag. Thus, the upper-left areas of Figure 10 can be explained by an airmass effect, but the bottom-right areas are mostly due to changing observing conditions, including the appearance of clouds in some of the observing nights.

Figure 10.

Figure 10. Relation between the measured 50% completeness magnitude (m50), the FWHM, and the airmass for mosaic averaged pointings in typical fields of the 13A (a), 14A (b), and 15A (c) survey campaigns. Gray lines show the expected relation for different FWHM at zenith. See the text for mode details.

Standard image High-resolution image

The previous efficiencies are based on single epoch measurements on full DECam mosaics, but our detections are based on image differences. To account for the additional loss of efficiency due to image differencing, we injected artificial stars and studied their S/Ns at injection and after recovery, concluding that the effect was typically a decrease in the S/N of up to $\sqrt{2}$, which would be expected when subtracting two images with the same noise level, or a shift in the efficiency versus magnitude relation of up to $2.5\,\,{\mathrm{log}}_{10}\,(\sqrt{2})\approx 0.38$ mag. This is because we did not use deep templates for image subtraction, but used the first observation at a given band that approached the lowest airmass values achievable (about 1.05 in 13A and 1.2 in 14A/15A, see Figure 8) of the first night, which had photometric conditions (second observation in the main band for the three campaigns), effectively amplifying the noise by up to $\sqrt{2}$ since the images are dominated by sky shot noise. Therefore, we shifted our efficiency-magnitude relations by 0.38 mag to account for the effect of image differences, but also used the original unshifted relation as the best-case scenario in what follows.

Overall, there are four relevant factors that made the survey depth shallower than expected: (1) the ETC available when designing the survey (v5) overestimated the limiting magnitudes by approximately 0.6 mag with respect to the latest ETC version (v6), which matches our empirical 50% completeness magnitudes well; (2) the effect of airmass was larger than expected, with a variation of up to 1.1 mag in the 50% completeness magnitudes instead of less than 0.3 mag using the relations from Stone & Baldwin (1983); (3) the observational conditions were not always good, with nights that had significantly larger than expected seeing values or even clouds; and (4) the difference imaging could lead to a loss in depth of up to 0.38 mag with respect to the individual epoch measured depths. The combination of all the previous effects can be seen in the large magnitude differences seen in Figure 9.

A practical description of the main observational parameters for the 13A, 14A, and 15A observational campaigns is shown in Table 3. Note that, for the 15A campaign, we had three DECam half nights separated by 2, 5, and 20 days from the end of the high cadence phase, where we also observed selected fields in the r and i bands.

Table 3.  HiTS Survey Description Summary

Semester Band Area (deg2) # nights # fields # epochs/night Exposure (s) Cadence (hr) Airmass Typical m50
13A u 120 4 40 4 173 2 1.0–1.6 23–24
14A g 120 5 40 4 160 2 1.1–2.1 23.5–24.5
15A g 150 6 50 5 87 1.6 1.1–1.9 23–24.5

Download table as:  ASCIITypeset image

5.2. Number of Detections

The lower than expected survey depths obtained in Section 5.1 (at least 0.6 mag of difference) should have an important effect on the predicted number of SBO events, which means that our initial assumptions during survey design had to be corrected. Thus, we have recomputed the expected number of events for different observational strategies using different assumptions about the survey depth. We use both the modified ETCs v5 and v6 for a different number of fields per night, assuming no airmass evolution or a realistic airmass evolution given a typical decl. and 4 or 5 epochs per field per night for 13A and 14A or 15A, respectively. We also test for the effect of the image difference process by assuming that it results in a loss of depth of 0.38 mag or assuming that it results in no loss in depth.

We found that using the ETC v5 instead of the ETC v6 led to an overestimation of the number of SNe that would be detected during SBO peak or twice during the first rest-frame day after shock emergence of approximately 2.5 in both cases. Using the relation between airmass and extinction from Stone & Baldwin (1983) to derive an excess extinction at a given airmass led to an overestimation of the number of SNe detected at SBO peak or twice within the same rest-frame day after shock emergence of approximately 1.8 and 1.4, respectively. Moreover, the loss in depth due to the difference imaging process can lead to an overestimation of up to approximately 1.8 in both quantities.

Using the ETC v6 and a realistic airmass evolution, we show the predicted number of SNe that would be detected at SBO optical peak or at least twice within the first rest-frame day after shock emergence as a function of the number of fields per night per epoch in Figure 11. We considered 4 continuous nights with 4 epochs per night in the u band, 5 continuous nights with 4 epochs per field per night in the g band, and 6 continuous nights with 5 epochs per field per night in the g band, as we did in 13A, 14A, and 15A, respectively. We also consider the two MZAMS distributions discussed in Section 3.4, M16.5 and M30, which differ in their upper mass cutoff limit.

Figure 11.

Figure 11. (a) Simulated number of SBO optical peaks (see Figure 2) as a function of the number of fields observed per night for the 13A, 14A, and 15A survey campaigns (see the text for details). The shaded areas show the possible effect of the image difference process, ranging from the case where the noise associated with the reference image is negligible to the case where it has a noise level similar to the science image. (b) Same as (a), but for the predicted number of SNe detectable at least twice during the first rest-frame day after shock emergence (see Figure 3). The gray vertical lines correspond to the actual number of fields observed in the respective survey campaigns.

Standard image High-resolution image

Figure 11 shows the predicted numbers of detections using the modified DECam ETC v6, for the M16.5 and M30 averaged ensemble of models as a function of the number of fields observed per night for the 13A, 14A, and 15A survey campaigns, as described in Section 3.4. We assume 4 epochs per field per night in the u band during 4 continuous nights for 13A, 4 epochs per field per night in the g band during 5 continuous nights for 14A or 5 epochs per field per night in the g band during 6 continuous nights for 15A. Shorter exposure times have smaller m50, which favors more nearby SNe that are easier to follow up, but which requires a faster real-time pipeline. We found no strong deviations as a function of the number of fields within the parameter range under consideration, i.e., wide/shallow surveys result in a similar number of events than narrow/deep surveys, with less than 20% and 30% variations with respect to the mean for the M16.5 and M30 distributions, respectively. The actual observational campaigns were optimized using the ETC v5, which resulted in maxima at a larger number of fields per epoch than the ETC v6. Wider/shallower strategies enable more follow-up capabilities, favouring larger numbers of fields per epoch or shorter exposure times depending on the available resources.

In Figure 12, we compare the predicted number of SNe computed using the empirical observational conditions (airmass, FWHM, sky brightness, and transparency) versus using the expected observational conditions (FWHM, sky brightness, and transparency), in order to estimate the effect of cloud cover and bad seeing in the 13A, 14A, and 15A campaigns. From this, we conclude that in the 13A campaign the atmospheric conditions led to an overestimation factor of approximately 1.1 for the number of SNe that would have been discovered at the SBO optical peak and 1.2 for those discovered twice within the first day after shock emergence. In the 14A campaign, the overestimation factors were approximately 1.4 and 1.6, respectively, and in the 15A campaign, approximately 1.7 and 1.6, respectively.

Figure 12.

Figure 12. Cumulative number of predicted SN detections during SBO optical peak (a) or at least twice during the first rest-frame day after shock emergence (b) as a function of redshift for the 13A, 14A, and 15A survey campaigns. We compare the cumulative numbers in the cases when the limiting magnitude is given by the ETC v6 with the expected FWHM, sky brightness, and transparency and in the case where the empirical probability of detection, P(m), is used (see Figure 7). The shaded areas show the possible effect of the image difference process, just as in Figure 11.

Standard image High-resolution image

In total, we estimate that the expected limiting magnitudes (using the unmodified ETC v5, the relation between extinction and airmass from Stone & Baldwin 1983, expected atmospheric conditions, and negligible reference image noise) led to an SBO number overestimation factor of up to 14x compared to using the empirical limiting magnitudes.

5.3. Progenitor Mass Distribution

The MZAMS distribution of the SNe found at SBO optical peak or twice within the first rest-frame day after shock emergence should differ considerably from the original SN II progenitor IMF distribution. More massive progenitor scenarios that become observed SNe II appear to have larger radii and more energetic explosions (Hamuy 2003; Müller et al. 2016), which would make them brighter after shock emergence. This should bias our discoveries toward larger MZAMS values in the M30 distribution. This can be seen in Figure 13, where we combine the 13A, 14A, and 15A campaigns and separate the predicted cumulative number of detections by associated progenitor model MZAMS for the M30 distribution (we lack enough model resolution for the M16.5 distribution) and the Tominaga et al. (2011) models. In the M30 distribution, about 70%–80% of the expected SBO optical peak detections should come from the weighted models with MZAMS of 20 or 25 M, and about 60%–70% in the case of the double detections at day 1 after shock emergence for each survey campaign. This means that the search for SBO or very young SNe will be strongly biased toward the more massive SN II explosions.

Figure 13.

Figure 13. Cumulative number of predicted SN detections during SBO optical peak (a) or at least twice during the first rest-frame day after shock emergence (b) as a function of redshift for the combined 13A, 14A, and 15A survey campaigns assuming the M30 distribution and the Tominaga et al. (2011) models. We show the cumulative number of SNe obtained with the empirical probability of detection P(m) and a perfect subtraction (the sum of the thick black lines in Figure 12), but separating the contributions of each IMF-weighted explosion model as explained in Section 3.4.

Standard image High-resolution image

5.4. Parameter Sensitivity

The number of predicted events is sensitive to the upper mass limit of the Type II supernova (SN II) IMF distribution, with a larger upper mass limit leading to more detections because they are generally easier to detect than lower mass, generally smaller radii explosions. As shown in Figure 13, an upper mass limit of 30 M would lead to about four times more SBO detections than an upper mass limit of 16.5 M, and about twice the number of SNe younger than one rest-frame day after shock emergence with at least two detections. We also tried a steeper IMF, with a γ of 2.3 instead of 1.3, and we expect approximately 50% fewer SBO detections for the M30 distribution.

We have also explored the sensitivity to explosion energy. More energetic models have brighter, narrower SBO optical peaks, and have significantly brighter shock cooling optical light curves. The 25 M MZAMS model from Tominaga et al. (2011), which is the dominant model in the M30 distribution, would produce 30% fewer or more optical SBO peak detections if instead of the 3 foe model we had used the 1 and 9 foe models, respectively. However, it would produce 2.7 and 2.6 times fewer or more SNe younger than one day after shock emergence than the 3 foe model, respectively. Thus, the explosion energy is not a very important parameter for the predicted number of SBO detections, but it is a relevant parameter for the number of very young SNe detected in their shock cooling phase.

We have also explored analytic models for the SBO (Nakar & Sari 2010) and early days of evolution of SNe II (Nakar & Sari 2010; Rabinak & Waxman 2011) and have found that they are effectively a two parameter family in the optical, where the two parameters are the progenitor radius and the ratio between explosion energy and progenitor mass. Larger radii, but also larger energy to mass ratio explosions produce brighter/slower light curves in these models. We have found that the models from Nakar & Sari (2010) are generally in better agreement with those from Tominaga et al. (2011) during the rise to maximum light, but there are significant differences among all these models, which suggest that the physical interpretation of early SN light curves should be treated with caution.

In addition to the model uncertainties, it is not clear how the SBO optical peak should be observationally defined. We define the end of the SBO optical peak for the Tominaga et al. (2011) models as the time when 90% of the magnitude change between the SBO peak and subsequent minimum (maximum in magnitudes) has occured, which is shown with thick lines in Figure 2. When trying a more conservative definition, as the time when only 50% of the previous magnitude change has occurred, we found a decrease of only 1% in the number of predicted SBOs. If we define the SBO optical peak end as the time when only 10% of the previous magnitude change has occured, we see a decrease in the number of predicted SBOs of only 40%. This low sensitivity to the definition of the end of SBO optical peak means that most predicted detections are expected to occur only once and very close to the SBO optical peak maximum (minimum in magnitudes). In fact, the number of SBO optical peak double detections is expected to be very small, only 18% of the SBO detections in the 90% magnitude difference definition, 3% of the SBO detections in the 50% magnitude change definition, and none in our simulations for the 10% magnitude difference definition.

5.5. Constraints on the SBO Optical Peak

In Figure 14, we plot the cumulative distribution of expected SBO detections as a function of redshift using the Tominaga et al. (2011) models with the M16.5 and M30 distributions for the combined 13A, 14A, and 15A campaigns. A dotted horizontal gray line indicates the number of predicted events from which a non-detection in the three survey campaigns has a probability of less than 5%. We conclude that, although the search for SBO optical peaks should be sensitive to the highest mass limit of the SN II IMF distribution, the shallower than expected depth of the survey implies that we cannot rule out either the M16.5 or M30 distributions for this family of models.

Figure 14.

Figure 14. Predicted cumulative number of SNe that should have been detected during SBO optical peak in the combined 13A, 14A, and 15A campaigns assuming the M16.5 or M30 distributions described in Section 3.4 for the selected models from Tominaga et al. (2011), assuming the empirical detection probabilities, P(m), and including the effect of the difference imaging as in previous figures. The horizontal dotted gray line indicates the expected number of events from which there is less than a 5% chance of having no detections.

Standard image High-resolution image

We then compare the previous results to those obtained using the brighter and longer-lived Nakar & Sari (2010) models and considering different model weigths, which resembles observed RSG stars discussed in Section 3.4. This is seen in Figure 15, where we show the total number of events that should have been detected with different assumptions about the SBO models, quality of the image difference and IMF weighting scheme. The corresponding non-detection probabilities are shown in Table 4, where we have found that the probability of having k detections in a Poisson distribution with mean λ is ${\lambda }^{k}{e}^{-\lambda }/k!$, which for k = 0 is eλ. The brighter and longer-lived models from Nakar & Sari (2010) models can be marginally rejected under all reasonable IMF distributions, especially the IMF distributions that favor larger progenitor masses. We cannot reject the Tominaga et al. (2011) models under any reasonable IMF distribution, but we marginally reject that most SBOs resemble those produced by their more massive 20 and 25 M at ZAMS models shown in Figures 2 and 3.

Figure 15.

Figure 15. Total number of expected SBO peak detections in the combined 13A, 14A, and 15A campaigns assuming the empirical detections probabilities, P(m), different SN II IMF distributions and the models from Tominaga et al. (2011) and Nakar & Sari (2010). The vertical lines range from the case where the difference imaging introduces a negligible amount of additional noise to the case where it increases the noise by $\sqrt{2}$, as done in previous sections. We assume the Salpeter-like M16.5 and M30 distributions, as well as IMF distributions representative of observed RSG stars in the Milky Way and Magellanic clouds (see Section 3.4), and show for reference the (non-physical) case that all SBOs resemble those produced by each of the selected models of a given mass (using weights of 0.524 for the given model and zero for the rest). The horizontal dotted and continuous gray lines indicate the expected number of events from which there is less than 5% and 1% probability of having no detections, respectively. We can marginally reject the brighter and longer-lived Nakar & Sari (2010) models under reasonable IMF distributions. We cannot reject the Tominaga et al. (2011) models under any reasonable IMF distribution, though we marginally reject that most SBOs resemble those produced by the more massive 20 and 25 M models. See Table 4 for corresponding non-detection probabilities.

Standard image High-resolution image

Table 4.  SBO Non-detection Probabilities for the Combined 13A, 14A, and 15A HITS Campaigns Assuming the SBO Models from Tominaga et al. (2011) and Nakar & Sari (2010) and different SN II IMF Distributions (see Figure 15) with an Optimistic Image Subtraction Depth (Negligible Noise in the Referece) and a Conservative Image Subtraction Depth (0.38 mag Difference)

SBO Models Tominaga et al. (2011) Nakar & Sari (2010)
IMF dist. min max min max
M16.5 0.40 0.60 0.003 0.04
M30 0.05 0.21 107 104
RSGMW 0.01 0.16 1011 105
RSGMC 0.03 0.09 109 106
13 M 0.15 0.36 0.01 0.06
15 M 0.34 0.54 0.03 0.14
20 M 0.01 0.06 1012 107
25 M 1014 108 1036 1019

Note. Bold values are values less than or equal to 0.05.

Download table as:  ASCIITypeset image

5.6. SN Candidates

Although no good SBO candidates were detected, several very young SN candidates were detected in the three HiTS campaigns, a few of them apparently less than a day after shock emergence, and more than a hundred SN candidates of all ages in total, whose positions were made available to the community (Förster et al. 2014a, 2014b, 2015a, 2015b, 2015c, 2015d, 2015e, 2015f, 2015g, 2015h, 2015i, 2015j, 2015k, 2015l, 2015m). We could spectroscopically classify about a dozen of these (see, e.g., Anderson et al. 2014b, 2015a, 2015b, 2015c; Walton et al. 2014a, 2014b; Förster et al. 2015n), confirming their SN nature in all cases. We have also used the models from Tominaga et al. (2011) to study the number of SNe II expected to be detected twice during their rise (because our template images were taken at the beginning of each run, we should only detect rising SNe) and found that we expect about 80 and 130 SNe II for the M16.5 and M30 distributions, respectively. Because these numbers are derived using an SN efficiency computed from actual observations (Strolger et al. 2015), this suggests that the total number of events we detect is roughly consistent with the observed rate of SNe within the relevant redshift range. Considering a contribution from other SN types (which will be the subject of future publications) and model uncertainties, this also suggests that the presence of other types of transients is not required to explain the number of detected events in our sample, and may even suggest a preference for the M16.5 distribution. We will present a catalog of calibrated light curves that will include follow-up resources other than DECam in a future publication, including host galaxy redshifts and post-SN images for the most interesting candidates. Interestingly, the SN candidates apparently younger than one day after shock emergence appear to show initially faster than expected rise times (see González-Gaitán et al. 2015; Garnavich et al. 2016; Tanaka et al. 2016), which suggests that the evolution of SN II light curves during the first days may not be explained by standard assumptions about their CSM.

6. SUMMARY

The first results of the HiTS search for SN SBOs were presented. With the current calibration scheme and data analysis pipeline, we see no clear evidence for RSG SBO optical early-time optical peaks in light curves resembling SNe II (with an S/N > 5). Based on a joint analysis of the three observational campaigns with our empirically derived limiting magnitudes, we conclude that ensembles of explosion models assuming a Salpeter-like IMF with an upper mass limit of either 16.5 or 30 M (M16.5 or M30 distributions, respectively) cannot be excluded for the Tominaga et al. (2011) models, but can be marginally excluded for the Nakar & Sari (2010) models (see Table 4). This result should be taken with caution given all the uncertainties associated to the distribution of SN II progenitor properties.

HiTS run in the optical using the DECam during the 13A, 14A, and 15A survey campaigns. The survey strategy could be described as several contiguous nights (4 nights in 13A, 5 nights in 14A, and 6 nights in 15A) of high cadence (2 hr in 13A and 14A and 1.6 hr in 15A), monochromatic (u band in 13A and g band in 14A and 15A), untargeted, varying airmass observations toward a large area of the sky (120 deg2 in 13A and 14A and 150 deg2 in 15A) with single epoch depths between 23 and 24.5 mag.

In March of 2013, a pilot phase of the survey was performed, with data being analyzed after the observation run had finished. In March of 2014, we performed the data processing and candidate filtering in real-time, the first real-time analysis of DECam data to our knowledge, though with very simple visualization tools, which delayed our reaction capability by a few hours. In 2015 February, we achieved the full data analysis, candidate filtering, and visualization process in real-time thanks to significant improvements in our visualization tools, which highlights the importance of fast visualization for real-time surveys. We processed more than 1012 pixels in a stream of 40 Mbps, which after processing resulted in a stream of about 3 new candidates per minute, 5–6 minutes after every exposure. As a result, more than 120 SN candidates were detected in total in real-time.

We computed empirical 50% completeness magnitudes analyzing deep stacked DECam images in relation to the individual epochs of the survey. The depth of the survey varied within each night typically by more than one magnitude as the observations had to be performed at varying airmasses in order to achieve the required cadence during the full night. We compared these values with the predictions of the public ETCs versions 5 and 6 (v5 and v6, respectively), which we modified to include the effect of airmass. We validated our modified ETCs studying the relation between these limiting magnitudes with the observed FWHM and airmasses. The elusive SBOs may have been detected if the actual survey depth had matched the initial limiting-magnitude estimations. However, the survey depth was overestimated due to several factors: an overly optimistic ETC available at the time (ETC v5), a stronger than expected airmass effect, worse than planned observing conditions and errors introduced by the image difference process.

During survey design, we used two figures of merit to determine the quality of an observational strategy, both defined in Section 3.4: (1) the number of SBO optical peak detections and (2) the number of SNe detected at least twice during the first rest-frame day after shock emergence. We have shown that an SBO optical peak detection is much harder to obtain than a double detection of the SN within the first rest-frame day. An obvious consequence is that an example of (2) does not mean that the SBO optical peak should have been seen in the data, which may explain why we have not seen optical SBO peaks in some data sets (see, e.g., Khazov et al. 2016).

Using the empirical limiting magnitudes with the models described in this analysis (from Nakar & Sari 2010; Tominaga et al. 2011), we evaluated these two figures of merit and found that the number of predicted events was most sensitive to the upper mass limit of the SN II IMF distribution. An upper mass limit of 30 M would lead to about four times more SBO detections than an upper mass limit of 16.5 M, and about twice the number of SNe younger than one day after shock emergence with at least two detections. The number of SNe detected twice during the first day after shock emergence appears to be even more sensitive to the explosion energy, varying by as much as the energy variation factor in the models tested.

An important consequence of marginally favoring the relatively dimmer and shorter-lived SBO optical peaks from the Tominaga et al. (2011) models is that, with our typical cadences, a real-time detection of an SBO will be unlikely to happen fast enough to react and observe it with other instruments. We expect more than 82% of our SBO detections to have only one detection before the end of the SBO optical peak with these models, thus we rely on the subsequent early rising SN light for their online identification. In fact, for the shock cooling SN light curve to rise to at least half a magnitude below the SBO optical peak maximum it takes typically more than half a day according to the more realistic Tominaga et al. (2011) models (see Figure 3), making the identification of SNe within a few hours of shock emergence incredibly challenging. This highlights the importance of having continuous high cadence observations during several nights followed by low-cadence follow-up observations in order to aid with the candidate selection and SBO identification via post-processing of the high cadence phase data.

7. IMPLICATIONS AND FUTURE WORK

Apart from the previously described SBO model constraints, an important contribution from this survey will be the detection of SN candidates younger than one day after shock emergence. A preliminary analysis shows that they have a very fast initial rise in their light curve inconsistent with the model light curves used in this analysis. They seem more consistent with shocks breaking into high density CSM in RSG stars (Gezari et al. 2015; González-Gaitán et al. 2015; Garnavich et al. 2016; Khazov et al. 2016; Tanaka et al. 2016). In order to compare their light curves to existing models, we are compiling host galaxy redshifts and post-SN explosion images for more precise and accurate absolute calibrations.

If the shock-CSM interaction in normal RSG stars significantly affects the SBO properties, which could be the case for the high density CSM SBOs, the evolution of SN II light curves during the first rest-frame day after shock emergence could be very different to that suggested by the models considered in this work. Thus, the detection of SNe during the first days after shock emergence could be a tool to constrain the properties of RSGs and their CSMs. Given the discrepancy in derived mass-loss rates between early and late times implied by these recent works, this could be a clue about the wind structure in RSGs (Mackey et al. 2014) or the latest stages of nuclear burning before explosion in these stars (Mcley & Soker 2014). These factors and our non-detection of RSG SBOs suggest that HiTS should switch to a lower cadence, multiwavelength survey mode.

If SBOs can be detected in a systematic fashion, they could provide an alternative probe for the upper mass limit of the SN II IMF distribution than pre-explosion progenitor detections (Smartt et al. 2009). This is because the UV radiation during the SBO phase is expected to sublimate most of the CSM dust that could be obscuring the SN II progenitors in pre-explosion images. Some evidence for dust column density changes after SBO exist for SN 2012aw (Fraser et al. 2012; Van Dyk et al. 2012), which appears to be on the high end of the progenitor mass distribution (although see Kochanek et al. 2012).

The high cadence strategy will be very important for future SBO surveys, since it is difficult to confirm an SBO detection without a previous non-detection and a subsequent drop in the light curve with a timescale comparable to the SBO optical peak. In a multichromatic survey, it may be possible to differentiate SBOs from early SN light curves based on their colors, but this requires either simultaneous multiwavelength observations or high cadence filter changes, which may be expensive for large etendue telescopes. In the case of LSST, intranight multicolor observations will be limited because of the limited number of allowed filter changes during its lifetime, so most SBO detections will most likely rely on high cadence monochromatic observations in the deep-drilling fields to be defined.

We have scaled our simulations to LSST's larger FoV, larger mirror area, smaller pixel size and shorter overhead times and found that by assuming 30 s exposure times we could observe 170 LSST fields per night with a cadence of 1.6 hr and would be able to find 13 times more SBOs and very young SNe than in our 15A strategy assuming 6 continuous nights of observations. Given that, in our simulations, 30 s LSST exposures would produce similar limiting magnitudes than our 87 s DECam exposures (a simple scaling suggests that 30 s g band LSST observations would be 0.1 mag deeper than 87 s g band DECam observations), this can be approximately explained by a combination of an increased number of fields that can be observed up to a limiting magnitude during one night (3.4 × larger) and LSST's larger FoV (3.2×), and is only slightly more than what would be obtained scaling the two telescope's etendues. We did a similar exercise for the Hyper Suprime-Cam instrument and we obtained a similar result, i.e., that the predicted number of SBO and young SNe should approximately scale with etendue.

Continuous observations with large etendue telescopes from space or from polar regions of the Earth could provide high cadence observations without the large airmass and limiting-magnitude variations experienced by this survey. In fact, the recent detection of an RSG SBO candidate with the large etendue, space-based Kepler observatory (Garnavich et al. 2016) combines both the continuous high cadence and large etendue required for this purpose. Alternatively, moderately large etendue, space-based X-ray or UV observatories may be better tools to look for SBOs in BSG and RSG stars, but with the GALEX mission now completed we will need to wait for future space missions for this to happen (e.g., Fraser et al. 2002). Interestingly, gravitational waves (Abbott et al. 2016) or neutrino detections (Hirata et al. 1987; Kistler et al. 2013) may be complementary methods for obtaining Earth-based high cadence observations of SBOs if candidates with a signature consistent with core collapse SNe in nearby galaxies within the detectors localization errors are targeted, but this will require either very-wide-field-of-view telescopes or arrays of robotic telescopes observing several galaxies simultaneously and able to reach the necessary absolute magnitudes shown in Figure 2.

Finally, it should be noted that the current observation and data analysis strategy followed by HiTS allowed us to study the optical universe with an unprecedented combination of total observed volume, high cadence for several continuous nights, and real-time data reduction including visualization, emphasizing the importance of interdisciplinary collaborations for astronomy. These observations are currently being used as a verification data set for the LSST software stack28 and will likely become a legacy data set for different scientific purposes.

We kindly thank N. Tominaga for providing detailed explosion models used in this analysis. We also thank C. Aragon, M. Graham, K. Maeda, R. Muñoz, E. Nakar, P. Protopapas, and F. Valdes for useful discussions, as well as the PESSTO collaboration for providing spectroscopic confirmations. F.F., J.S.M., and S.G. acknowledge support from Basal Project PFB-03. F.F., M.H., G.P., H.K., F.O., L.G., S.G., and P.E. acknowledge support from the Ministry of Economy, Development, and Tourism's Millennium Science Initiative through grant IC120009, awarded to The Millennium Institute of Astrophysics (MAS). F.F. acknowledges support from Conicyt through the Fondecyt Initiation into Research project No. 11130228. F.O., S.G., P.H., and G.C. acknowledge support from FONDECYT postdoctoral grants 3140326, 3130680, 3150460, and 3140566, respectively. F.F., J.C.M., P.H., G.C., and P.E. acknowledge support from Conicyt through the Programme of International Cooperation project DPI20140090. G.P. acknowledges support by the Proyecto Regular FONDECYT 1140352. We acknowledge support from Conicyt through the infrastructure Quimal project No. 140003. Powered@NLHPC: this research was partially supported by the supercomputing infrastructure of the NLHPC (ECM-02). This project used data obtained with the Dark Energy Camera (DECam), which was constructed by the Dark Energy Survey (DES) collaboration. Funding for the DES Projects has been provided by the U.S. Department of Energy, the U.S. National Science Foundation, the Ministry of Science and Education of Spain, the Science and Technology Facilities Council of the United Kingdom, the Higher Education Funding Council for England, the National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign, the Kavli Institute of Cosmological Physics at the University of Chicago, Center for Cosmology and Astro-Particle Physics at the Ohio State University, the Mitchell Institute for Fundamental Physics and Astronomy at Texas A&M University, Financiadora de Estudos e Projetos, Fundação Carlos Chagas Filho de Amparo, Financiadora de Estudos e Projetos, Fundação Carlos Chagas Filho de Amparo à Pesquisa do Estado do Rio de Janeiro, Conselho Nacional de Desenvolvimento Científico e Tecnológico and the Ministério da Ciência, Tecnologia e Inovação, the Deutsche Forschungsgemeinschaft and the Collaborating Institutions in the Dark Energy Survey. The Collaborating Institutions are Argonne National Laboratory, the University of California at Santa Cruz, the University of Cambridge, Centro de Investigaciones Energéticas, Medioambientales y TecnológicasMadrid, the University of Chicago, University College London, the DES-Brazil Consortium, the University of Edinburgh, the Eidgenössische Technische Hochschule (ETH) Zürich, Fermi National Accelerator Laboratory, the University of Illinois at Urbana-Champaign, the Institut de Ciències de l'Espai (IEEC/CSIC), the Institut de Física d'Altes Energies, Lawrence Berkeley National Laboratory, the Ludwig-Maximilians Universität München and the associated Excellence Cluster universe, the University of Michigan, the National Optical Astronomy Observatory, the University of Nottingham, the Ohio State University, the University of Pennsylvania, the University of Portsmouth, SLAC National Accelerator Laboratory, Stanford University, the University of Sussex, and Texas A&M. University.

APPENDIX: HITS FIELDS

Table 5.  Fields Observed During the 2013A Semester, Mostly in the u Band

Semester Field Name R.A. decl.
    (hr) (deg)
2013A Blind13A_01 10:36:32.960 −29:11:29.00
2013A Blind13A_02 10:43:03.600 −27:05:42.80
2013A Blind13A_03 10:38:35.450 −25:00:03.20
2013A Blind13A_04 10:44:42.390 −22:54:12.40
2013A Blind13A_05 10:55:12.630 −22:54:15.10
2013A Blind13A_06 10:49:16.180 −24:59:59.10
2013A Blind13A_07 10:53:56.110 −27:05:44.60
2013A Blind13A_08 10:47:38.500 −29:11:28.00
2013A Blind13A_09 10:58:43.680 −29:11:28.20
2013A Blind13A_10 11:04:48.510 −27:05:43.70
2013A Blind13A_11 10:59:57.150 −24:59:58.30
2013A Blind13A_12 11:05:43.380 −22:54:13.40
2013A Blind13A_13 11:16:13.750 −22:54:14.60
2013A Blind13A_14 11:10:37.970 −24:59:58.60
2013A Blind13A_15 11:15:41.070 −27:05:43.30
2013A Blind13A_16 11:09:49.250 −29:11:27.40
2013A Blind13A_17 11:20:54.390 −29:11:27.90
2013A Blind13A_18 11:26:33.450 −27:05:42.80
2013A Blind13A_19 11:21:19.040 −24:59:57.40
2013A Blind13A_20 11:26:44.600 −22:54:13.20
2013A Blind13A_21 11:37:14.920 −22:54:13.70
2013A Blind13A_22 11:31:59.860 −24:59:57.70
2013A Blind13A_23 11:37:26.010 −27:05:42.50
2013A Blind13A_24 11:31:59.970 −29:11:26.50
2013A Blind13A_25 11:43:05.120 −29:11:27.40
2013A Blind13A_26 11:48:18.410 −27:05:41.90
2013A Blind13A_27 11:42:40.870 −24:59:56.50
2013A Blind13A_28 11:47:45.780 −22:54:12.00
2013A Blind13A_29 11:58:16.120 −22:54:12.70
2013A Blind13A_30 11:53:21.710 −24:59:56.40
2013A Blind13A_31 11:59:10.950 −27:05:41.50
2013A Blind13A_32 11:54:10.700 −29:11:25.30
2013A Blind13A_33 12:05:15.840 −29:11:26.10
2013A Blind13A_34 12:10:03.360 −27:05:41.00
2013A Blind13A_35 12:04:02.680 −24:59:55.10
2013A Blind13A_36 12:08:46.960 −22:54:10.60
2013A Blind13A_37 12:19:17.350 −22:54:10.70
2013A Blind13A_38 12:14:43.550 −24:59:55.20
2013A Blind13A_39 12:20:55.930 −27:05:40.30
2013A Blind13A_40 12:16:21.420 −29:11:23.80

Note. 4 epochs per night per field were observed during 4 consecutive nights. The semester, field name, right ascension (R.A.), and declination (decl.) are shown.

Download table as:  ASCIITypeset images: 1 2

Table 6.  Fields Observed During the 2014A Semester, Mostly in the g Band

Semester Field Name R.A. decl. 2015A
    (hr) (deg) Field Name
2014A Blind14A_01 10:08:46.320 −02:05:44.81 Blind15A_33
2014A Blind14A_02 10:00:20.640 −02:05:44.81 Blind15A_28
2014A Blind14A_03 09:58:48.000 +00:09:00.00 Blind15A_27
2014A Blind14A_04 10:00:28.800 +02:12:36.00 Blind15A_26
2014A Blind14A_05 10:09:32.880 +02:05:44.81 Blind15A_35
2014A Blind14A_06 10:12:28.800 +00:00:00.00 Blind15A_34
2014A Blind14A_07 10:17:17.760 −02:05:44.81 Blind15A_38
2014A Blind14A_08 10:12:22.560 −04:11:29.62 Blind15A_39
2014A Blind14A_09 10:21:52.800 −04:57:00.00 Blind15A_42
2014A Blind14A_10 10:20:28.800 −06:31:12.00 Blind15A_40
2014A Blind14A_11 10:22:00.000 −08:06:00.00
2014A Blind14A_12 10:25:00.000 −09:34:12.00
2014A Blind14A_13 10:32:48.000 −08:48:00.00
2014A Blind14A_14 10:36:35.520 −06:17:14.42 Blind15A_50
2014A Blind14A_15 10:31:47.280 −04:11:29.62 Blind15A_49
2014A Blind14A_16 10:36:40.320 −02:05:44.81 Blind15A_48
2014A Blind14A_17 10:41:29.760 −04:11:29.62
2014A Blind14A_18 10:46:21.360 −02:05:44.81
2014A Blind14A_19 10:41:31.200 +00:00:00.00 Blind15A_47
2014A Blind14A_20 10:46:21.360 +02:05:44.81
2014A Blind14A_21 10:56:02.640 +02:05:44.81
2014A Blind14A_22 10:51:12.000 +00:00:00.00
2014A Blind14A_23 10:56:02.640 −02:05:44.81
2014A Blind14A_24 10:51:12.000 −04:11:29.62
2014A Blind14A_25 11:00:54.240 −04:11:29.62
2014A Blind14A_26 11:05:43.680 −02:05:44.81
2014A Blind14A_27 11:00:52.800 +00:00:00.00
2014A Blind14A_28 11:05:43.680 +02:05:44.81
2014A Blind14A_29 11:15:24.960 +02:05:44.81
2014A Blind14A_30 11:10:33.600 +00:00:00.00
2014A Blind14A_31 11:15:24.960 −02:05:44.81
2014A Blind14A_32 11:10:36.720 −04:11:29.62
2014A Blind14A_33 11:20:18.960 −04:11:29.62
2014A Blind14A_34 11:25:06.240 −02:05:44.81
2014A Blind14A_35 11:20:14.400 +00:00:00.00
2014A Blind14A_36 11:25:06.240 +02:05:44.81
2014A Blind14A_37 11:34:47.280 +02:05:44.81
2014A Blind14A_38 11:29:55.200 +00:00:00.00
2014A Blind14A_39 11:34:47.280 −02:05:44.81
2014A Blind14A_40 11:30:01.440 −04:11:29.62

Note. 4 epochs per night per field were observed during 5 consecutive nights. The semester, field name, right ascension (R.A.), and declination (decl.) are shown, as well as the 2015A field name for those matching the positions of fields observed during the 2015A semester.

Download table as:  ASCIITypeset images: 1 2

Table 7.  Fields Observed During the 2015A Semester, Mostly in the g Band

Semester Field Name R.A. decl. 2014A
    (hr) (deg) Field Name
2015A Blind15A_01 09:13:31.152 −06:18:04.68
2015A Blind15A_02 09:08:53.952 −04:18:02.81
2015A Blind15A_03 09:13:30.840 −02:17:49.30
2015A Blind15A_04 09:08:53.952 −00:17:59.06
2015A Blind15A_05 09:13:31.152 +01:42:02.81
2015A Blind15A_06 09:22:45.552 +01:42:02.81
2015A Blind15A_07 09:18:08.352 −00:17:59.06
2015A Blind15A_08 09:22:45.552 −02:18:00.94
2015A Blind15A_09 09:18:08.352 −04:18:02.81
2015A Blind15A_10 09:22:45.552 −06:18:04.68
2015A Blind15A_11 09:31:59.952 −06:18:04.68
2015A Blind15A_12 09:27:22.752 −04:18:02.81
2015A Blind15A_13 09:31:59.952 −02:18:00.94
2015A Blind15A_14 09:27:22.752 −00:17:59.06
2015A Blind15A_15 09:31:59.952 +01:42:02.81
2015A Blind15A_16 09:41:14.352 +01:42:02.81
2015A Blind15A_17 09:36:37.152 −00:17:59.06
2015A Blind15A_18 09:41:14.352 −02:18:00.94
2015A Blind15A_19 09:36:37.152 −04:18:02.81
2015A Blind15A_20 09:41:14.352 −06:18:04.68
2015A Blind15A_21 09:50:28.752 −06:18:04.68
2015A Blind15A_22 09:45:51.552 −04:18:02.81
2015A Blind15A_23 09:50:28.752 −02:18:00.94
2015A Blind15A_24 09:49:33.312 −00:04:07.46
2015A Blind15A_25 09:50:28.752 +02:09:46.01
2015A Blind15A_26 10:00:28.800 +02:12:36.00 Blind14A_04
2015A Blind15A_27 09:58:48.000 +00:09:00.00 Blind14A_03
2015A Blind15A_28 10:00:20.760 −02:05:44.81 Blind14A_02
2015A Blind15A_29 09:55:43.560 −04:05:46.68
2015A Blind15A_30 10:00:20.760 −06:05:48.55
2015A Blind15A_31 10:09:35.160 −06:05:48.55
2015A Blind15A_32 10:04:02.520 −04:05:46.68
2015A Blind15A_33 10:08:46.394 −02:05:44.81 Blind14A_01
2015A Blind15A_34 10:12:28.800 +00:00:00.00 Blind14A_06
2015A Blind15A_35 10:09:32.887 +02:05:44.81 Blind14A_05
2015A Blind15A_36 10:18:47.287 +02:05:44.81
2015A Blind15A_37 10:21:33.607 +00:05:42.94
2015A Blind15A_38 10:17:17.839 −02:05:44.81 Blind14A_07
2015A Blind15A_39 10:12:22.570 −04:11:29.62 Blind14A_08
2015A Blind15A_40 10:20:28.800 −06:31:12.00 Blind14A_10
2015A Blind15A_41 10:28:47.760 −06:31:12.00
2015A Blind15A_42 10:21:52.800 −04:57:00.00 Blind14A_09
2015A Blind15A_43 10:27:25.440 −02:08:27.52
2015A Blind15A_44 10:32:02.640 −00:08:25.66
2015A Blind15A_45 10:27:25.440 +01:51:36.22
2015A Blind15A_46 10:36:39.840 +01:51:36.22
2015A Blind15A_47 10:41:31.200 +00:00:00.00 Blind14A_19
2015A Blind15A_48 10:36:40.217 −02:05:44.81 Blind14A_16
2015A Blind15A_49 10:31:47.285 −04:11:29.62 Blind14A_15
2015A Blind15A_50 10:36:35.527 −06:17:14.42 Blind14A_14

Note. 5 epochs per night per field were observed during 6 consecutive nights followed by three non-consecutive half nights. The semester, field name, right ascension (R.A.) and declination (decl.) are shown, as well as the 2014A field name for those matching the positions of fields observed during the 2014A semester.

Download table as:  ASCIITypeset images: 1 2

Footnotes

Please wait… references are loading.
10.3847/0004-637X/832/2/155