Brought to you by:

The following article is Open access

On the Impulsive Heating of Quiet Solar Corona

and

Published 2021 July 27 © 2021. The Author(s). Published by the American Astronomical Society.
, , Citation Vishal Upendran and Durgesh Tripathi 2021 ApJ 916 59 DOI 10.3847/1538-4357/abf65a

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

0004-637X/916/1/59

Abstract

The solar corona consists of a million degree Kelvin plasma. A complete understanding of this phenomenon demands the study of quiet Sun (QS) regions. In this work, we study QS regions in the 171 Å, 193 Å, and 211 Å passbands of the Atmospheric Imaging Assembly (AIA) on board the Solar Dynamics Observatory, by combining the empirical impulsive heating forward model of Pauluhn & Solanki with a machine-learning inversion model that allows uncertainty quantification. We find that there are ≈2–3 impulsive events per minute, with a lifetime of about 10–20 minutes. Moreover, for all the three passbands, the distribution of power-law slope α peaks above 2. Our exploration of correlations among the frequency of impulsive events and their timescales and peak energy suggests that conduction losses dominate over radiative cooling losses. All these findings suggest that impulsive heating is a viable heating mechanism in QS corona.

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 4.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

Full disk images of the Sun taken in the extreme ultraviolet (EUV) and X-rays, consist of three features: active regions (ARs), coronal holes (CHs), and quiet Sun (QS). The ARs are the brightest, CHs appear darkest, whereas QS regions appear like a background. Note that while the number of ARs is a function of solar activity cycle, QS is present irrespective of the activity phase. There is a lot of literature studying the coronal heating in the ARs (see Reale 2014, for a review)—however, studies of coronal heating in QS and CHs are sparse (see, e.g., Tripathi et al. 2021). To address the physics of coronal heating in general, a comprehensive study of the heating in QS regions is of paramount importance (Aschwanden 2014).

One of the most popular mechanisms to heat the corona is via impulsive events, viz. nanoflares (see, e.g., Parker 1988). Impulsive events are transients generated through the dissipation of magnetic stresses or waves (see, e.g., Antolin et al. 2008). Solar atmosphere presents us with plethora of impulsive events at various energies and timescales, viz. flares (Benz 2008; Tripathi et al. 2004), microflares (Hannah et al. 2008; Schadee et al. 1983; Subramanian et al. 2018; Chifor et al. 2006, 2007), AR transient brightenings (Testa et al. 2013; Gupta et al. 2018; Tripathi 2021; Vilangot Nhalil et al. 2020; Rajhans et al. 2021), transition region blinkers (Harrison 1997), UV bombs (Peter et al. 2006; Gupta & Tripathi 2015), Ellerman bomb (Ellerman 1917; Pariat et al. 2004; Isobe et al. 2007), and other activities such as jets (Mulay et al. 2016; Raouafi et al. 2016). It is also observed that the properties of loops in ARs are well described by the impulsive heating scenario (Ghosh et al. 2017; Tripathi et al. 2009; Warren et al. 2008; Gupta & Tripathi 2015; Winebarger et al. 2013b, and references therein). Such a scenario has also been studied independently in ARs using a variety to techniques like time lag analysis (Viall & Klimchuk 2012, 2013, 2015, 2016, 2017), differential emission measure and analysis of Doppler shifts (see, e.g., Tripathi et al. 2010, 2011, 2012; Winebarger et al. 2011; Warren et al. 2012; Winebarger et al. 2013a; Subramanian et al. 2014; Del Zanna et al. 2015), hydrodynamic modeling (see, e.g., Bradshaw et al. 2012; Reep et al. 2013; Cargill et al. 2015; Barnes et al. 2016), magneto hydrodynamic modeling (see, e.g., Rappazzo 2015; Rappazzo et al. 2017; Knizhnik et al. 2018, 2019; Knizhnik & Reep 2020), and empirical models (see, e.g., Jess et al. 2019, 2014). Thus, it is natural to assume that the coronal heating in QS may also be governed by impulsive heating. However, since the QS regions have a very diffuse structure, it is not possible to count individual events and understand the energetics of these events in the QS.

To maintain the corona at a million degrees, the frequency distribution of impulsive events must follow a power-law distribution in energy—i.e., $\tfrac{{dN}}{{dW}}\propto {W}^{-\alpha },$ with α > 2 (see Hudson 1991). Observations do show that impulsive events in the corona follow a power-law distribution—however, there are a range of α values reported in the literature (see, for example, Figure 6.14 in Aschwanden 2019).

One of the most significant caveats in these results is due to the assumption that the each detected event is a single entity. But, we know that small-scale impulsive events may happen at a sub-resolution scale. Hence, individual brightenings may not be single entity but may consist of many tiny events. This may lead to the under counting of events particularly at lower energies. Moreover, the observations of QS radiance in the UV and EUV bands, both in space and time, show log-normal distribution (Pauluhn et al. 2001). Thus, the QS radiance might be generated due to some form of a Markovian process (Pauluhn & Solanki 2007; Gorobets et al. 2016).

To mitigate the above mentioned issues, Pauluhn & Solanki (2007) proposed an empirical model for the heating of the QS corona. Hereafter, we call it the Pauluhn and Solanki Model (PSM). In brief, PSM approximates the response of the plasma to a unit heating event as an exponential rise and fall of intensity. The amplitude of the heating event is sampled from a power-law distribution, with the frequency of occurrence of the events being kept as a free parameter. The resultant light curve is then a combination of a multitude of these events. Hence, this addresses the sub-pixel resolution scale of these structures. Moreover, the resulting intensities are also shown to be log-normal, mimicking the observations. Finally, the observed power-law distribution of energetic events is also incorporated in this model, thereby enabling us to understand the viability of impulsive events maintaining the observed intensity in a given light curve. Thus, PSM may provide an excellent proxy for the generation of the QS coronal intensities.

Pauluhn & Solanki (2007) generate the parameters for different observations by comparing similarity of intensity distribution and the global Morlét wavelet power (Torrence & Compo 1998) of simulated light curves with those of observed light curves. The comparison is qualitative—a sufficient good match in distribution, and the frequency location of peaks in the power spectrum were taken to represent a good match between the observation and simulation. Although the comparison had a sound basis, it was done by eye, and needs to be on a more quantitative foundation.

The problem of obtaining parameters from the observed light curve thus becomes an inversion problem. In general, the inversion approaches depend primarily on generating important features from the light curves, which then have a one-to-one mapping with the parameter set of the PSM. This is essentially performed qualitatively by Pauluhn & Solanki (2007). Since it is quite nontrivial to objectively pick out features for inversion, one trains an inversion model to pick out abstract features and perform the inversion. In this case, an optimization principle guides the mapping from light curves to parameter set developed by the inversion model.

Tajfirouze & Safari (2012) and Bazarghan et al. (2008) employed this method by using a probabilistic neural network (PNN). Under this scheme, every simulated light curve is classified, where each class has a unique combination of free parameters. The PNN is trained on the full set of simulated light curves. Finally, the observed light curves are fed into the PNN, which then assigns each of them to one of the learned classes. For their study, Tajfirouze & Safari (2012) used ≈10,000 light curves (at max). These light curves corresponded to CHs, QS, and ARs on the Sun obtained from the Atmospheric Imaging Assembly (AIA; Lemen et al. 2012) on board the Solar Dynamics Observatory (SDO; Pesnell et al. 2012) and data from the Extreme UltraViolet Imager on board Solar TErrestrial RElations Observatory (STEREO; Kaiser et al. 2008). On average, they obtained α > 2 for all the regions.

This method is a great start to a tricky stochastic inversion problem. But, such an approach must be well validated by an exhaustive testing set. Moreover, the classification of light curves imposes a discretization constraint on the parameter set, which depends on the grid resolution of the simulation. Thus, any assigned class to a given light curve may change on improving the grid resolution. In other words, we do not know the confidence level of the PNN for each inversion.

We further emphasize that Tajfirouze & Safari (2012) used poor resolution of AIA 171 Å data (90 s cadence and 2'' spatial resolution), and further binned the data spatially by 3 × 3 and 5 × 5 windows. Such a binning will average out the small-scale events, and thus, will not allow the use of full potential of AIA observations. Here, we use the full spatial and temporal resolution observations taken using AIA and perform a regression on the target parameter set using a convolutional neural network (CNN; LeCun et al. 2015), that circumvents the potential caveats in the work of Bazarghan et al. (2008) and Tajfirouze & Safari (2012) as detailed above.

A CNN is used to preserve information of the timescales of features in the light curve along with the distribution of intensity values. A CNN may be thought of as a set of kernels, which perform convolution on the input and return a convolved output. The kernel size can be interpreted as the scale over which information is summarized. Several such kernels are operated on a given input, and these outputs are passed through a nonlinear function called the activation function. Successive kernels are thus sensitive to local scales of the corresponding input, but that input itself may be an extremely nonlinear transformation of the original light curve.

Furthermore, we do not use a simple CNN for the inversion. Instead, we use an inception module, which improves its performance (see Szegedy et al. 2015). The inception module provides kernels with multiple sizes at the same level and makes is possible to perform multiscale analysis at each level. By employing a CNN to capture the scales of variation present in the data, we circumvent grid resolution issues by considering the target as a regression problem. We can effectively interpolate between simulated grid points while performing the inversion. With the CNN, we also obtain associated inversion uncertainties by perturbing the network.

The rest of the paper is arranged as follows. We describe our data sets and associated noise in Section 2 and Section 3, respectively. The modeling and validation of the network is discusses in Section 4. The results are presented in Section 5, followed by a summary and conclusions in Section 6.

2. Observations and Data

For this work, we used the observations recorded by AIA on board SDO. AIA observes the Sun's atmosphere in the UV and EUV bands using eight different passbands sensitive to plasma at different temperatures (O'Dwyer et al. 2010; Boerner et al. 2012). Here, we used the data taken using 171 Å, 193 Å, and 211 Å passbands. These images are taken with a pixel size of ∼0.6'' and an approximate time cadence of 12 s. We have chosen these particular passbands since the count rates in these passbands for QS are large, when compared to others (see Table 2 in O'Dwyer et al. 2010).

Monitoring the EUV images on SolarMonitor 1 , we identified QS patches during 2011 and 2019, where no activity was observed. Details of the two data sets (DS1 and DS2) are given in Table. 1. In total, we obtained eight continuous hours of data for each set, which corresponds to 2400 time steps. All the images are aligned to the first image and are exposure time normalized. The full field of view (FOV; single snapshot) for DS1 and DS2 as observed in 171 Å is displayed in the left panels of Figures 1 and 2 with the corresponding spatial distribution of intensities in the right panels.

Figure 1.

Figure 1. 171 Å image of QS corresponding to DS1, and the corresponding histogram of intensity.

Standard image High-resolution image
Figure 2.

Figure 2. Same as Figure 1 but for DS2.

Standard image High-resolution image

Table 1. Data Set Details

IdentifierDS1DS2
Start time2011-08-14 T00:00:002019-05-02 T00:00:00
End time2011-08-14 T08:00:002019-05-02 T08:00:00
Reference time2011-08-14 T00:00:002019-05-02 T00:00:00
Xcen, Ycen192'', 749''19.0'', 211.5''
FOVx, FOVy230'', 116''346.0'', 269.0''
InstrumentAIAAIA
Passband171,193,211171,193,211
Exposure normalizeTrueTrue
Cadence12 s12 s

Download table as:  ASCIITypeset image

To study the distribution of the intensity, for each pixel we create light curves of intensity in all three passbands. We plot sample light curves from both data sets, for one passband, and their corresponding distribution in Figures 3 and 4 for a single pixel. Note that we also show the time series of magnetic flux density of the corresponding pixel taken from the Helioseismic and Magnetic Imager (HMI; Scherrer et al. 2012) on board SDO corresponding to the AIA FOV. The line-of-sight (LOS) magnetograms are obtained by HMI at approximately 45 s cadence with a pixel size of 0.5''. We map the HMI data to the same plate scale as that of AIA. The error in HMI LOS measurements is estimated to be ±10 G (Yeo et al. 2014; Couvidat et al. 2016), and this is depicted in the figure as the black horizontal line. The time series for intensity and magnetic flux density are shown in panels (b) and (c). The distributions are shown in panels (a) and (d), respectively, which demonstrate the log-normal distribution, as was previously observed by Pauluhn & Solanki (2007) in Solar Ultraviolet Measurements of Emitted Radiation (SUMER) observations and Tajfirouze & Safari (2012) for AIA observations.

Figure 3.

Figure 3. Intensity and magnetic field intensity time series of 1 pixel from the FOV of DS1, and their corresponding distributions, as labeled. The 10 G noise level is indicated in (c) and (d).

Standard image High-resolution image
Figure 4.

Figure 4. Same as Figure 3 but for DS2.

Standard image High-resolution image

3. Noise Characterization of the Light Curves

The observed light curves, shown in Figures 3 and 4, have inherent noise, which is essentially dominated by photon shot noise. To mitigate this, while preventing over-smoothing (and thus averaging over real events), we use a procedure that we call Finding kneemo.

Finding kneemo is based on existing knee analyses performed in machine learning. Broadly, the goal of the algorithm is to monitor a performance metric against the free parameter, which, in our case, is the size of the smoothing window. The window size for which we observe a drastic improvement in performance metric is taken as the boxcar window. The change point is generally known as the knee.

The knee determination is extremely qualitative, though some methods exist that quantify this well (see, for example, Salvador & Chan 2004). In our analysis, we consider a random light curve for a pixel in our data set, along with its error time series, which is obtained from aia_bp_estimate_error.pro. We smooth the light curve using a boxcar of box size varying between 1 and 100, and obtain its modified signal-to-noise ratio (S/N). We then plot the obtained S/N against the box size in Figure 5, along with the asymptotes of the S/N, and find their point of intersection. This point is then selected as the boxcar window size. From Figure 5, we find that the asymptotes intersect at boxcar size of five time points. Therefore, we use this value to enhance the S/N. Note that we have run this analysis on several light curve within our data set and have found a consistent result. Thus, we performed boxcar averaging with a box size of five time points for all the light curves in our data set. An example plot with the original and the smoothed light curve is shown in Figure 6.

Figure 5.

Figure 5. S/N variation with boxcar window. The black dots represent the variation, while the red dotted lines represent the asymptotes. The vertical black line shows the approximate point of intersection (marked as blue)

Standard image High-resolution image
Figure 6.

Figure 6. Comparison of original and smoothed light curves by Finding kneemo.

Standard image High-resolution image

4. Methods

4.1. Nanoflare Model

The forward model employed here, as mentioned earlier, is the PSM. This is an empirical model based on two key observations, i.e., log-normal spatial and temporal distribution of QS intensities (Pauluhn et al. 2001), and power-law distribution of energies from flares to microflares (see, e.g., Aschwanden 2019). The algorithm may be summarized by asking the following questions:

  • 1.  
    What is the probability of a flare to occur at a given time step?
  • 2.  
    If a flare is meant to occur at the given time step, what would its peak energy be?
  • 3.  
    How long will the flare last once it has occurred, i.e., the duration of the flare?

In this model there are five free parameters: the event or flaring frequency (pf ), i.e., the probability of a flare to occur at a given time; the duration of the individual flare event (τ); the power-law slope (α) and the minimum (ymin) and the maximum (ymax) energy that is allowed for an individual flare events, which provides the bounds of the power law. An example simulated light curve, depicting the formation of light curve from individual events is shown in Figure 7. It may be seen from here that given only the envelope light curve, the individual events may not be inferred. We note that the simulations are performed over a large parameter space (see Table 2), while fixing ymin and ymax, to perform the inversion of light curves from all the three passbands with the same inversion model.

Figure 7.

Figure 7. An example showing light-curve generation using the nanoflare generation model, similar to Figure 2 of Pauluhn & Solanki (2007).

Standard image High-resolution image

Table 2. Simulation Grid Parameters Note that all parameters are in code units

ParameterRangeStep Size
pf [0.05, 0.95)Steps of 0.05
α [1.1, 3.0)Steps of 0.1
τ [1, 100)Steps of 2.0
ymax 0.3 
ymin 0.03 

Download table as:  ASCIITypeset image

We generate the light curves for a length of 5L + 1600, where L is the length of the light curve (2400 in our case), and reject 800 samples from either side to remove boundary effects. The remaining light curve is folded five times to get a final light curve of length L. 2 This is done to minimize the effects of the initial seed for random number generators. The observed light curves are normalized by their median values, following Bazarghan et al. (2008). Hence, the radiances reported from hereon have no associated units.

4.2. Inversion

We perform the inversion using 1D CNN. In this approach, generally, there are convolution layers followed by an activation layer. The activation, as we have seen in Section 1, is a nonlinear function that forms the core of complex learning ability of any NN. We use the function Elu as defined in Equation 1 as activation for all layers except the last, since Elu enables the network to train faster and generalize better (Clevert et al. 2015). For the final layer, we have no activation since this is a regression problem mapping to a continuous variable.

Equation (1)

A graphical representation of the model architecture is shown in Figure 8. As depicted in the figure, there are input/output layers, convolution layers (where we implicitly assume the activation function to be present) and fully connected (FC) layers. In the tensor shape, None is generally used to denote a variable size, which in this case represents the number of light curves to be inverted during a single forward pass. The convolution layers (marked in red) are given a four-dimensional shape, representing (height, width, input channel, and output channel), and the stride size is given by an integer. Note that channel here is not to be confused with AIA passbands, and that since we have used a 1D signal, the height is set to 1. After a suitable number of convolutions, the array is unrolled to 1D, and captured by FC layers (marked in yellow). For more information on general CNN and training see (Goodfellow et al. 2016, Chapter 9).

Figure 8.

Figure 8. CNN architecture used in this work. The blue boxes indicate input/output layers, and other colors indicate trainable layers. The tensor shapes are given in square brackets as [], and the number outside, for the convolution layers is the stride size. FC denotes the FC layers.

Standard image High-resolution image

4.3. Data Preparation and Training

We divided the simulated light curves into training set (80%) and testing set (20%). However, before feeding in any data to the CNN, it must be prepared appropriately to make sure all the parameters are of the same scale. The training set parameters are rescaled between 0 and 1, and the testing set parameters are rescaled using the training statistics, as is the standard procedure in machine learning. All light curves are also rescaled between 0 and 1.

We train the CNN by feeding in the training set light curves, and obtain a three free parameter set (pf , τ, and α). This obtained parameter set is them compared with the original target parameter set using an error metric, which is used to update the kernel values of the CNN. The error metric is the sum of the two terms defined as

Equation (2)

where x is the target parameter set, $\hat{x}$ is the predicted parameter set, and the summation is performed over all observations, and all parameters.

There are certain free parameters that need to be fixed while developing any NN. These free parameters are called hyperparameters, while the trainable parameters of the NN are called the weights of the NN. For training the CNN, we used the Adam optimizer (Kingma & Ba 2014), which is a stochastic optimization algorithm. The size of the update at each step is controlled by the hyperparameter called the learning rate.

Overfitting is a serious issue in NN training whereby the model starts fitting the noise in the model and stops generalizing. This may result in erroneous results and interpretation of inversion. To prevent overfitting, we use dropout (Hinton et al. 2012), which essentially switches off neurons randomly with a fixed probability for every forward pass. The training hyperparameters are summarized in Table. 3.

Table 3. Training Hyperparameters

HyperparameterValue
Cost function ${{ \mathcal L }}_{1}$(prediction,target)+${{ \mathcal L }}_{2}$(prediction,target)
OptimizerAdam optimizer with default values
Learning rate1e − 3
Dropout rate0.2
No. of iterations3000

Download table as:  ASCIITypeset image

4.4. Uncertainty Estimation

A CNN generates a single prediction for a given forward pass. However, in general, there exist two kinds of uncertainties—namely epistemic and aleatoric—associated with any such predictions (Kendall & Gal 2017). Epistemic uncertainty relates to model uncertainty due to unexplored weight space of the neural network. Aleatoric uncertainty relates to the inherent uncertainty in the target parameter values. In our study, there is no aleatoric uncertainty because the parameters are defined by us. Hence, we have only epistemic uncertainty.

Deficiencies in model training, resulting in unexplored weight space, can occur if not enough data is provided during model training. Hence, this effect can be simply minimized by increasing the size of training set. However, the epistemic uncertainty measure, while informing us about the deficiencies in fitting, may also inform us about any outliers in the data set. Throughout the analysis in this work, the PSM is assumed to be the ground truth, i.e., it fully describes the observed light curves to infer parameters. This is obviously never the case with any model. Hence, departures of the observations from simulations, where the PSM does not fully explain the given observation, would behave as outliers. Hence, higher uncertainties associated with the parameters inferred from the observations tells us either there are deficiencies in model fitting in certain regimes, or the light curve is not explained well by the PSM. However, we note that it is practically impossible to disentangle these effects (see, e.g., Kendall & Gal 2017).

The epistemic uncertainty may be estimates by application of dropout (Hinton et al. 2012; Díaz Baso et al. 2019). In addition to being used to prevent overfitting, Dropout can also be used to create perturbations, and obtain the variability in the predictions (Gal 2016). Since the neurons switched off in every forward pass are random, we perform a Monte Carlo forward pass to obtain multiple realizations of the CNN, and present the mean, standard deviation from the passes. Thus, the error bars reported for the estimated parameters of an individual light curve are the standard deviation obtained from dropout.

4.5. Inversion Performance

To asses the performance of our CNN, in Figure 9, we display scatter plots between the target and predicted values of pf (left panel), τ (middle panel), and α (right panel). As can be readily noted, the predicted values lie very close to the target values, thereby validating the performance of our network on a test set. This may be quantified using the coefficient of determination (R2) defined as

Equation (3)

where 〈 x 〉 represents the mean of target set, x = {xi } represents the target values, and $\hat{x}=\{{\hat{x}}_{i}\}$ represents the predicted values. In this case, i corresponds to number of points in the test set, i.e., R2 is computed separately for each target parameter. The R2 values are 0.990, 0.999, and 0.97 for pf , τ, and α, respectively, showing excellent performance of our network. Our network is now ready to be fed in with the observed intensity light curve.

Figure 9.

Figure 9. Correlation plot of target and predicted parameter values from our CNN, for pf (left panel), τ (center panel), and α (right panel). This quantifies the generalizability of the CNN from the training set.

Standard image High-resolution image

5. Results

Now we discuss the application of the network on the observed light curves. We first discuss the results obtained for a single light curve in Section 5.1. Then in Section 5.2, we move on to discuss the results obtained for all light curves obtained for all the three AIA passbands. We next explore the various correlations between our parameters and perform an analysis of the involved energetics in Section 5.3.

5.1. Application of the CNN on a Single Light curve

For representative purpose, we choose the intensity light curve for a random pixel from both DS1 and DS2 and obtain the corresponding simulated light curve. It is important to note that since PSM is a statistical model that generates a representation of the observations statistically, one should not perform a point-by-point comparison of the simulations with the observations. A simple change in the seed of the random number generator can change the exact times when events occur. Furthermore, since the amplitude of events is sampled from a distribution, the random seed value can also change the event amplitude at particular times. However, these seeds cannot change the overall statistical properties of the light curve. Thus, a comparison of the observation and simulation must be done using statistical properties of light curves (e.g., intensity distribution, frequencies showing enhances power), rather than a pointwise comparison of light curves. Once a good representative simulated light curve is obtained, the corresponding parameter set is taken to characterize the observed light curve.

In Figures 10 and 11, we show the comparison between observed (orange) and simulated (blue) light curves (panel (a)), kernel density estimation (KDE) of intensity distribution (panel (b)), the global Morlét power spectra (panel (c)), and the cumulative distribution function (CDF; panel (d)). Note that both the observed and simulated light curves are normalized by their median. The parameter sets denoted for the simulated light curves are the mean values of the obtained parameter distribution by performing 1000 Monte Carlo simulations and are denoted at the bottom of the figures.The KDE can be understood to be essentially a continuous extension of histogram (see, e.g., Chen 2017). Note that pf and τ are defined as per time step in Figure 9, and may be converted into real units as

and

The pf denoted henceforth is given as number of events per minute, while τ is given as timescale in minutes. The α remains a dimensionless parameter.

Figure 10.

Figure 10. The comparison of a representative light curve obtained for 171 Å passband from DS1 with a simulated light curve. Observations are shown in translucent orange and simulations are shown in blue. Panel (a): normalized observed and simulated light curves; panel (b): KDE of observed and simulated light curves; panel (c): global Morlét power for observation and simulations; and panel (d): CDF comparison of simulation and observation.

Standard image High-resolution image
Figure 11.

Figure 11. Same as Figure 10, but for DS2.

Standard image High-resolution image

From panel (a) in Figures 10 and 11, we can see an excellent statistical correspondence between the observed and simulated light curve. This is corroborated by the match in their corresponding KDEs (panel (b)) and CDFs (panel (d)). Furthermore, the Morlét wavelet power spectrum shows peaks at corresponding frequencies for both the observed and simulated light curve. These results confirm that the inversion model was able to learn both the time series and distribution properties corresponding to the three free parameters. We further emphasize that value of α inferred in these two cases is ≥ 2, which in turn suggests that events with smaller energy are dominantly responsible for generation of radiance of these particular examples.

From Figures 10 and 11, we find a clear relation between the goodness of representation of simulated light curve (using the CDF and Morlét power) and the spread of parameters obtained by Monte Carlo simulations. Consider the percentage uncertainty (i.e., uncertainty/mean value)—we find this quantity is approximately 3% in pf for DS1, 6% in pf for DS2; ∼5% in τ for DS1 and ∼4.5% DS2; and ∼12% for DS1 and ∼15% for DS2. Thus, the inversion model is more certain of the parameters of DS1 than DS2, which is also reflected in the relative mismatch of Morlét power between the observation and simulation for DS2 over DS1, at the first two peaks (note the difference in y-axis limits in panel (c)). Thus, such an uncertainty measure, along with the Monte Carlo forward pass can help us explain which parameters are strongly influencing the quality of a given inversion, assuming PSM as the ground truth.

5.2. Multi-light curve—Multi-passband Analysis

Since our network gives reliable results for the light curve obtained for a random pixel in both data sets, we now take all of our light curves (331,967 light curves per passband), for the three AIA passbands, and pass them through the network to obtain the relevant parameter set for each light curve. Due to operational constraints, we perform only 100 Monte Carlo forward passes in this case. We emphasize that the obtained parameter set for 100 and 1000 Monte Carlo forward passes are statistically the same for a limited handpicked set of representative light curves.

For both of our data sets, we first divide each light curve by its median value, rescale between 0 and 1, perform the Monte Carlo forward pass through the CNN, and obtain the mean parameter set. Finally, we concatenate the parameter set across the whole FOV for both data set separately for each AIA passband to improve our statistics. This concatenation can be done since all light curves are from QS regions and are evolving independently. Figure 12 displays the distribution of flaring frequency pf (panel (a)), duration τ (panel (b)), and power-law slope α (panel (c)) for this concatenated data set. The solid blue curves are for the 171 Å observations, the red dashed–dotted curve is for the for 193 Å observations, and the green dashed curve is for the 211 Å observations.

Figure 12.

Figure 12. Distribution of inferred parameter set for pf (panel (a)), τ (panel (b)), and α (panel (c)) over both data sets. The colors are distributed as blue (171 Å), red (193 Å), and purple (211 Å).

Standard image High-resolution image

The plots reveal that the distribution of all three parameters for all passbands are remarkably similar. The pf distribution peaks at ∼2.2 events per minute for 171 Å and at ∼2.4 events per minute for 193 Å and 211 Å, with a range of values between one and four events per minute. The distribution of τ peaks near 12 minutes for 211 Å, 14 minutes for 193 Å, and 16 minutes for 171 Å, implying a slight temperature dependence. However, we emphasize that since the AIA passbands are multithermal, this inference should be taken with caution.

The distribution of α, plotted in panel (c), ranges between 1.0 and 8, with the peak at ∼2.3 for all three passbands. When we consider the whole set of light curves, we find that ∼62% of light curves in 171 Å have α ≥ 2, while the fraction is ∼61% and ∼54%, respectively, for the 193 Å and 211 Å passbands, respectively. Thus, we find a reduction in dominance of lesser energy events as we progressively probe the QS in hotter passbands. We also note that the falloff for all three parameters goes from the hotter passband (i.e., 211 Å) falling off first, followed by progressively cooler channels (193 Å and 171 Å, respectively).

To further understand any peculiarities exhibited by the QS for the two domains of α, we plot the variation of pf τ with α in Figure 13. To boost the S/N, we have averaged pf τ within a constant bin of α. Note that the bin size for averaging is obtained from Doane (1976). The error bars shown are 3σ standard error. 3 Following Pauluhn & Solanki (2007), the factor pf τ may be interpreted as the ratio of excitation (i.e., intensity generation by pf ) to damping (i.e., intensity dissipation by τ) for a given pixel. Here, we investigate the dominance of one over the other.

Figure 13.

Figure 13. Variation of pf τ with α. The vertical dashed line marks α = 2. The error bars are 3σ standard errors.

Standard image High-resolution image

The plot reveals that there is almost no change in the excitation to damping ratio for α < 1.8 and is independent of α in this regime. However, for α ≥ 1.8 the dynamics change, and we find a reduction in the ratio with increasing α. The larger error bars at the end are due to poor statistics. But even considering the variation until α = 4, we find a considerable reduction in the ratio with α, presenting an increasing dominance of damping over excitation. Thus, there is either a reduction in excitation, or an increase in damping, or both, which comes into play once the smaller events start dominating radiance generation.

5.3. Energetics

With the parameter set obtained, we can now investigate the involved energetics. A simple way to understand the energetics is by comparing the behavior of the slope α vis-à-vis the other free parameters. A large slope implies a predominance of smaller energies, while a small slope implies predominance of larger energies.

To quantify relations in terms of the peak intensity of a nanoflare, Pauluhn & Solanki (2007) defined the average peak nanoflare radiance (Eavg) as

Equation (4)

Eavg is a measure of the average peak nanoflare radiance value for a given time series. Using Equation (4), we estimate the values of Eavg for each pixel and study its relationship with flaring frequency (pf ) as well as flare duration (τ), through scatter plots between these parameters.

In Figure 14, we plot pf (top panel) and τ (bottom panel) as a function of Eavg for all three wavelengths. Note that we have averaged the free parameters within a constant bin of Eavg, and our error bars are 3σ standard errors.

Figure 14.

Figure 14. Inferred dependence of free parameters (binned) on Eavg over different passbands. Blue color represents 171 Å, red 193 Å, and green 211 Å. Top: pf is plotted per Eavg bin; bottom: τ is plotted per Eavg bin. The error bars indicate 3σ.

Standard image High-resolution image

The plots reveal a slight tendency of decreasing pf as a function of Eavg, albeit there is sharp decline in the beginning. However, τ monotonically increases with increasing Eavg until about Eavg = 0.085 and shows saturation thereafter. Moreover, the plot further suggests a systematic lowering of flaring duration, being the largest for the coolest passband (171 Å).

6. Summary, Discussion, and Conclusion

The QS coronal region provides a wealth of data to narrow down and understand the underlying heating processes. Assuming the underlying heating mechanism to be impulsive, coronal light curves may be approximated using empirical statistical models to infer physics-inspired parameters. To this end, we construct an inversion model of the PSM using CNN and validate it using a separate test set. We perform the inversion on two coronal data sets and infer the free parameters for the three AIA passbands, viz. 171 Å, 193 Å, and 211 Å.

We find that the light curves inverted using the CNN and observed light curves are in excellent statistical agreement, considering the CDF and KDEs (see Figures 10 and 11). Note that we are not concerned with a point-to-point match at each time step between the simulation and the observation. A change in the seed of the random number generator is enough to shift the location of individual events and their amplitudes in simulation. However, this does not cause any changes in the statistical properties of the light curve, which is what we are primarily interested in. The global wavelet power shows the simulations and observations to have peaks at similar frequencies, validating that the simulation and observation both have enhanced power in similar frequencies. The quality of approximation may be understood by the epistemic uncertainty of our CNN, obtained by the application of dropout to perturb the model.

We find the distribution of all parameters to be similar for light curves from all three passbands. The flaring frequency lies within the range of 1–4 events per minute, with a peak at ∼2.3 events per minute for all three passbands. Similarly, the flaring duration has a range of values between 5 and 30 minutes. However, unlike the flaring frequency, the peak of the distribution of flaring duration shows temperature dependence, being the highest (∼16 minutes) for the coolest 171 Å channel, ∼14 minutes for 193 Å, and lowest (∼12 minutes) for the hottest 211 Å channel.

The power-law index α has a range of values between 1 and 8. The distribution of α peaks at 2.3 for all three passbands of AIA, viz. 171 Å, 193 Å, and 211 Å. This finding strongly suggests that nanoflare heating is indeed a viable source of energy in the quiet corona (see, e.g., Hudson 1991). We also find the fraction of light curves giving α ≥ 2 progressively reduces from cooler to hotter passbands. We further note that there are a significant number of pixels where α < 2. This is suggestive that low energy events may not be dominant everywhere. However, note that the viability of these low energy events also relies on the flaring frequency pf , requiring a full analysis of the energetics.

Our finding further suggests that there is a change of dynamics for pixels with α < 1.8 and α ≥ 1.8 (from Figure 13). In the former case, pf τ is nearly constant, while in the latter case, it reduces with increasing α. This may also be observed in Figure 14. Note that α = 2 corresponds nearly to Eavg = 0.08 and the right tail of Eavg represents α < 2. Here, a definite increase of pf with decreasing Eavg is seen, which is interpreted as excitation increasing with decreasing Eavg (and thus increasing α). However, we also find that τ reduces with reducing Eavg. Thus, the increase in damping counters that in excitation in the α ≥ 2 regime, causing a reduction in the ratio in Figure 13. Thus, the increase is excitation is essentially nullified by the increase in damping.

We note that our inferred pf distribution peaks at ∼2–3 events per minute, while Tajfirouze & Safari (2012) obtained a mean pf of ∼0.33 events per minute. This may be explained by the better temporal cadence and spatial resolution of our data, whereby our simulation captures much smaller transient events. However, note that our inferred pf corresponds to an average waiting time of ∼30 s, which is much smaller than the waiting times of ∼230 s suggested from simple geometric arguments (Klimchuk 2015).

There is a discrepancy in the τd (decay time) and α derived here with those obtained by Tajfirouze & Safari (2012). The τd being about a factor of 6 smaller in our case (∼60 minutes in Tajfirouze & Safari 2012 to ∼10 minutes in our case) and α peaking near 2.3 in our case, compared to a mean α of 2.6 in Tajfirouze & Safari (2012). This discrepancy may be attributed to the fact that the data used in this study are at much higher spatial and temporal resolution than in Tajfirouze & Safari (2012). Therefore, we must have captured smaller events with much shorter timescales (as also alluded to by Tajfirouze & Safari 2012). Similarly, the discrepancy in power-law index α may be attributed to the high spatial resolution data used in the present work. Moreover, in our simulation, we are sampling flares within a larger energy range with larger ${y}_{\max }/{y}_{\min }$ than those by of Tajfirouze & Safari (2012). However, note that our obtained cooling timescales (∼600 s) are indeed of the order of cooling timescales in the corona obtained by Klimchuk (2015) (∼1000 s).

Next, we find that the pf decreases with Eavg (see the top panel in Figure 14). The decrease of pf with Eavg is interpreted as a decrease in peak energy released per flare with increasing frequency. Thus, we can either have intermittent, high energy events or sustained low energy events. This variation of pf with Eavg is similar to the observation of the relation obtained between peak flare flux and waiting times by Hudson (2020) (see also Sarkar et al. 2019). Furthermore, the inverse relation between pf and Eavg, (or a direct relation between the waiting time and succeeding nanoflare energy) was a necessary ingredient needed to reproduce the observed Emission Measure distribution with temperature in ARs by Cargill (2014). Since the rise time is given as a fraction of the decay time in this setup, we do not distinguish between pre-flaring and post-flaring times. Thus, this may point to the presence of a reservoir of energy that may be exhausted by frequent, small-energy events or intermittent, large-energy events. However, we emphasize that the change in pf is tiny (2–3 events per minute, as can be seen from Figure 14) when compared to the total timescale of these events (10–15 minutes across all passbands). Thus, we must view this interpretation with caution.

The timescale τ is seen to increase with Eavg (see the bottom panel in Figure 14). This shows an increase in flare timescale corresponds to an increase in average flare energy. This weak correlation is similar to the weak correlation observed between peak flux and flare timescale by (Veronig et al. 2002, see Figure 3). However, we may seek to explain this relation qualitatively as below.

For an isothermal, optically thin plasma, the observed intensity is directly proportional to electron number density (O'Dwyer et al. 2010), i.e., ${DN}\propto {n}_{e}^{2}$. From Cargill (1994), we find that

Equation (5)

and

Equation (6)

where τc and τr are conductive and radiative cooling times, respectively.

Combining the equations of timescale and intensity, we obtain

Equation (7)

Thus, for a conduction cooling dominated plasma, the timescale τ should increase with the emitted intensity, while for a radiative cooling dominated plasma, we expect the opposite. Our results show a direct relation between τ and the peak flare intensity Eavg and qualitatively suggest that in such events, conduction losses are dominant over radiative losses, assuming a constant flaring frequency. This is similar to the results obtained by Gupta et al. (2018), Subramanian et al. (2018), and Rajhans et al. (2021) for tiny transient brightenings.

Finally, the flaring timescales are seen to be largest in the coolest passband and are seen to decrease from the cooler to hotter passbands. This indicates the decreasing dominance of conduction loss over radiative loss (but the conduction loss still dominates), as would be the case for cooling loops (see, e.g., Klimchuk 2006; Viall & Klimchuk 2012; Tripathi et al. 2009; Gupta et al. 2015). We emphasize that this is true under the assumption of constant flaring frequency since only ∼2–3 events are happening per minute, whereas our total (rise+fall) time in consideration is ∼15 minutes.

It is important to emphasize that the PSM is very well suited for explaining the observed light curves from the QS region obtained at much higher temporal and spatial resolution than was initially studied by Pauluhn & Solanki (2007). We note that although we have improved the inference by quantifying the uncertainties, the PSM may further be developed by incorporating the plasma filling factor and effective area in the forward model, as has also been suggested in the original paper (Pauluhn & Solanki 2007). Note that in the PSM, the radiance distribution is considered to be same as energy distribution. However, note that there may be a number of further smaller events, which may or may not produce detectable signatures in the intensity images. It is also possible that many events produce a signature in one passband, and not in another. Therefore, the distribution reported using the radiance is just a lower limit of the total number of events. Hence, incorporation of filling factors in the PSM is an important next step in improving the model.

Moreover, note that we have fixed the ${y}_{\max }/{y}_{\min }$, limiting the energy range considered for the simulation. Therefore, it would be prudent to keep this ratio as a variable quantity. We have tried to develop an inversion model with varying ymin over three orders of magnitude—however, the inversion model did not give adequate performance (the R2 for α was ∼0.88 on the testing set). Thus, introducing the radiance ratio as a free parameter creates a large degeneracy in parameter space (already having been hinted to by Pauluhn & Solanki 2007), which cannot be solved by our neural network or the optimization scheme. Hence, further modeling and training strategies are required to disentangle such highly degenerate parameter spaces. Furthermore, it has been reported that the distribution of flare waiting times (inverse of flaring frequency) follow a time-dependent Poisson process, which gives rise to a power-law tail of waiting time distributions (Wheatland & Litvinenko 2002). Thus, the assumption of a constant flaring frequency (which translates to exponential waiting time distribution) may also need to be revised in the PSM.

It is important to note that a lot of the magnetic field strength is below the 10 G limit (see Figures 10 and 11. However, the variability in the magnetic field strength suggests that the corona is extremely dynamic even in the quiet regions. Hence, for any study to characterize the heating of quiet corona, we need magnetic field measurement with much higher sensitivity and at greater resolution.

Finally, note that while our current analysis has been focused on the QS corona, Bazarghan et al. (2008) have applied PSM to AR corona successfully, although using data at a very low spatial and temporal resolution. Thus, we expect our results to also extend to AR corona. Furthermore, it would be interesting to apply our inversion model to chromospheric observations (see also Jess et al. 2014). This may be performed on spectral lines with the Interface Region Imaging Spectrograph, or using chromospheric observations in near-ultraviolet from the Solar Ultraviolet Imaging Telescope (Tripathi et al. 2017; Ghosh et al. 2016), on board the Aditya-L1 mission (Seetha & Megala 2017) of the Indian Space Research Organization.

We sincerely thank the referee for constructive comments and suggestions. This work is partly supported by the Max-Planck Partner Group of Max-Planck Institute for Solar System Research, Göttingen, Germany. We acknowledge the use of data from AIA and HMI. AIA is an instrument on board SDO, a mission for NASA's Living With a Star program. We would like to sincerely thank Dr. A. Pauluhn for sharing the IDL codes of the PSM, which was rewritten and parallelized in Python. The authors also thank Prof. Ranjiv Misra of IUCAA for discussion and comments on degeneracy in error surface and scale-free processes.

Software: The analysis was primarily performed in Python, using the TensorFlow (Abadi et al. 2016) package for machine learning—data download alone was done using SolarSoft (Freeland & Handy 1998). Computational tools used were NumPy (Van Der Walt et al. 2011; Oliphant 2006; Harris et al. 2020), SciPy (Virtanen et al. 2020), and Astropy (Price-Whelan et al. 2018; Astropy Collaboration et al. 2013). Parallelization was performed using Multiprocessing (McKerns et al. 2012), and plotting using Matplotlib (Hunter 2007) and Seaborn (Waskom & the Seaborn Development Team 2020). Most of the code was written in the Jupyter environment (Kluyver et al. 2016). The codebase is under expansion for future work and shall be made available on request.

Footnotes

  • 1  
  • 2  

    Folding essentially divides up the light curve in five equal chunks of length L and gets the average curve from these chunks.

  • 3  

    Standard error is defined as $\sigma /\sqrt{N}$, where σ is the standard deviation in the sample present in the bin, and N is the number of points in the sample.

Please wait… references are loading.
10.3847/1538-4357/abf65a