The following article is Open access

SuNeRF: 3D Reconstruction of the Solar EUV Corona Using Neural Radiance Fields

, , , , , , , , , , and

Published 2024 January 23 © 2024. The Author(s). Published by the American Astronomical Society.
, , Citation Robert Jarolim et al 2024 ApJL 961 L31 DOI 10.3847/2041-8213/ad12d2

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

2041-8205/961/2/L31

Abstract

To understand its evolution and the effects of its eruptive events, the Sun is permanently monitored by multiple satellite missions. The optically thin emission of the solar plasma and the limited number of viewpoints make it challenging to reconstruct the geometry and structure of the solar atmosphere; however, this information is the missing link to understand the Sun as it is: a 3D evolving star. We present a method that enables a complete 3D representation of the uppermost solar layer (corona) observed in extreme ultraviolet (EUV) light. We use a deep-learning approach for 3D scene representation that accounts for radiative transfer to map the entire solar atmosphere from three simultaneous observations. We demonstrate that our approach provides unprecedented reconstructions of the solar poles and directly enables height estimates of coronal structures, solar filaments, coronal hole profiles, and coronal mass ejections. We validate the approach using model-generated synthetic EUV images, finding that our method accurately captures the 3D geometry of the Sun even from a limited number of 32 ecliptic viewpoints (∣latitude∣ ≤ 7°). We quantify the uncertainties of our model using an ensemble approach that allows us to estimate the model performance in the absence of a ground truth. Our method enables a novel view of our closest star and is a breakthrough technology for the efficient use of multi-instrument data sets, which paves the way for future cluster missions.

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 4.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

The outermost layer of the solar atmosphere, the corona, is highly structured by its magnetic field. Observations in extreme ultraviolet (EUV) light allow us to study this structure and its evolution in the low corona in an unprecedented way. Several space missions image the solar atmosphere in specific wavelength bands in an effort to understand the Sun and its effects on Earth. The Solar Terrestrial Relations Observatory (STEREO; Kaiser et al. 2008) consists of two satellites on slightly displaced 1 au orbits, which cause a steady drift of 22fdg5 yr−1 of the two satellites relative to Earth. By adding the Solar Dynamics Observatory (SDO; Pesnell et al. 2012), which observes the Sun from Earth's perspective, the Sun can be simultaneously observed from three viewpoints.

On short timescales, the magnetic field evolves into highly structured localized regions (active regions; van Driel-Gesztelyi & Green 2015) that give rise to solar eruptions and ejections of plasma into interplanetary space. Both are a direct risk for our space assets (Temmer 2021). The scientific understanding of preeruptive structures is key for advancing our ability to predict eruptions, and yet it remains elusive (e.g., Patsourakos et al. 2020). This was one of the motivations and early science targets of the STEREO mission (e.g., Aschwanden et al. 2009). With two vantage points, the 3D triangulation of specific spatial points is possible from imaging data (Inhester 2006; Aschwanden et al. 2008; Bemporad 2009; Liewer et al. 2009). However, for the optically thin medium of the solar atmosphere, assigning observed intensities to a specific spatial point leads to ambiguous results due to the effects of integration along the line of sight (e.g., Aschwanden 2011). In other words, the intensity at each pixel is the integral of all emitting and absorbing plasma along the line of sight. Consequently, the idealized view of a single emitting or solid point is insufficient to reconstruct the solar atmosphere. Tomographic reconstructions provide a different approach by reconstructing the 3D temperature and density profile of the solar EUV corona utilizing multiviewpoint observations and the solar rotation (Frazin et al. 2009; Vásquez et al. 2009; Kramar et al. 2014; Cho et al. 2020). This approach has typically a temporal resolution of about 14 days and is not suited to study regions characterized by fast dynamics (Vásquez 2016).

On long timescales, the Sun evolves over an 11 yr cycle, which is distinctly marked by global changes (Hathaway 2015). During solar minimum, the magnetic field is mostly poloidal, with a low number of active regions clustered near the solar equator. In this period, coronal holes, which enclose open field line configurations, are present at the solar poles (Cranmer 2009) and can extend down to low latitudes. With the transition to the maximum of the cycle, and the corresponding reconfiguration to a toroidal magnetic field, active regions emerge at high latitudes, and the polar coronal holes vanish. During the maximum phase, mostly small coronal holes appear at low latitudes, and the number and complexity of active regions drastically increases. These phenomena are a direct effect of the solar dynamo and evolving magnetic field.

To relate both small- and large-scale changes to the physical processes that generate the magnetic field of the Sun and, consequently, other Sun-like stars, a complete picture of the Sun and its atmosphere is required. A frequently applied approach is synoptic maps, which use observations from only one instrument and make use of the solar rotation period of 27 days. Here, an ∼1° central slice is consecutively extracted to complete the image over the course of one full solar rotation. The observed slices are then reprojected onto the surface of a sphere to obtain a 2D image (Figure 7). A similar approach results in synchronic maps, which combine simultaneous (or nearly simultaneous) observations from different satellites located at different positions into a single image of the Sun (e.g., Caplan et al. 2016).

Both approaches have significant shortcomings that primarily affect extended and/or long-lived structures due to their temporal evolution and line-of-sight overlapping (e.g., a lifetime of days–weeks for small active regions; van Driel-Gesztelyi & Green 2015). Recent space-based observations originate in the ecliptic plane (latitudes of approximately ±7° or less) and are strongly limited in terms of the number of simultaneous observations. These constraints have consequences for the spherical assumption, where extended features (e.g., coronal loops) become warped toward the solar limb. While there are observing periods where full coverage of the entire Sun exists (2011–2014), the observations of the poles are limited due to the limiting viewing angle of the ecliptic-orbiting telescopes. Only the combination of simultaneous imaging from three (or more) viewpoints can provide the maximum information possible for this difficult-to-access location.

To address these challenges, we devised a novel approach to provide a complete 3D reconstruction of the global solar EUV corona. Our method builds upon a state-of-the-art deep-learning method for 3D scene representations (Neural Radiance Fields, NeRFs; Mildenhall et al. 2021). The goal is to reconstruct the 3D geometry of a scene from a set of images and their known viewpoints. For a sufficiently large data set (e.g., about dozens of images), the underlying 3D representation can be recovered, and images from any viewpoint can be rendered. We adapt this method to match the physical reality of the Sun using a simplified radiative transfer approach and a geometric sampling approach to create a solar radiance field, Sun Neural Radiance Fields (SuNeRFs; see Section 2), that accounts for the line-of-sight ambiguity of an optically thin source (see Bintsi et al. 2022). In addition, we account for insufficient coverage and dynamic changes by including a time component into our model.

SuNeRFs can be deployed to consistently combine observations from multiple vantage points into a 3D representation and then render observations at arbitrary viewpoints. In addition, this approach goes one step further and directly enables height estimates of coronal structures, solar filaments, coronal hole profiles, and solar eruptive events.

2. Method

To reconstruct the 3D solar atmosphere, we consider both simulated data and observation sequences from SDO/Atmospheric Imaging Assembly (AIA) and STEREO/Extreme Ultraviolet Imager (EUVI; A and B) in specific wavelength channels. Our goal is to obtain a scene representation from these image sets and known viewpoints (see Mildenhall et al. 2021). From the resulting representation, we can render novel viewpoints, extract height information, and analyze the plasma distribution in terms of emitting and absorbing material.

2.1. Simulated Data

To validate our method, we leverage a magnetohydrodynamic simulation, developed by Predictive Science Inc., of the solar corona used to forecast the state of the solar atmosphere prior to the 2019 July 2 total solar eclipse (see the website 10 and Boe et al. 2021, for more details). The data set consists of a single snapshot from which we extract full-disk images from 256 viewpoints at a resolution of 1024 × 1024 pixels (2farcs63 pixel–1) in three different wavelengths (171, 193, and 211 Å). We preprocess the simulated data analogously to the observational data introduced in Section 2.2.

2.2. Observational Data

From 2010 to 2014, three satellites capable of imaging in similar EUV channels were orbiting the Sun, namely, the SDO equipped with the AIA instrument (Lemen et al. 2011) and the twin STEREO spacecraft, each equipped with an EUVI (Wülser et al. 2004). All three satellites orbit in the ecliptic plane, with SDO in geosynchronous orbit and the STEREO satellites progressively advancing (STEREO-A) and trailing (STEREO-B) Earth. When the separation angle between SDO and each of the STEREO spacecraft exceeded 90°, it allowed a view of the full Sun at one instant for the first time.

An important issue in the context of this project is the design, performance, and operational differences among the various instruments. We apply Instrument-To-Instrument translation (ITI; Jarolim et al. 2024) to translate full-disk observations from STEREO to the SDO domain. With ITI, we can account for differences in resolution, calibration, and filter bands between the individual instruments. The AIA and EUVI instruments have common observing wavelengths (171 and 304 Å). For the wavelengths 195 and 284 Å, we use ITI to approximate the 193 and 211 Å channels of AIA (see Jarolim et al. 2024). Although labeled as single wavelength channels, the SDO/AIA and STEREO/EUVI spectral channels actually result from spectral integration as specified by the spectral response of the instruments and wavelength filters (Szenicer et al. 2019). The multichannel translation of ITI is a prerequirement to have an appropriate approximation to combine homogeneous observations of this channel. The resulting homogenized data set serves as approximation of multiviewpoint SDO/AIA observations in four spectral bands (171, 193, 211, and 304 Å).

We resample all data to 1farcs2 pixel–1. The focal length f, as defined by our ray tracing method (discussed below), is then accordingly computed individually for each observation,

Equation (1)

where W refers to the width of the image. Note that this focal length is computed from the satellite position and observed image and is not directly related to the actual focal length of the instrument due to the adjustment of the image size and resolution. Furthermore, we only consider square images. In this context, we analyze the propagation of diverging rays from the observer, with the parameter f specifying the angle of view. This approach departs from the assumption of small angles, where rays would be considered parallel. For each observer, we specify the time of the observation, the distance from the Sun, and the Carrington longitude and latitude.

2.3. Model Training

We adapt the approach from Mildenhall et al. (2021). We use a neural network to represent the 3D scene and ray tracing methods for learning the representation. Information on the entire scene is stored in the weights of the neural network, rather than having an explicit grid representation. To be precise, we train a neural network for each individual scene, in contrast to classical deep-learning applications, which learn to perform a general task.

For each image pixel, we determine the light ray path using known information from the observer position and viewing angle. Along this path, we sample points (x, y, z) within the simulation volume and use classical ray tracing methods to compute the integrated pixel value. In the NeRF approach proposed by Mildenhall et al. (2021), each point is associated with a color and density. The pixel value is computed by the weighted sum of the colors of the sampled points along the ray, where the weights are the cumulative sum of the density values. In other words, this approach computes the likelihood of a ray being absorbed along the line of sight. The training images, together with the observer position, serve as the target for model training. By iteratively updating the neural network to match the pixel values of the training images, we obtain the underlying scene representation. This approach has shown the ability to obtain photorealistic images at interpolated positions (Mildenhall et al. 2021; Li et al. 2022).

In contrast to a classical NeRF, where the volume is composed of colors and densities, the solar atmosphere consists of emitting and absorbing plasma. In addition, NeRFs require dozens of images in order to learn the scene representation, while only three simultaneous observations are available for observations of the solar EUV corona. Therefore, we apply four primary changes to the NeRF approach. (1) We replace density and color with wavelength channel–specific emission and absorption. These capture how plasma with temperatures observed by a given wavelength channel emits and absorbs light. Each pixel value refers to the total intensity that is observed along the line of sight for a given wavelength channel and is expressed in units of data number per second (DN s−1). (2) We utilize the temporal evolution and solar rotation to obtain more viewpoints (see Frazin et al. 2005). From this approach, we obtain a full scan of the Sun per observing instrument within one solar rotation (∼27 days). For the definition of viewpoints, we consider all observations in heliographic Carrington coordinates, which are solarcentric. We explicitly include time in the query coordinates of the NeRF to account for a consistent treatment of dynamic changes in the solar atmosphere. (3) We account for the known geometry of the Sun by only sampling points up to the solar surface. With this, we avoid an invalid placement of emitting plasma from below the solar surface. The change to monochromatic intensity values, instead of colors, further implies that there is no need for a background (see Mildenhall et al. 2021). (4) An optional component for NeRFs is a viewing-angle dependence of the pixel color. For our implementation, we do not consider any view dependence of the observed intensities.

In Figure 1, we provide an overview of our approach. For our 3D representation, we use a fully connected sinusoidal representation network (Sitzmann et al. 2020) with eight layers, where each contains 512 neurons, resulting in about 2M free parameters. Our model takes four input coordinates (x, y, z, t) and outputs the estimated emission epsilon(x, y, z, t) and absorption κ(x, y, z, t) coefficient for a given wavelength. We use an exponential and ReLU activation (setting negative values to 0) for the emission and absorption coefficients, respectively. With this, we obtain positive values for both coefficients. We then compute the total emission E at each point by multiplying with the spacing of the sampled points ds,

Equation (2)

Each emitted intensity value propagates through the line of sight to the observer, where absorption can reduce the intensity value at points between the observer and the point of emission. We model absorption values between [0, 1], where 1 refers to full transmission and 0 to total absorption. We compute the absorption A for each sampled point based on the spacing,

Equation (3)

The total intensity I is then computed as the sum of the absorption reduced emissions,

Equation (4)

where i and k refer to the points along the sampled ray and N refers to the total number of sampled points along the ray. Therefore, each emission E is reduced by all absorption values A in the range from the emitting point to the observer. We note that the first point is always transmitted without absorption (A1 = 1), and the last sampled absorption point has no effect.

Figure 1.

Figure 1. Overview of the proposed method for 3D reconstruction of the solar EUV corona. Our method consists of five steps. (1) For each pixel in the input video sequence, we sample points along the ray path ranging from −1.3 to 1.3 R. The endpoints of the rays passing through the Sun are fixed to the solar surface to account for its opaqueness. (2) The sampled points are passed through the neural network, which outputs the emission epsilon and absorption κ coefficient per point. (3) For each ray, we construct the emission and absorption along the ray. (4) The total observed intensity is then computed by integrating all sampled points, where each intensity value is reduced by absorption along the path of propagation (from the origin to the observer). (5) The predicted intensity value is then compared to the actual pixel value, which serves as a loss function for our model training. Per update step, we optimize a set of 32,768 rays. By iteratively fitting all pixel values, we obtain a complete spatial and temporal representation of the solar corona.

Standard image High-resolution image

For all computations, the spatial scales are normalized to 1 R. We use positional encoding, which implies periodic boundary conditions at 2π. Therefore, we scale the coordinates to units of 2 R prior to the encoding (simulation volume of [−2π, 2π] R) to avoid reflections at the boundary.

For training, we adjust the pixel-value range based on the average maximum value (see Jarolim et al. 2024) and use an asinh stretch to mitigate the contributions by bright pixels. To account for this scaling, we use the same stretch function when computing the final pixel values (Iscaled) of our rendered images,

Equation (5)

with a = 0.05.

We apply two sampling strategies to prioritize points from the most important regions. (1) We only sample from −1.3 to +1.3 R with respect to the solar center, such that the default sampling range spans 2.6 R. In addition, we stop the sampling where light rays intersect the solar surface (1 R), such that a ray that is pointing at the center of the solar disk is sampling a range of 0.3 R. Therefore, the total line of sight is dependent on the viewing direction. We note that this sampling is independent of the observer distance. (2) Similar to Mildenhall et al. (2021), we use a coarse and a fine network. For the coarse network, we sample 64 points uniformly distributed along the ray. The emission values of the coarse network are then used as a probability distribution to sample 128 additional nonuniformly distributed sampling points. These points and the 64 uniformly sampled points are then used as input in the fine network, which serves as the primary training objective. For all our evaluations, we use the outputs of the fine network.

For our model training, we select random rays from the full set of image pixels, where the number of rays per batch is adjusted to the available computational resources. For our reconstructions, we consider each wavelength separately. Therefore, our model is trained with observations from a single wavelength channel.

We initiate the model training with a centered crop of the image, where we use pixels within [−1000'', 1000''] in helioprojective coordinates. We train on this data set for one epoch (∼100,000 iterations) to prevent divergence at the beginning of the model training and emphasize the reconstruction of the solar disk in the initial phase of training. Model training is then continued with the full field of view of the instruments.

When training for smaller fields of view (subframes), we use a model trained on the full field of view as an initial starting point. This limits divergences and already incorporates the global geometry of the Sun. In the second training step, we crop the subframes at a fixed coordinate point from the full-disk observations, which serves as our new training set. Therefore, pixels outside of the region of interest are neglected and no longer suitable for further analysis after fitting the model to the subregion. The reduced spatial information that the model needs to fit enables the increase of temporal and spatial resolution (Section 3.4).

During model training, we separate one image of the training set to validate performance. We monitor the peak signal-to-noise ratio (S/N), mean squared error (MSE), and structural similarity index (SSIM; Wang et al. 2004). For all our model runs, we found a monotonic performance increase and a similar performance independent of the initialization. We note that starting with a field of view that is too large can result in irreversible divergence within the first epoch.

During each training step, we optimize both the coarse and fine models to match the target pixels. We use MSE as a loss function.

We introduce an additional regularization of the modeled absorption that reduces nonphysical results above the poles. Viewpoints from the ecliptic provide no observations of the solar surface and therefore only model low-intensity values. This allows for arbitrary absorption values at these regions. We add an additional loss factor ${{ \mathcal L }}_{\mathrm{regularization}}$ to our optimization that suppresses absorption values κ above 1.2 R to more strongly penalize large absorption at greater heights,

Equation (6)

where d refers to the distance from the solar center in R. When training with full field-of-view images, we add the regularization to the MSE loss, whereas we set the regularization to 0 when training with subframes (i.e., no polar regions).

2.4. Model Evaluation

To evaluate our method, we compute the peak S/N and SSIM between our reconstruction and the ground-truth image. Here, we compare scaled images (see Equation (5)) to avoid bright pixels dominating our evaluation.

To provide an uncertainty estimation, we train five independent models with randomly initialized weights. To generate uncertainty maps, we render the same viewpoint per model and compute the standard deviation of the resulting set of images. Therefore, for regions that are less constrained by observations (e.g., single viewpoint), different solutions can be found. This degree of freedom is then reflected by our uncertainty metric.

Calculating height from imaging data is challenging, since there exists no exact spatial point that can be associated with an observed intensity value. We determine height information by computing the average distance from the solar center for all sample points along the ray using channel-specific emission as a weight. In other words, for each ray, we compute the height of the average emission value. This implies that dark regions (e.g., coronal holes, filament channels) appear as elevated structures, since the primary emission is coming from higher layers, and should not be considered for mapping the spatial location of the solar feature.

For the visualization of 2D slices of the solar atmosphere, we sample radial points along a given longitude. We pass each point through our network to obtain the respective channel-specific emission and absorption coefficients. For the emission, we use an asinh stretch, analogous to Equation (5).

As a baseline approach for rendering novel viewpoints, we use a simple reprojection. More specifically, we combine observations at a given time from STEREO-A and -B and SDO into a synchronic (Carrington) map. In the case of overlapping regions, we compute the mean value of the intersecting pixels. For missing regions (i.e., polar regions that are not covered by any instrument), we use the mean value over the entire image as a filling value. To synthesize novel viewpoints, we project the map onto a unit sphere and rotate the perspective to the new viewpoint.

This approach has several shortcomings. By projecting to a 2D map, we lose the off-disk information; consequently, this method cannot account for the faint emission that is observed over the solar limb. This further approximates the atmosphere as flat, neglecting the height information when projecting to new viewpoints. This is especially problematic for extended structures (e.g., coronal loops), where projection effects become especially prominent closer to the limb. Finally, combining overlapping regions is problematic. Due to the projection effects, the observations of the individual instruments do not align pixelwise in the synchronic map. Therefore, computing the mean of the overlapping pixels leads to blurred features.

Here, we only consider a basic approach as our baseline but note that there are more advanced data homogenization methods (e.g., Hamada et al. 2020) and image stitching methods (e.g., Caplan et al. 2016) available.

3. Results

We validate our method against simulations (Section 3.1) and observations (Sections 3.2 and 3.3). Additionally, we rely on our method's ability to capture the volume's temporal evolution to reconstruct a solar eruption at high temporal cadence (Section 3.4).

3.1. Validation with Simulated EUV Observations

To validate our method, we use the simulated data introduced in Section 2.1. For our model training, we consider each wavelength separately and use observations viewed from ∣latitude∣ ≤ 7° from the ecliptic (i.e., 32 images). This constraint matches the observing capabilities of the SDO and STEREO satellites, which are solely located on the ecliptic plane. Our data set corresponds to a single snapshot; therefore, we consider a static atmosphere where we set the temporal coordinate to 0. For our evaluation, we focus on the comparison of ground-truth and SuNeRF reconstructed images. With this, we assess the global plasma distribution and can estimate the model performance with standard image metrics (i.e., peak S/N, SSIM).

In Table 1, we summarize the quantitative evaluation of a baseline spherical reprojection and SuNeRF reconstructions. We compute the peak S/N, SSIM, relative mean absolute error (MAE), and relative mean error (ME) based on images from nonecliptic viewpoints that are not seen by the model during training (i.e., a test set of 224 images) and the synthesized SuNeRF images. As a baseline, we use a standard reprojection of synchronic maps, where each point is mapped to a spherical surface (Section 2.4). We note that this method cannot account for the off-limb emission of the Sun, while our SuNeRF method can take the extent of the 3D corona into account. The evaluation shows that our method provides a clear improvement over the baseline approach. Moreover, the metrics averaged over all considered wavelength channels indicate a strong similarity to the ground truth, with an MAE of 0.5%, an SSIM of 0.97, and a peak S/N of 40.4 dB, which throughout outperforms the baseline with an MAE of 3.2%, an SSIM of 0.73, and a peak S/N of 22.6 dB. The ME indicates that there is no systematic over- or underestimation (maximum ME: −0.3%).

Table 1. Summary of the Quantitative Evaluation of SuNeRFs and Baseline (i.e., Reprojected Synchronic Maps) Methods

MethodPeak S/N (dB)SSIMMAEME
Baseline—171 Å20.30.665.2%−1.2%
Baseline—193 Å23.60.683.4%−1.1%
Baseline—211 Å33.00.840.9%−0.3%
SuNeRF—171 Å34.70.950.9%−0.3%
SuNeRF—193 Å43.10.980.3%0.05%
SuNeRF—211 Å43.30.980.3%0.02%

Note. We evaluate both methods on the test data set (224 nonecliptic viewpoints), where we compare the peak S/N, SSIM, relative MAE, and relative ME.

Download table as:  ASCIITypeset image

In Figure 2, we compare ground-truth and reconstructed EUV images at different latitudes. Our method results in almost identical images to the ground truth. The difference maps indicate that errors primarily originate from off-limb features and pixelwise shifts. In addition, there is only a small quality decrease with increasing latitudes (e.g., peak S/N from 46 to 41 dB for 193 Å), which suggests a valid approximation of the 3D geometry of the solar corona. This is also visible from the pixelwise error maps. Our SuNeRF model achieves the highest similarity in quiet Sun regions, as well as the interior of coronal holes. On the other hand, features such as active regions are more complex, in addition to spanning a larger range of emission values. This results in larger differences when compared to the ground truth. Larger errors are expected close to the limb as the model integrates over a larger range of uncertainties. Therefore, independent of the viewing angle, inferences on the limb will have an increased uncertainty.

Figure 2.

Figure 2. Qualitative evaluation of the SuNeRF reconstruction of simulation data. We compare the ground-truth image (first row) with our SuNeRF reconstructions (second row) at different latitudes. The pixelwise difference shows where our reconstruction deviates from the ground-truth reference (third row). The uncertainty maps are computed from an ensemble of five different runs (fourth row). The SuNeRF reconstructions appear indistinguishable at larger scales. The primary errors occur close to off-limb and due to pixelwise shifts, while large-scale structures are consistent. The uncertainty maps align with regions of increased error, which enables a performance estimate in the absence of a reference image.

Standard image High-resolution image

To further estimate model errors, we render images from an ensemble (Lakshminarayanan et al. 2016) of five SuNeRFs with different random weight initializations and compute the standard deviation per pixel (Figure 2). The resulting uncertainty maps show a good spatial overlap with the pixelwise difference maps. We perform a pixelwise comparison of the uncertainty estimates and the MAE over the full test set (see also Appendix A for pixelwise scatter plots), where we find a Pearson correlation of 0.63 and a Spearman correlation of 0.69. This indicates that the ensemble approach reflects the model errors and can serve as an error estimate in the absence of a ground truth.

3.2. Complete Observation of the Solar Poles in EUV

We evaluate our method using multiviewpoint EUV observations from two different space missions, SDO and STEREO, with the latter mission consisting of twin satellites, STEREO-A and STEREO-B. All three spacecraft orbit in the ecliptic plane, with SDO in geosynchronous orbit around Earth and the STEREO spacecraft leading (STEREO-A) and trailing (STEREO-B) Earth. Even for constellations where we have full 360° coverage of the Sun, we find that three simultaneous images are insufficient for our model training. Therefore, we make use of the solar rotation that allows us to have a full scan of the Sun in ∼27 days, even for a single instrument. For our model training, we consider a sequence of 14 days at 1 hr cadence (from 2012 August 24 00:00 to 2012 September 7 00:00 UT; a total of 681 observations). We explicitly encode time as a coordinate for the sampling along light rays (Figure 1), which allows SuNeRFs to model temporal changes. We randomly sample 32,768 rays within the volume that we then use in our training set. Our model is trained for ∼250,000 iterations (three epochs) until the model converges (about 2 days on 8 × A100 GPUs). For our analysis, we primarily focus on the 193 Å channel, but we also provide reconstructions of the 171, 211, and 304 Å channels. Before training, we adjust instrumental differences, including differences in the wavelength bands, using ITI (Jarolim et al. 2024), allowing us to generate homogeneous observations of SDO and STEREO.

In Figure 3, we show snapshots from a global reconstruction of the solar atmosphere. The arrows in Figure 3(a) indicate the viewpoints used for the reconstruction, with latitudes ranging from −4° to 8°. The central image shows the reconstruction of the solar south pole at 2012 August 30 00:00:00 (UT). This directly enables us to further investigate the polar coronal hole, which we segment using a deep-learning tool for automatic coronal hole detection (CHRONNOS; Jarolim et al. 2021). We then compare two approaches in identifying the coronal hole boundaries. (1) We use the individual satellite observations (i.e., full-disk images) to identify coronal holes and then reproject the combined binary maps to the polar viewpoint (blue contours). (2) We use SuNeRF to reconstruct the polar observation and then apply the detection method (red contours). The comparison shows that the strong projection effects only allow for a partial identification of the coronal hole boundary from the ecliptic plane. Coronal holes are also frequently obscured by loop systems in adjacent active regions, which makes detection and consequent space-weather predictions challenging. Our method can account for these projection effects by mapping the 3D geometry (see the online animation of Figure 3). From our comparison to simulated coronal holes, the SuNeRF reconstruction provides a more consistent reference to estimate the coronal hole boundary, specifically in the case of strong reprojection effects, where errors are expected to be in the range of the uncertainty maps (Figure 3(b)).

Figure 3. Global reconstruction of the solar EUV corona on 2012 August 30 00:00:00 (UT). (a) Complete image of the solar south pole as reconstructed by our method from STEREO/EUVI and SDO/AIA observations (left). Reconstructions in 171, 211, and 304 Å are shown on the right. The image at the center shows the reconstruction in 193 Å, where the polar coronal hole is best visible. Coronal hole boundaries detected using CHRONNOS are indicated as contour lines. Blue lines refer to spherical reprojections from the ecliptic perspective, and red lines are obtained from the SuNeRF reconstruction. The arrows indicate the positions of individual observations used for training, and the color coding refers to the temporal distribution of each instrument (green: STEREO-A; red: STEREO-B; blue: SDO). (b) Observations rendered for different latitudes and a fixed longitude (white arrow in (a)). The baseline spherical reprojection (observations stitched into a synchronic map; first row) shows artifacts, provides poor performance for overlapping observations, and is unable to resolve the polar coronal hole. The SuNeRF method (second row) provides consistent observations at all latitudes. The uncertainty maps (third row) indicate regions where higher model errors are to be expected. The EUV color bars are given in units of DN s−1, and the 193 Å images use the same data scaling throughout. An animation of the SuNeRF reconstruction of the global EUV corona in the 193 Å channel on 2012 August 29 (UT) is available in the online journal. The animation first shows the Sun from ecliptic viewpoints, followed by the polar viewpoints. The final part highlights how our model can capture the Sun's temporal evolution, including the reconstruction of a solar eruption.

(An animation of this figure is available.)

Video Standard image High-resolution image

In panel (b), we compare viewpoints from different latitudes against the corresponding baseline reprojections. SuNeRF clearly recovers the polar coronal hole by combining the multi-instrument data. Artifacts occur only for the off-limb region that is not sampled by the ray tracing (i.e., bottom left at −90°).

In the bottom row of Figure 3(b), we provide uncertainty estimates derived via an ensemble of five individual runs performed with random initializations. The uncertainty maps serve as a proxy for the model error in the absence of a ground truth (Lakshminarayanan et al. 2016). The uncertainty is generally higher for observations than for the model data in Section 3.1. This uncertainty likely arises from the increased complexity of the observations (i.e., the high-frequency features are more difficult to match) compared to the smoother atmosphere in the simulations. Note the higher number of small regions (e.g., bright points) with errors >10%. These could be associated with a misalignment in temporal evolution, where the 1 hr temporal resolution may be too coarse to capture small brightness variations. Overall, large-scale features (i.e., active regions, coronal holes, filaments) are consistently reconstructed by our method, as can be gauged by the uncertainty maps, and the main errors are primarily related to pixelwise shifts (e.g., the coronal hole boundary).

3.3. Height Profiles

Using SuNeRFs, we can now determine the height of the solar features from the 3D reconstructions. In Figure 4(a), we extract a meridional slice to analyze the emission profile from 1.0 to 1.3 R. From north to south, we observe the emission profile of an active region, a coronal hole (blue), a filament channel (orange), and the polar coronal hole. The 2D coronal slice in Figure 4(b) shows regions of lower emission close to the surface. A possible explanation for the modeled emission profile is that our model accounts for the solar chromosphere and transition region, which are not visible in EUV due to temperatures being below the million K range, to generate EUV emission. In contrast, bright active region loops are rooted in the solar photosphere. The coronal hole can be clearly identified from the decreased emission of the entire column. The solar filament channel can be identified from the dark cavity that appears as a distinct and sharp drop in intensity. Lower layers show a similar brightness as the quiet Sun corona, where the reduced emission primarily originates from the dark filament channel. The most distinct emission decrease occurs at the solar south pole, where we observe the polar coronal hole (Section 3.2). From the emission, we can estimate the relative plasma density by assuming that the emission is proportional to the squared density. This provides a direct estimate of the coronal hole boundaries, which appear as distinct jumps in the integrated emission profiles for the polar and low-latitude coronal holes (Figure 4(c)).

Figure 4.

Figure 4. Emission profile of the solar corona as estimated by SuNeRFs. (a) Intercalibrated STEREO-A observation, where the red line indicates the analyzed slice through the solar atmosphere (fixed longitude). The blue and orange segments indicate the coronal hole and the filament channel, respectively. (b) SuNeRF-rendered view, rotated such that the atmospheric slice in (a) coincides with the limb. In the solar atmosphere above said slice, we compute the emission at each height. The 2D atmospheric profile clearly shows the bright active region close to the equator. The coronal holes (blue line and south pole) show a strong decrease in intensity. The black dashed line indicates the radial average height of channel-specific emission, which varies between 15 and 50 Mm (0.02–0.07 R). (c) Square root of the normalized integrated channel-specific emission in the analyzed slice. The emission refers to the radially outgoing total emission, which we use as a proxy for the plasma density distribution. The blue segment shows the exact coronal hole profile. The orange segment is associated with the solar filament and appears as a sharp drop in the profile (coronal cavity).

Standard image High-resolution image

3.4. Reconstruction of Solar Eruptive Events

To investigate solar transient events, we increase the temporal cadence of observations in our training set and focus on a subframe instead of the full field of view. Here, we apply our method to the solar eruption that started at 2012 August 31 19:30 (UT), and we use observations from STEREO-B and SDO in the 304 Å channel. We use the parameters learned from the 304 Å model in Section 3.2 as a starting point, which already provides the general geometry and structure of the considered time period. We focus the training on 1024 × 1024 pixel subframes centered at the erupting solar filament (Carrington longitude of 90° and latitude −20°) at a 1 minute cadence and train our model for an additional 100,000 iterations (∼four epochs).

In Figure 5(a), we give an overview of the filament eruption event and show SuNeRF-rendered images during and after the eruption (top and bottom row, respectively). We provide height maps that give the distance from the solar center for the observed emission (Section 3.3). The reconstruction shows that the prominence rises slowly (yellow arrow), followed by a rapid ejection of the filament plasma into interplanetary space (coronal mass ejection). Afterward, the hot flare ribbons, which are clearly mapped close to the solar surface, are formed, in agreement with our physical understanding of eruptions (blue arrow). In addition, we show the total absorption along the line of sight, where the filament structure can be clearly identified. The ejection can be related to the depletion of absorbing plasma.

Figure 5. Solar eruption observed on 2012 August 31 in the 304 Å channel. (a) Full-disk SDO/AIA image before the onset of the eruption and subframes of our SuNeRF reconstruction. SuNeRF images (left column) are rendered during (top) and after (bottom) the event using the SDO viewpoint. The middle column shows heliocentric height estimates. The right column shows the integrated absorption along the line of sight, indicating dense plasma. (b) Temporal evolution of the event as observed for a slice through the center of the solar filament (blue line in panel (a)). We plot the spatial distribution of emission (top) and absorption (bottom) in the atmosphere above the slice. The reconstruction shows a slow rise of the filament, as indicated by regions of strong absorption, and rapid ejection of coronal mass, both visible in emission and absorption. After the ejection, two flare ribbons can be clearly mapped close to the surface (height maps), and postflare loops become visible in the emission slices. An animation of the solar filament eruption from 2012 August 31 is available in the online journal. Left: reconstructed EUV observations from the SDO perspective using the 304 Å channel. Center: height above the surface as estimated by our model. Right: total absorption along the line of sight as estimated by our model. The reconstructions clearly show the ejection of dense plasma into interplanetary space.

(An animation of this figure is available.)

Video Standard image High-resolution image

Triangulation is a useful approach for height estimations of dense plasma but can lead to ambiguous results for an optically thin medium (Aschwanden 2005, 2011). With the full mapping of the solar atmosphere, we can better analyze the underlying 3D structure and study emission and absorption profiles along their radial extent. In Figure 5(b), we show the time evolution of an extracted slice of the modeled solar atmosphere (blue line in Figure 5(a)). The top row shows the emission profile. After the eruption (∼19:40), postflare loops and their related flare ribbon footpoints are formed (blue arrow). The loop height can be directly estimated from the profiles (∼0.7 R or 500 Mm above the solar surface). The absorption profiles show the evolution of the filament plasma and allow one to determine the direction and velocity of the erupting structure.

4. Discussion

The analysis in Section 3 demonstrates that SuNeRF can use multi-instrument/multiviewpoint data to create virtual instruments and provide a realistic 3D representation of the Sun. The method is able to reconstruct regions that are only sparsely observed by the individual instruments, like the solar poles. Comparisons to a baseline reprojection demonstrate that our method provides a significant advancement for the synthesis of novel viewpoints. The quantitative evaluation with simulation data shows that our method provides almost identical reconstructions, where differences mostly occur close to the solar limb and in active regions (Section 3.1). Extended regions with less spatial variability (e.g., coronal holes, quiet Sun) are less viewpoint-dependent and thus result in smaller reconstruction errors. Uncertainty maps allows us to determine model errors in the absence of a ground-truth reference. For the application to SDO/AIA and STEREO/EUVI data, uncertainties are higher than for simulation data, which we associate with the increased complexity of the corona (e.g., active regions and the solar limb are structurally more complex and detailed than our validation data) and the additional modeled temporal evolution (e.g., the model must now account for the different lifetimes of dynamically evolving solar features).

The SuNeRF reconstructed polar view allows for a more consistent detection of the polar coronal hole boundary than the combined detections from the ecliptic plane (Section 3.2), which demonstrates that our method can largely reduce projection effects. With the use of 2D slices through the solar atmosphere, we demonstrate that our method can provide additional insights into solar eruptive events (Section 3.4) and provides a direct estimation of (reprojected) coronal hole boundaries and profiles of solar filaments (Section 3.3). This provides the basis to directly detect coronal holes from the reconstructed emission profile.

Our method uses video sequences of EUV data for detailed reconstructions with only three simultaneous observing instruments. This is explicitly important for regions that are not covered at a given point in time but can be approximated from previous and preceding frames. In Section 3.4, we analyzed a solar eruption that is constrained by two viewpoints and reduced the field of view while increasing the temporal resolution of the training data. The reconstruction allows one to extract additional height information of the event, which can be useful for the estimation of space-weather impacts (Thernisien et al. 2009). From our model, we can directly map flare-related emission (e.g., flare ribbons, loops) and the ejection of dense plasma (i.e., filament eruption). A possible extension that would improve the tracking of coronal mass ejections is the application to coronagraph data that can monitor the propagation through the corona and interplanetary space.

A limiting factor of our method is the long training time and large computational demand. Training from scratch requires days to reconstruct a single sequence, even in a high-performance computing environment. For our training, the global structures are learned within the first few thousand iterations, but reconstructing details, such as thin coronal loops, requires the full training time. Our method reconstructs sequences of 15 days, which already provides 3D reconstructions in real time, but individual analyses would require a full simulation run. Extending the temporal range results in a decrease in spatial and temporal resolution due to the limited capacity of the model and increased training demand. A suitable strategy to reconstruct large data sets would be a continuous simulation, where new frames are added to the end of the sequence while frames at the beginning are truncated and model weights are sequentially extracted. Individual events could then be analyzed by branching and fine-tuning the model with high-cadence data of the region of interest. This could provide a higher-level data archive that exploits additional information from the available observations. Note that a larger number of parameters could potentially capture more details of the 3D volume but requires more computing power and training time, which is typically the main limitation of our approach.

Using the multichannel observations from STEREO/EUVI and SDO/AIA and methods of differential emission measure (DEM), SuNeRF can be extended to give 3D reconstructions of the densities and temperatures in the solar corona. This approach is similar to DEM tomography (Frazin et al. 2005, 2009; Vásquez et al. 2009; Vásquez 2016). Here, the SuNeRF approach has the potential to improve the spatial and temporal resolution of the reconstructions, which can be particularly important for the study of faint structures (e.g., coronal prominence cavities; Vásquez et al. 2009) and transient events (e.g., EUV waves).

The analysis presented here takes advantage of multiviewpoint observations captured with an optimal separation angle between individual spacecraft for full coverage of the entire Sun. The application to a reduced number of viewpoints and the variation in reconstruction quality could inform future constellation missions (e.g., required coverage). The recent Solar Orbiter mission (SolO; Müller et al. 2020) provides additional EUV observing capabilities (Extreme-Ultraviolet Imager, EUI; Rochus et al. 2020) at nonecliptic angles that can be valuable in further verifying the SuNeRF reconstructions (e.g., by inferring the SolO viewpoint and comparing it to actual SolO observations) or to extend the training data (e.g., including an additional viewpoint). We note that this requires an additional preprocessing to intercalibrate the SolO/EUI and SDO/AIA instruments.

Our method achieves a 3D representation of the atmosphere of our closest star by adapting NeRFs. This successful application has implications for other domains in astrophysics and geophysics. NeRF-based algorithms could provide improved 3D reconstructions from planetary, lunar, and asteroid surveys, even for complex atmospheric environments (e.g., clouds, dust, and gas), all of which are domains likely to soon need such methods as multipoint observations become more commonplace (e.g., cubesats). A future extension of our method is the use of more physical constraints, with emission and absorption deduced from density and temperature profiles. This provides an additional opportunity to include physics-informed losses (Raissi et al. 2019) and to connect plasma density with magnetic fields through magnetohydrodynamic equations (Jarolim et al. 2023).

Acknowledgments

This work has been enabled by the Frontier Development Lab (FDL 11 ). FDL is a collaboration between the SETI Institute and Trillium Technologies Inc., in partnership with NASA, Google Cloud, Intel, NVIDIA, and many other public and private partners. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Aeronautics and Space Administration. The authors would like to thank the FDL organizers, the SETI institute, Trillium, the FDL partners and sponsors, and the reviewers for their constructive comments during the research sprint. The authors also want to thank Yarin Gal and Chedy Raissi for valuable discussions. Finally, the authors also thank Google and NVIDIA for providing access to computational resources without which this project would not have been possible. The authors acknowledge the financial support by the University of Graz. A.V. was also supported by NASA grant 80NSSC19K1261 and the STEREO program support to FDL. B.T was also supported by NASA grants 80NSSC21K1796 and NNG04EA00C.

Facilities: SDO (AIA) - , STEREO (EUVI) - .

Software: Astropy (Astropy Collaboration et al. 2013, 2018), Sunpy (Barnes et al. 2020; Mumford et al. 2020), PyTorch (Paszke et al. 2019).

Appendix A: Extended Evaluation of Simulation Data

In Figure 6, we show a comparison between the test set of the simulated data and our SuNeRF reconstructions. We perform a pixelwise comparison that shows a clear correlation between the SuNeRF values and the ground-truth reference. The uncertainty estimate and MAE have a different scaling but clearly show a linear correlation, in agreement with the correlation coefficients in Section 3.1.

Figure 6.

Figure 6. Correlation between SuNeRF reconstruction and ground-truth reference. Left: 2D histogram of simulation data (ground truth) and our SuNeRF reconstructions. Right: 2D histogram of MAE and model uncertainty. The color bars give the total number of pixels per bin.

Standard image High-resolution image

Appendix B: Synchronic Maps

Using full 3D reconstructions of the Sun, we can render true synchronic maps, where each pixel corresponds to an observation normal to the solar surface. In other words, we can obtain ideal reprojected synchronic maps. For this, we sample 512 points from 1.0 to 1.3 R and compute the observed pixel intensity analogous to the ray tracing in Section 2. We sample a full Carrington map (360° × 180°), where we use a pixel size of 0fdg025 × 0fdg025 pixel–1. In Figure 7, we compare a conventional synchronic map and an adjusted synchronic map (Caplan et al. 2016) with our SuNeRF map. The conventional map shows projection effects toward the poles, the combination of overlapping regions leads to blurred results, and unobserved regions are interpolated with the mean value (e.g., north pole). The adjusted map improves overlapping regions but leads to artifacts where features do not spatially align (blue circle). From our SuNeRF map, we obtain consistent reconstructions, where solar features at the disk center are similar to the original observations.

Figure 7.

Figure 7. Comparison of synchronic maps on 2012 August 30 00:00 (UT). (a) Conventional synchronic map combining three simultaneous observations from different viewpoints using the reproject function by SunPy. (b) Adjusted synchronic map optimized for coronal hole detections (see Caplan et al. 2016). (c) SuNeRF synchronic map. Reprojected maps can result in poor alignment of features, while our SuNeRF approach consistently combines the three observations (blue circle). Reprojection effects at higher latitudes are corrected by our approach, leading to a clearer identification of the polar coronal hole boundary (green circle). Note that the resolution of our SuNeRF reconstruction is reduced as compared to the synchronic maps, which is related to the limited capacity of the neural representation. The values are given in units of DN s−1.

Standard image High-resolution image

Footnotes

Please wait… references are loading.
10.3847/2041-8213/ad12d2