This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy. Close this notification
Brought to you by:

High-cadence Imaging and Imaging Spectroscopy at the GREGOR Solar Telescope—A Collaborative Research Environment for High-resolution Solar Physics

, , , , , , , , , and

Published 2018 May 11 © 2018. The American Astronomical Society. All rights reserved.
, , Citation Carsten Denker et al 2018 ApJS 236 5 DOI 10.3847/1538-4365/aab773

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

0067-0049/236/1/5

Abstract

In high-resolution solar physics, the volume and complexity of photometric, spectroscopic, and polarimetric ground-based data significantly increased in the last decade, reaching data acquisition rates of terabytes per hour. This is driven by the desire to capture fast processes on the Sun and the necessity for short exposure times "freezing" the atmospheric seeing, thus enabling ex post facto image restoration. Consequently, large-format and high-cadence detectors are nowadays used in solar observations to facilitate image restoration. Based on our experience during the "early science" phase with the 1.5 m GREGOR solar telescope (2014–2015) and the subsequent transition to routine observations in 2016, we describe data collection and data management tailored toward image restoration and imaging spectroscopy. We outline our approaches regarding data processing, analysis, and archiving for two of GREGOR's post-focus instruments (see http://gregor.aip.de), i.e., the GREGOR Fabry–Pérot Interferometer (GFPI) and the newly installed High-Resolution Fast Imager (HiFI). The heterogeneous and complex nature of multidimensional data arising from high-resolution solar observations provides an intriguing but also a challenging example for "big data" in astronomy. The big data challenge has two aspects: (1) establishing a workflow for publishing the data for the whole community and beyond and (2) creating a collaborative research environment (CRE), where computationally intense data and postprocessing tools are colocated and collaborative work is enabled for scientists of multiple institutes. This requires either collaboration with a data center or frameworks and databases capable of dealing with huge data sets based on virtual observatory (VO) and other community standards and procedures.

Export citation and abstract BibTeX RIS

1. Introduction

Challenges posed by "big data" certainly became a topic in solar physics with the launch of space missions such as the Solar and Heliospheric Observatory (SOHO; Domingo et al. 1995) and the Solar Dynamics Observatory (SDO; Pesnell et al. 2012). Synoptic full-disk images (visible, UV, and EUV), magnetograms, and Doppler maps are the core data products of both missions. However, the SDO pushed the limits of spatial resolution to 1'' and a cadence of 12 s. The results are about 300 million images and a total data volume of more than 3.5 PB during the 7 yr mission time so far. Data flow and processing for the SDO are described in Martens et al. (2012) concerning both hardware and software aspects, particularly the data conditioning for helioseismology and near real-time data products needed in space weather prediction and forecast. Many of these data products are available from the Joint Science Operations Center5 (JSOC) at Stanford University. Considering the volume of data, various institutes around the world hold partial or full copies of SOHO and SDO for advanced in-house processing.

The data volume of ground-based synoptic full-disk observations also significantly increased over the years, even though it did not reach the magnitude of space data. Both helioseismology and space weather applications require continuous observations; thus, telescope networks such as the Global Oscillation Network Group (GONG; Leibacher 1999) and the Global Hα Network (Denker et al. 1999; Steinegger et al. 2000) are a natural choice to overcome the day-night cycle. Other important synoptic data sets are hosted at the Digital Library6 of the U.S. National Solar Observatory (NSO), e.g., the photospheric and chromospheric vector magnetograms of the Synoptic Optical Long-term Investigations of the Sun (SOLIS; Keller et al. 2003; Henney et al. 2009) program, or at institutional data repositories, e.g., full-disk images of the Chromosheric Telescope7 (ChroTel; Kentischer et al. 2008), operated by the Kiepenheuer Institute for Solar Physics in Freiburg, Germany, and coronal images8 of the Coronal Multichannel Polarimeter (CoMP; Tomczyk et al. 2008) and the COSMO K-coronagraph (Tomczyk et al. 2016) from the Mauna Loa Solar Observatory. These data are typically accessible via FTP archives or can be requested via web-based query forms. High spectral resolution spectroscopy and spectropolarimetry were historically the domain of ground-based telescopes. The Japanese Hinode (Kosugi et al. 2007) space mission with its 50 cm Solar Optical Telescope (SOT; Tsuneta et al. 2008) changed this field by providing high spectral and spatial resolution full Stokes polarimetry with moderate temporal resolution but high sensitivity (Ichimoto et al. 2008). Hinode data are publicly available in the Data ARchive and Transmission System (DARTS; Miura et al. 2000) at the Institute of Space and Astronautical Science (ISAS) in Japan and are mirrored to data centers around the world (Matsuzaki et al. 2007).

Access to high-resolution ground-based data is often difficult because they were obtained in campaigns led by principal investigators (PIs) and their teams. In addition, poor documentation and offline storage of high-resolution data (tape drives and local hard-disk drives) have hampered efforts to make them openly accessible. Fortunately, the situation is improving, as more and more data holdings adopt the open-access paradigm. For example, high-resolution data from the 1.6 m Goode Solar Telescope (Cao et al. 2010) at Big Bear Solar Observatory are listed on the observatory website and can be requested via web interface. In preparation for the next generation of large-aperture solar telescopes, the NSO tested a variety of observing modes, including proposal-based queue observations executed by experienced scientists and observatory staff. These service-mode data9 also became publicly available and are accessible via the NSO Digital Library infrastructure.

At the moment, the Daniel K. Inouye Solar Telescope (DKIST; McMullin et al. 2014; Tritschler et al. 2016) is under construction, with first light anticipated in 2020, and the European Solar Telescope (EST; Collados et al. 2010a, 2010b) completed its preliminary design phase and was included in the 2016 road map of the European Strategy Forum for Research Infrastructures (ESFRI). Concepts for the exploitation of DKIST, with special emphasis on the "big data" challenge, were presented in Berukoff et al. (2016). The current generation of 1 m class solar telescopes like GREGOR (Schmidt et al. 2012) and the Goode Solar Telescope can be considered stepping stones toward unveiling the fundamental spatial scales of the solar atmosphere, i.e., the pressure scale height, the photon mean free path, and the elementary magnetic structure size. All scales are of the order of 100 km or even smaller. This provides the impetus for large-aperture solar telescopes and high-resolution solar physics, i.e., to approach temporal and spatial scales that are otherwise only accessible with numerical radiative MHD simulations (e.g., Vögler et al. 2005; Rempel & Cheung 2014; Beeck et al. 2015).

In this article, we introduce high-resolution ground-based imaging (spectroscopic) data obtained with the 1.5 m GREGOR solar telescope and describe our approaches to data processing, analysis, management, and archiving. We show by example how the dichotomy of ground-based versus space-mission data and synoptic versus high-resolution data affects these approaches. In addition, we present the collaborative research environment (CRE) for GREGOR data as a concept tailored toward the needs of the high-resolution solar physics community. This is complementary to research infrastructure such as the Virtual Solar Observatory10 (VSO; Hill et al. 2004), which was established to allow easy access to solar data from various space missions, as well as from ground-based observatories. The VSO reduces the effort of locating and downloading different data in different archives by providing a common web-based interface. The VSO design minimizes the hardware and software resources of federated archives and often simplifies integration of new data. The VSO does not store any data but integrates federated archives by offering a central registry and interfaces for distributed queries to multiple independent data repositories. Alternative access to solar data is provided by the SolarSoft (Bentley & Freeland 1998) library, which is written mainly in the Interactive Data Language11 (IDL). Virtual observatory implementations in solar physics also exist on the European level with the European Grid of Solar Observations (Bentley 2002) and, more recently, the SOLARNET Virtual Observatory12 (SVO). Currently, only an SVO prototype is available. However, data can be searched based on events, data set specific parameters, and cotemporal observations.

In the following, we provide a comprehensive overview of GREGOR high-resolution data—from the photons arriving at the detector to the final data products. In Section 2, we describe the telescope, its post-focus instruments, and the particulars of high-cadence imaging. The data-processing pipeline, its design, and its relation to other software libraries is introduced in Section 3. Data management and access to the GREGOR Fabry–Pérot Interferometer (GFPI) and High-resolution Fast Imager (HiFI) archive13 (Section 4) comprise the domain-specific answers to the challenges provided by "big data" in solar and stellar astronomy. Finally, the conclusions in Section 5 develop a perspective for CREs in high-resolution solar physics and explore future extensions allowing database research.

2. GREGOR Solar Telescope and Instrumentation

2.1. GREGOR Solar Telescope

The 1.5 m GREGOR solar telescope is the largest telescope in Europe for high-resolution solar observations (see Soltau et al. 2012 for the origin of the telescope's name and why it is formatted in capital letters). Located at Observatorio del Teide, Izaña, Tenerife, Spain, the telescope exploits the excellent and stable seeing conditions of a mountain-island observatory site. The concept for the telescope's mechanical structure (Volkmer et al. 2012) and the open design employing a foldable-tent dome (Hammerschlag et al. 2012) allow for wind flushing of the telescope platform, which minimizes dome and telescope seeing. The GREGOR telescope uses a double Gregory configuration (Soltau et al. 2012) to limit the field of view (FOV) to a diameter of 150'' to facilitate on-axis polarimetric calibrations in the symmetric light path and provide a suitable f-ratio to the GREGOR Adaptive Optics System (GAOS; Berkefeld et al. 2012) and the four post-focus instruments: the Broad-Band Imager (BBI; von der Lühe et al. 2012), GREGOR Infrared Spectrograph (GRIS; Collados et al. 2012), GFPI (Denker et al. 2010; Puschmann et al. 2012 and references therein), and HiFI (C. Denker et al. 2018, in preparation), which replaced the blue imaging channel (BIC) that used much slower and outdated facility cameras. In the present study, we discuss data processing, analysis, management, and archiving for the last two instruments.

2.2. GFPI

The GFPI is a tunable, dual-etalon imaging spectrometer, where the etalons are placed near a conjugated pupil plane in the collimated beam. The etalons were manufactured by IC Optical Systems (ICOS) in Beckenham, UK. They have a 70 mm diameter free aperture and possess both high finesse (${{ \mathcal F }}_{\mathrm{eff}}=45\mbox{--}50$) and high reflectivity (R ≈ 95%). Their coatings are optimized for the spectral range 5300–8600 Å, where the instrument achieves a spectral resolution of ${ \mathcal R }\approx 250,000$. Spectral scans are recorded with two precisely synchronized Imager QE CCD cameras from LaVision in Göttingen, Germany. One camera acquires narrowband filtergrams, whereas the simultaneous broadband images enable image restoration of the full spectral scan using various deconvolution techniques.

The 12 bit images with 1376 × 1040 pixels are captured at a rate of up to 10 Hz for full frames, depending on exposure time. The image scale of about 0farcs04 pixel−1 yields an FOV of 55'' × 42''. The image acquisition rate can be doubled with 2 × 2 pixel binning. The relatively small full-well capacity of 18,000 e and low readout noise of the detectors are well adapted to the instrument design, considering the small full width at half maximum (FWHM) of the double-etalon spectrometer of just 25–40 mÅ, the maximum quantum efficiency of 60% at 5500 Å, and the short exposure times (texp = 10–30 ms) needed to "freeze" the seeing-induced wavefront distortions. The typical cadence of Δt = 20–60 s for a spectral scan provides very good temporal resolution so that dynamic processes in the solar photosphere and chromosphere can be resolved.

However, both exposure time and cadence are already compromises because the high spectral resolution leads to a low number of incident photons at the detector, and small-scale features potentially move or evolve significantly in the given time interval. Faster detectors combining low readout noise, comparatively small full-well capacity, good quantum efficiency, and a high duty cycle with respect to the exposure time mitigate these limitations and are considered for future upgrades. In principle, the GFPI can be operated in a polarimetric mode, which produces spectral scans of the four Stokes parameters so that the magnetic field vector can be inferred for each pixel. However, validation of the polarimetric mode is still in progress. Consequently, imaging polarimetry is not covered in this study.

Sample data products obtained from a scan of the strong chromospheric absorption line Hα 6562.8 Å are compiled in Figure 1 to illustrate the GFPI's science capabilities. Ellerman bombs (see Rutten et al. 2013 for a review), as opposed to microflares, are typically observed as enhanced line-wing emission in Hα, Hβ, Hγ, etc. and have typical lifetimes of 1.5–7 minutes with a maximum of around 30 minutes (Pariat et al. 2007). To resolve their evolution requires fast spectral scans with imaging spectroscopy and ex post facto image restoration to uncover small-scale dynamics within the photosphere and chromosphere. On 2014 August 14, active region NOAA 12139 was observed with the GFPI focusing on a complex group of sunspots and pores (see broadband image in Figure 1). Image restoration using Multi-Object Multi-Frame Blind Deconvolution (MOMFBD; Löfdahl 2002; van Noort et al. 2005) was applied to the spectral data to enhance solar fine structure. The blue and red line-wing filtergrams reveal two features with different spectral characteristics: an Ellerman bomb (Ellerman 1917) as localized, small-scale brightenings (area "a") and a system of dark fibrils with strong downflows associated with newly emerging flux (area "b"), respectively.

Figure 1.

Figure 1. Imaging spectroscopy of active region NOAA 12139 observed on 2014 August 14 with the GFPI in the strong chromospheric absorption line Hα 6562.8 Å: broadband image (top left) near the Hα spectral region, line-core intensity image (bottom left), blue line-wing image at Δλ = −0.70 Å (top middle), red line-wing image at Δλ = +0.60 Å (bottom middle), and chromospheric Doppler velocity derived with the Fourier phase method (top right), where blue and red represent upflows and downflows, respectively. Samples of spectral profiles (bottom right) are shown for areas "a" (black), "b" (red), quiet Sun ("QS"; blue), and Fourier transform spectrometer (FTS; Wallace et al. 1998) spectral atlas (green). The bullets on the quiet-Sun profile denote the positions of the three displayed filtergrams.

Standard image High-resolution image

High spatial resolution imaging and imaging spectroscopy belong to the standard observational techniques of large-aperture solar telescopes. Various studies based on GFPI data illustrate the potential of the instrument. Recently, Kuckein et al. (2017b) used GFPI Ca ii 8542.1 Å filtergrams to study sudden chromospheric small-scale brightenings. The combination of ground-based, high-resolution imaging spectroscopy and synoptic EUV full-disk images from space reveals that the brightenings belong to the footpoints of a microflare. To further investigate the bright kernels and the central absorption part (below the flaring arches), spectral inversions of the near-infrared Ca ii line are performed with the NICOLE code (Socas-Navarro et al. 2015). The retrieved average temperatures reveal rapid heating at the brightenings (footpoints) of the microflare of about 600 K. The inferred line-of-sight (LOS) velocities at the central absorption area show upflows of about −2 km s−1. In contrast, downflows dominate at the other footpoints.

In another study, Verma et al. (2018) used high-resolution imaging and spectroscopic GFPI data in the photospheric Fe i 6173.3 Å spectral line to infer the three-dimensional velocity field associated with a decaying sunspot penumbra (Figure 2). The velocities in the decaying penumbral region deviate from the usual penumbral flow pattern because of flux emergence in the vicinity of the sunspot. The detailed analysis is based not only on GFPI data but also HiFI and GRIS observations, which provide further photospheric and chromospheric diagnostics. In a study with a similar setup for GREGOR multiwavelength and multi-instrument observations, Felipe et al. (2017) investigated the impact of flare-ejected plasma on sunspot fine structure, i.e., a strong light bridge, which experienced localized heating and changes of its magnetic field structure.

Figure 2.

Figure 2. Three-dimensional flow field observed in active region NOAA 12597 on 2016 September 24. The horizontal flows were derived from a time series of restored GFPI broadband images at 6122.7 Å, whereas the Doppler velocities were determined from a restored narrowband scan of the photospheric Fe i 6173.3 Å line. Color-coded local correlation tracking (LCT; November & Simon 1988; Verma & Denker 2011) vectors are superposed onto the average LOS velocity map, which was scaled at ±0.7 km s−1. The black and white contours delineate the umbra–penumbra and penumbra–granulation boundaries, respectively.

Standard image High-resolution image

2.3. High-resolution Fast Imager

Detectors based on scientific complementary metal-oxide-semiconductor (sCMOS) technology have become an alternative to standard CCD devices in astronomical applications, in particular when high-cadence and large-format image sequences are needed. In early 2016, we installed HiFI at the GREGOR solar telescope, where it observes the blue part of the spectrum (3850–5300 Å) using a dichroic beam splitter in the GFPI's optical path. Two Imager sCMOS cameras from LaVision (LaVision 2015) are synchronized by a programmable timing unit and record time series suitable for image restoration either separately for each channel or by making use of both channels at the same time.

In typical observations, two of three spectral regions are selected, i.e., Ca ii H 3968.0 Å, Fraunhofer G-band 4307.0 Å, and blue continuum 4505.5 Å (see Figure 3). The width of the filters is around 10 Å, so typical exposure times reach from a fraction of a millisecond to a few milliseconds. Thus, observations are not "photon-starved," as often encountered in imaging spectropolarimetry, where the transmission profile of (multiple) Fabry–Pérot etalons can be as narrow as 25 mÅ. Different count rates in both channels are balanced by choosing suitable neutral density filters. Short exposure times are essential to freeze the wavefront aberrations in a single exposure. The 2560 × 2160 pixel images with an FOV of 64farcs× 54farcs6 are digitized as 16 bit integers and recorded with a data acquisition rate of almost 50 Hz. Thus, the image scale is about 0farcs025 pixel−1, or about 18 km on the solar surface at disk center. The diffraction-limited resolution of the GREGOR telescope with a diameter D = 1.5 m at a wavelength λ = 4000 Å is α = λ/ D = 0farcs055. According to the Nyquist sampling theorem, HiFI images are critically sampled at the shortest wavelength of the standard interference filters.

Figure 3.

Figure 3. Blue continuum image 4505.5 Å of active region NOAA 12529 obtained with HiFI at 08:37 UT on 2016 April 11. The image was restored from a time series of 100 short-exposure images with the speckle-masking method implemented in KISIP.

Standard image High-resolution image

In the standard HiFI observing mode, sets of 500 images are captured in each channel in 10 s and continuously written to a RAID-0 array of SSDs at a cadence below 20 s. The imaging system has achieved a sustained write speed summed over both channels of up to 660 MB s−1. To limit the final data storage requirements, only the best 2 × 100 images of a set are kept for image restoration and further data analysis. These settings are already a compromise, considering that solar features move with velocities of several kilometers per second in the photosphere and several tens of kilometers per second in the chromosphere, often exceeding the speed of sound. Eruptive phenomena in the chromosphere reach even higher velocities, in excess of 100 km s−1. Thus, in the standard observing mode, solar features moving at more than 2 km s−1 and traversing a pixel in less than about 10 s will be blurred, and their proper motions will not be properly resolved. However, considering that HiFI provides mainly context data and only some pixels are exposed to high velocities, this compromise is acceptable. If higher temporal resolution is required—for example, when tracking small-scale bright points or following the evolution of explosive events—changing the observing mode is an option, i.e., reading out smaller subfields to increase the data acquisition rate or dropping frame selection to keep all observed frames.

2.4. Implications of High-cadence Imaging

As demonstrated above, many instrument designs in solar physics rely on high-cadence imaging. The data challenge arises from the combination of several factors: the daytime correlation timescale of the seeing, the evolution timescale of the solar features, and large-format detectors. The latter are needed to either catch transient events or observe large-scale coherent features that provide structuring and connectivity in the solar atmosphere. To illustrate the implications, we carried out a small experiment with the HiFI cameras. Images were acquired at 135 Hz for only a small FOV of 640 × 640 pixels, i.e., 16farcs× 16farcs3 on the solar disk. Three sets of 50,000 images were written to disk with short interruptions at a sustained rate of 220 MB s−1. An extended study based on an even higher image acquisition rate is presented in Denker et al. (2018), where we evaluated image-quality metrics and the impact of frame selection for AO-corrected images on image restoration with the speckle-masking technique (Weigelt & Wirnitzer 1983; de Boer 1993; von der Lühe 1993).

The median filter gradient similarity (MFGS; Deng et al. 2015) is an image-quality metric recently introduced into solar physics. The results for G-band images of the full time series are depicted in the top panel of Figure 4, whereas the two lower panels show successively shorter time periods centered around the moment of best seeing conditions. The AO system locked on a decaying pore in active region NOAA 16643 that contained a light bridge, umbral dots, and indications of penumbra-like small-scale features, i.e., elongated, alternating bright and dark features at the umbra–granulation boundary, as well as chains of filigree in the neighboring quiet Sun that point radially away from the pore's center (Figure 5). In the standard HiFI observing mode, 2 × 100 images are selected within a 10 s period, as indicated by the gray vertical bars in the lower right panel of Figure 4. This observing mode relies on a special implementation of frame selection (Scharmer 1989; Scharmer & Löfdahl 1991; Kitai et al. 1997), where not just the best solar image but a set of high-quality images is chosen for image restoration.

Figure 4.

Figure 4. Temporal evolution of the seeing conditions at the GREGOR solar telescope on 2017 March 25. Three series of 50,000 images were captured at 135 Hz (top), and the corresponding MFGS values were computed and are plotted for G-band images. The gray rectangle covers a 2 min period, which is depicted at higher resolution (bottom left). A 10 s period at even higher temporal resolution highlights the frame selection process (bottom right). The best images, which are used for image restoration, are marked by gray vertical lines.

Standard image High-resolution image
Figure 5.

Figure 5. Decaying pore with light bridge observed in active region NOAA 12643 on 2017 March 25. All G-band images were restored with MFBD, with the exception of the best raw image (bottom left). The best nsel images from a 10 s time interval were used for the restored image, which is based on nmod restored Zernike modes. The rms-intensity contrast c and the MFGS value m are used as image-quality metrics. The images are scaled individually between minimum and maximum intensity, and the FOV is 16'' × 16''.

Standard image High-resolution image

Obviously, even at this very high cadence, the image-quality metrics still show strong variations, despite some clustering of the best images in Figure 4, indicating that seeing fluctuations occur at even higher frequencies. Considering what lies ahead for high-resolution imaging, the data acquisition rates for the next generation of 4k × 4k pixel detectors recording at 100 Hz will amount to about 3 GB s−1. This exceeds today's typical data acquisition rates of less than 1 GB s−1 but demonstrates that the data challenge will persist for years to come and, in particular, for the next generation of large-aperture solar telescopes, such as DKIST and EST. In Denker et al. (2018), we demonstrated that an image acquisition rate of facq = 50 Hz is an appropriate choice, considering the marginal benefits in quality for the frame-selected images with respect to the demands on camera detectors, network bandwidth, data storage, and computing power.

Multi-frame blind deconvolution (MFBD; Löfdahl 2002) is another commonly used image-restoration technique in solar physics. Figure 5 compiles some results based on the best set of G-band images. In Denker et al. (2018), we established a benchmark for G-band images, i.e., an MFGS value of m = 0.65, where image restoration with the speckle-masking technique becomes possible. Thus, the seeing conditions were only moderate-to-good on 2017 March 25. The aforementioned threshold was determined for the full FOV of the sCMOS detector, whereas the present observations used a much smaller FOV covering the immediate neighborhood of the AO lock point. The top row of Figure 5 demonstrates that, even compared to telescopes with 0.7–1 m apertures, data obtained with larger 1.5 m class telescopes require a significant increase in the number of Zernike modes (nmod ≈ 300) in the restoration process. This significantly increases the computation time and poses challenges for imaging with 4 m class solar telescopes. MFBD has an advantage over speckle masking because it requires a smaller number of images for a restoration, in particular, when the images are of high quality taken under very good or excellent seeing conditions. Thus, restored time series with higher cadence become possible. Already, nsel = 20 selected images deliver good restorations. Note that MFGS loses its discriminatory power for restored images, so that other metrics, like the rms-intensity contrast c, become more important. A likely explanation is that the MFGS metric is sensitive to the fine-structure contents of an image, which is mainly encoded in the phases of the Fourier-transformed image. Thus, once almost diffraction-limited information is recovered, the MFGS metric reaches a plateau. The image contrast, on the other hand, is more closely related to the Fourier amplitudes.

3. sTools Data Pipeline

The Optical Solar Physics research group at AIP operates two of GREGOR's facility instruments with GFPI and HiFI. The software package sTools (see Kuckein et al. 2017a for a brief introduction) provides, among other features, the data-processing pipeline for these two instruments. Major parts of the software were developed from 2013 to 2017 within the SOLARNET14 project, which is a "Research Infrastructures for High-Resolution Solar Physics" program following the Integrated Infrastructure Initiative (I3) model supported by the European Commission's FP7 Capacities Program. The software package was written from scratch but builds on the code development for and experiences gained with data reduction tools for the Göttingen Fabry–Pérot Interferometer (Bendlin et al. 1992; Puschmann et al. 2006; Bello González & Kneer 2008) and the Interferometric BIdimensional Spectropolarimeter (IBIS; Cavallini 2006).

Mainly written in IDL, sTools utilizes other IDL libraries with robust and already validated programs whenever available. The SolarSoftWare (SSW) system (Bentley & Freeland 1998; Freeland & Handy 1998), for example, offers instrument-specific software libraries and utilities for ground-based instruments and space missions, which includes database access and powerful string processing for metadata. The MPFIT (Markwardt 2009) package is primarily used for spectral-line fitting; the Coyote Library for image processing, graphics, and data I/O (Fanning 2011); and the NASA IDL Astronomy User's Library15 for reading and writing data in the Flexible Image Transport System (FITS; Wells et al. 1981; Hanisch et al. 2001) format. Data from imaging spectropolarimetry benefits especially from the possibility of writing FITS image extensions with individual headers (Ponz et al. 1994); i.e., polarization state, wavelength position, and calibration information are saved along with the corresponding filtergram. The result is a compact (a few GB), self-describing data set that can serve as input for various spectral analysis and inversion codes.

Computationally intense applications, in particular image restoration, make use of parallel computing implemented in other programming languages. Here, the IDL programs typically condition the input data and collect the output data for further processing. Currently, sTools provides interfaces for the Kiepenheuer Institute Speckle Interferometry Program (KISIP; Wöger & von der Lühe 2008) and MOMFBD. Recently, we started making use of the IDL built-in functions for parallel processing, so that other time-consuming parts of the data processing are also more efficiently implemented.

All newly developed IDL routines use the prefix stools_ to avoid name space collisions with other libraries. No subfolders exist within sTools, with an exception for documentation and individual IDL routines from external sources, which are not part of the aforementioned libraries. Instrument-specific and functional dependencies are declared in the naming schema for routines, e.g., stools_gfpi_ for GFPI data processing or stools_html_ for creating summary web pages for the observed data sets. In general, we aim to separate instrument-specific code from multipurpose routines, which can be shared among applications. The majority of the sTools programs are independent of the computer hardware and the site where the software is installed. Site specifications are encapsulated in structures, which are defined in specific configuration routines (stools_cfg_). The same applies to expert knowledge, e.g., about spectral lines, filter properties, camera settings, telescope details, etc., which is collected in configuration routines that return structures with the required parameters based on tag names. For example, many fitting routines rely on information regarding the width of the spectral line and spectral sampling. Based on the configuration information, the most suitable and validated fit parameters are chosen. The configuration routines are the only programs, which may have to be adapted for specific sites, or if new observing modes are carried out with different filters and spectral lines.

While writing the sTools routines, we placed special emphasis on proper documentation using standard IDL headers, which can be extracted with the doc_library procedure, on meaningful in-line comments, descriptive variable names, and a consistent programming style and formatting. Testing routines for specific data-processing steps is typically performed for several data sets with different observing characteristics. This ensures that new calibration steps or updated procedures do not lead to unintended consequences. In some cases, when implementing new observing setups for scanning multiple lines or using nonequidistant line sampling, major conceptual changes of the code become necessary. These test data are used to ensure that previously reduced and calibrated data are unaffected by the aforementioned changes. If changes affect already calibrated data, they will be reprocessed. The version of sTools is tracked so that, if needed, data calibrated with older versions can be recovered. These changes are documented on the project website, and major updates will be accompanied by data-release publications in scientific journals. Initial data calibration for one data set takes about 1 day of computation time on a single processor. On the other hand, reprocessing all GFPI quick-look data, which utilizes the already calibrated data, takes just a few days. The latter data serve as a benchmark to validate the performance of new or updated programs. The sTools data pipeline is currently installed at AIP and the German solar telescopes on Tenerife. The source code is maintained internally on an Apache Subversion16 (SVN) version control system. The latest released version can be downloaded by registered users as a tarball from the GREGOR website at AIP (http://gregor.aip.de).

4. Data Management and Data Archive

4.1. Point of Departure

High-resolution solar observations are taken at 1 m class ground-based solar telescopes around the world and with Hinode/SOT from space, where the latter provides, in many respects (i.e., imaging and spectropolarimetry), the closest match to HiFI and GFPI data. Consequently, scientists who work with such high-resolution data are among the primary target groups for utilizing GFPI and HiFI data. In addition, we especially foresee close interactions with the solar physics community on spectral inversion codes and numerical modeling of the highly dynamic processes on the Sun. Therefore, our immediate priority is providing a CRE fostering collaborations among researchers with common interests. Our goal is to raise awareness of and stimulate interest in complex and heterogeneous spectropolarimetric data sets, which are inherent trademarks of imaging spectropolarimeters with a broad spectrum of user-defined observing sequences. Offering a data repository to the whole solar physics community or the general public at large will enhance the impact of these high-resolution data products. However, in this case, it may be advantageous to highlight key data sets obtained under the best seeing conditions or of particularly interesting events like solar flares. In any case, data access, as described in the data policy (Section 4.2), is granted to everyone who registers for the GREGOR GFPI and HiFI data archive. At a later stage, when the high-level data products have matured, a more differentiated access will be instantiated, dropping the registration requirement for the publicly available data.

The data-specific challenges for GFPI and HiFI are as follows.

  • 1.  
    The campaign- and PI-oriented nature of the data, which results in different combinations of post-focus instruments with changing setup parameters, e.g., selection of diverse spectral lines, dissimilar spectral and spatial sampling, and different cadences.
  • 2.  
    Observations through Earth's turbulent atmosphere, which brings about data with fast-changing quality and necessitates image restoration, where different restoration algorithms introduce multiplicity in the major data levels.
  • 3.  
    The complexity of the data sets, which requires considerable efforts to condition the data products for a broader user base—even at the level of quick-look data.
  • 4.  
    The availability of personnel and financial resources for maintaining long-term access to data and for quality assurance beyond the typical funding cycle of third-party financial support.

Fortunately, AIP's mission includes the development of research technology and e-infrastructure as a strategic goal. Thus, the collaboration between e-science and solar physics allowed us to develop a tightly matched solution to the aforementioned data-specific challenges.

4.2. Data Policy

The GREGOR consortium (i.e., the Kiepenheuer Institute for Solar Physics, Max Planck Institute for Solar System Research, and Leibniz Institute for Astrophysics Potsdam) and GREGOR partners (i.e., the Instituto de Astrofísica de Canarias and Astronomical Institute of the Academy of Sciences of the Czech Republic) agreed that, in principle, all data shall be publicly available. In the following, we use the term "consortium" when referring to GREGOR members and partners.

Since observing proposals contain proprietary information and original ideas of PIs and their teams, the data of PI-led observing campaigns will be embargoed for 1 yr. The embargo period can be extended upon request for another year when observations are related to PhD theses. Quick-look data are publicly available after storage at the GREGOR GFPI and HiFI archive and subsequent processing, typically after 4–6 weeks, and these data are not subject to the embargo period. This way, the PI of an observing campaign can be contacted to inquire about potential collaborations even within the embargo period. The consortium encourages, but does not require, collaboration with the originator of the data. All work based on GREGOR data is required to include an acknowledgment (see below), and the consortium asks the authors to include appropriate citations to the GREGOR reference articles published in 2012 as a special issue of Astronomische Nachrichten (Vol. 333/9). A detailed version of the data policy is publicly available on the GREGOR website at AIP.

4.3. GREGOR GFPI and HiFI Data Archive

The archive is based on the Daiquiri17 framework, which was developed by AIP's R&D group "Supercomputing and E-Science." Daiquiri is designed for creating highly customized web applications for data publication in astronomy. The features of Daiquiri comprise rich tools for user management, an SQL query interface to enable users to directly enter database queries via the website, means to download the results of queries in different formats following standards of the International Virtual Observatory Alliance18 (IVOA; Quinn et al. 2004), and plotting functions. The interaction with Daiquiri can be scripted using its VO universal worker service (UWS; Harrison & Rixon 2016) interface.

For the GREGOR archive, we use Daiquiri's user management to implement a custom user-registration workflow. New users first register on the GREGOR data portal, and they do not have to belong to the GREGOR consortium. Therefore, registration is possible for anyone with an interest in GREGOR high-resolution data. To emphasize the intended collaborative nature of this research infrastructure, in the following, we use the terms "GREGOR collaboration" or "collaborators" collectively for all registered users. After registration, one of the project managers confirms the new user, again through the GREGOR portal. Only after this organizational confirmation takes place will members of the technical staff activate the user. This procedure is necessary because users receive CRE privileges allowing them processing data on institute-owned computers and editing of websites and blog entries.

Along with their accounts for the web portal, users also obtain Secure Shell (SSH) login credentials for the data access node (Section 4.4). The SSH protocol facilitates efficient and fast distribution of the data products while ensuring modern security and maintainability (in particular, firewall configuration) when compared to the popular but insecure File Transfer Protocol (FTP). On the data access nodes, an elaborate permissions system ensures access restrictions and data security. This is implemented using Linux Access Control Lists (ACLs), which extend the usual file permissions (user/group/world) common on UNIX systems. Users are organized in groups that gain write permissions to certain subdirectories of the GREGOR archive. These ACL directories, which typically comprise data for a specific instrument, data-processing level, and observing day, can have multiple groups, allowing for fine-grained access by the registered users (e.g., when accessing embargoed data), as well as the GREGOR archive administrators.

Besides the user-registration workflow, Daiquiri is also used to set up the GREGOR web pages, some with access restrictions, some public. This includes the data products generated by the sTools data-processing pipeline (see Section 3). Access to the data was initially limited to the GREGOR consortium but is now open to all registered users of the GREGOR GFPI and HiFI data archive. The data sets generated by sTools are currently in the process of being converted to FITS files containing image extensions. This reduces the level of complexity, as metadata are available, and many images of a spectral scan or in a time series are already aggregated. Once converted to FITS format, these data will be integrated into an SQL database and can be accessed using Daiquiri's query functionality.

4.4. Collaborative Research Environment

The large number and high resolution of images and spectra, as well as the computational effort for their postprocessing, demands capable and efficient structures for storage and data management. To make the best use of GREGOR data and to encourage their usage, we implemented a dedicated CRE at AIP. This research infrastructure acted initially as a central hub for the storage and processing of different data products, as well as their distribution within the GREGOR consortium, but it is now open to all interested scientists. The CRE provides data space and data access with different levels of authorization, in addition to computational resources and customized tools for analysis and processing. Participation in the CRE is managed by the GREGOR consortium lead for GFPI and HiFI. Finally, collaborators have the option via the CRE to publish selected and curated "science-ready" data for the solar community, including a minted DOI registered with DataCite.19

Over the last decade, AIP provided similar CREs for several projects. For collaborations working on simulations of cosmological structure formation, such as Constrained Local UniversE Simulations (CLUES; Gottlöber et al. 2010) and MultiDark (Riebe et al. 2013), AIP hosts hundreds of TB of file storage and a relational database with about 100 TB of carefully curated particle information, halo catalogs, and results of semi-analytical galaxy models. These data products are available via the CosmoSim20 database provided by the E-Science group at AIP. However, observational collaborations, e.g., the Multi Unit Spectroscopic Explorer21 (MUSE; Bacon et al. 2010; Weilbacher et al. 2014) collaboration and the Radial Velocity Experiment22 (RAVE; Steinmetz et al. 2006) survey, also rely on a CRE for data management and processing.

The CRE hardware is integrated into the Almagest cluster at AIP. The whole cluster consists of over 50 nodes and about 3 PB of raw disk space. The different machines are connected through a high-performance InfiniBand network. Throughout the cluster, we use the Linux distribution CentOS as an operating system. Directly allocated for the GREGOR CRE are the following.

  • 1.  
    One shared compute node acting as a login node and to share the data among the collaboration. Collaborators can log in to this machine using the SSH protocol and copy data to their workstations for further processing and scientific analysis.
  • 2.  
    Two storage nodes with 80 TB of storage space reserved for GREGOR data. Each of the nodes contains 24 hard disks, which are combined into one logical RAID volume using a ZFS23 file system. Using ZFS's send-receive feature, the disk content of the two nodes is mirrored, and they are physically located in different buildings to minimize the risk of data loss due to catastrophic events.
  • 3.  
    A dedicated compute node, hosting the sTools pipeline, which is directly connected to the archive. This machine allows users to run their own programs and to visualize GFPI and HiFI (raw) data. This computer is also used for generating and updating the web pages with quick-look data. Remote users can log in by SSH and use virtual network computing (VNC) for a desktop-like environment.
  • 4.  
    Additional resources are supplied to host the web applications for the data archive, connecting to the internet, and backups.
  • 5.  
    Two internal compute nodes with mirrored installations of the sTools data-processing pipeline. One of the nodes with a 64-core processor and 256 GB RAM is dedicated to image restoration and hosts the SVN repository of the sTools code, including external libraries. The workstations of the researchers on the AIP campus can NFS-mount the respective volumes containing data and software libraries so that data processing and analysis are also possible locally.
  • 6.  
    In addition, a copy of the sTools data-processing pipeline is installed at the German solar telescopes at Observatorio del Teide, Izaña, Spain. The computer network includes workstations and a dedicated multicore computer for data processing and image restoration on site. In particular, the MFGS-based image selection for the HiFI data is carried out immediately after the observations with an easy-to-use GUI written in IDL as an interface to sTools.

4.5. Data Levels

We distinguish three major levels of data products within the GREGOR GFPI and HiFI data archive.

  • 1.  
    Level 0 refers to raw data acquired with GFPI and HiFI. The data are written in a format native to the DaVis software of LaVision, which runs both instruments. A short ASCII header declares the basic properties of the images (e.g., DaVis version number, image size in pixels, number of image buffers, type of image compression, if applicable, etc.), which is followed by either compressed or uncompressed binary data blocks. Another free-format ASCII header, containing auxiliary information like a time stamp with microsecond accuracy, is placed after the data blocks at the end of the file. This time stamp results from the programmable timing unit (PTU), which provides external trigger signals for image acquisition. Specific settings of the observing mode and instrument parameters are saved in additional text files for each image sequence or spectral scan. These set files include, for example, telescope and AO status, as well as a separate time stamp for the observing time based on the camera computer's internal clock, which is synchronized with a local GPS receiver at the observatory.
  • 2.  
    The huge amount of large-format, high-cadence HiFI data (up to 4 TB day–1 with the current setup) requires reducing the data already on site, directly after the observations. This is the standard procedure and includes dark and flat-field corrections. In addition, the image quality is determined with the MFGS metric, which facilitates frame selection. Only the best 100 out of 500 images in a set are kept. The calibrated image sequences (level 1) of the two synchronized sCMOS cameras are written as FITS files with image extensions. The metadata contained in the primary and image headers are partially SOLARNET-compliant (level 0.5), which means that they do not contain all mandatory SOLARNET keywords and have not used any SOLARNET/FITS standard keywords in a way that is in conflict with their definitions. HiFI level 0 data are typically deleted once the frame-selected and calibrated level 1 data are safely stored in the GREGOR archive. HiFI level 0 images are only kept when the seeing conditions were excellent, fast or transient events were captured, or special observing programs were carried out (see Denker et al. 2018). The processing time for restoring a single HiFI image (level 2) from a set of 100 images with KISIP takes several tens of minutes on the 64-core compute node.
  • 3.  
    Level 1 GFPI data are typically created at AIP after the observing campaign, which reflects the high complexity of data from imaging spectroscopy. In most cases, the multiple images per wavelength point are simply destretched and coadded with no image restoration. However, all other calibration steps (e.g., alignment of narrowband and broadband images, blueshift and prefilter-curve corrections) are carried out so that scientific exploitation of level 1 data is possible. After preprocessing level 1 data, i.e., after dark and flat-field corrections and determining the alignment of narrowband and broadband images, a copy of the narrowband and broadband images is saved that serves as the starting point for image restoration (level 2), thus avoiding preprocessing level 0 data twice. Only the best spectral scans are then chosen for level 2 processing, i.e., image restoration with MOMFBD or speckle deconvolution. The processing time of a single scan with MOMFBD takes several hours on the 64-core compute node.
  • 4.  
    Level 1 data are the starting point for creating quick-look data products such as time-lapse movies, Doppler velocity maps, and overview graphics for seeing conditions and observing parameters.
  • 5.  
    Level 2 data are restored data using MOMFBD and KISIP (see Section 3). These data-processing steps are only included on demand, considering the significant amount of computational resources. The best image restoration scheme is chosen by the researcher working with the data, and it is not unusual to select different ways to restore images or spectral scans depending on user preferences or a specific science case. Once the spectral scans are restored, other calibration steps still need to be applied, such as blueshift and prefilter-curve corrections, before physical parameters such as Doppler velocities or other spectral-line properties can be determined.

Level 1 HiFI data (and occasionally level 0) and level 0 GFPI data are transferred from the GREGOR telescope to AIP by regular 2.5 inch external hard-disk drives with 2–4 TB storage capacity. Smaller data sets are often transferred over the internet. However, the physical transfer using hard disks is preferred over copying over the internet due to limited bandwidth and network reliability at the observatory site. Data from all GREGOR instruments can be stored on site on a 100 TB storage array for up to 3 months. Keeping at least two copies of the data at different locations during the data transfer to the GREGOR archive mitigates potential data loss. Data integrity during the transfer is assured by monitoring the transfer logs and verifying that all files with correct sizes were transferred. The total amount of data that was acquired with GFPI and HiFI, as well as with the now-obsolete facility cameras of BIC, is summarized in the last row of Table 1. The other entries in the rows for 2014–2016 refer to the number of scientific data sets that were obtained in various observing campaigns. The data volume refers to the sum over all data levels. GFPI level 1 data are roughly twice the size of level 0 data. Furthermore, the bulk of the data volume arises from level 0 data for BIC and level 1 data for HiFI.

Table 1.  Number of Data Sets and Stored Data Volume for GREGOR's Instruments

  GFPI BIC HiFI
2014 48 37  
2015 69 42  
2016 41 4 46
2017 35   17
  10.0 TB 13.6 TB 20.3 TB

Note. The abbreviations are defined as follows: GFPI: GREGOR Fabry-Pérot Interferometer; BIC: blue imaging channel; HiFI: High-resolution Fast Imager.

Download table as:  ASCIITypeset image

Finally, 70 users are currently registered in the CRE and are mainly from the GREGOR consortium. However, increasingly, external users (at the moment, 10) have registered with the CRE, most of them participating in (coordinated) observing campaigns or the SOLARNET Access Program promoting observing campaigns with Europe's telescopes and instruments for high-resolution solar physics. In the meantime, routine observation started at GREGOR in 2016, and we already see an influx of new international collaborators who will certainly broaden the user base of the GREGOR CRE.

4.6. Use Case: GREGOR Early Science Phase

The GREGOR early science phase took place in 2014 and 2015, where members of the GREGOR consortium collaboratively carried out observing campaigns, i.e., in 2014 with the individual instruments and in 2015 with multi-instrument setups. Notably, a 50 day observing campaign with GFPI and BIC (the predecessor of HiFI) was carried out in 2014 July and August. In a joint effort, scientists from all involved institutes submitted observing proposals, which were evaluated and condensed into a list of top-priority solar targets and feasible observing modes and strategies. The observations were carried out by experienced observers of all institutes together with novice observers (not necessarily novice scientists) to strengthen their observing skills and familiarize them with the new instruments. The data, which were acquired with many different setups, were used to develop, test, and improve the sTools data-processing pipeline. Finally, the level 1 data were stored in the GREGOR GFPI and HiFI data archive so that all scientists were able to access and analyze them. Collaborations on specific studies, with the aim of publishing them in scientific journals, were coordinated using the blog facilities of the Daiquiri framework. This also includes the organization of GREGOR science meetings at AIP. The News Blog of the GREGOR CRE allowed us to publish poster presentations given at solar physics meetings (e.g., the SOLARNET IV meeting in Lanzarote in 2016) and newly appearing journal articles and conference proceedings to a broader section of the solar physics community and the interested public. The first results of the GREGOR early science phase were published in 2016 in the journals Astronomy & Astrophysics (Vol. 596) and Astronomische Nachrichten (Vol. 337/10), and more up-to-date GFPI and HiFI science publications are referenced in Section 2.2.

4.7. Future Plans

In the future, we plan to expand the data archive and data access infrastructure considerably. With a release of the data to the general solar community, new means of data access will be necessary beyond users actively utilizing the CRE to interact with like-minded researchers. As before, the latter type of access will be managed via registration and subsequent confirmation by the collaboration. While collaborators will still be able to retrieve data through SSH connections, this is not suitable for public access by the general scientific community.

Therefore, we will extend the GREGOR archive toward a public data portal, which will offer a full search on all files in the archive (unless they are still embargoed) and downloads using the HTTP protocol, either through the browser or using command line tools. The selection of files will be based on the metadata of the level 1 and 2 data of the GFPI and HiFI instruments. The archive will also allow for SQL queries to create custom result sets based on any scientific criteria. Standards defined and adopted by IVOA, like UWS and Table Access Protocol (TAP), will expedite accessing the data through clients using interoperable application programming interfaces (APIs). The results of the queries of a user and the queries themselves will be stored in a personal database to be retrieved at the user's convenience without repeating the query to the whole GREGOR database.

5. Conclusions and Prospects

High-resolution solar observations are confronted with short timescales in both the Sun's and the Earth's atmospheres. Whereas the first is related to the evolution and dynamics of solar features, the latter imposes strong boundary conditions for image restoration. The underlying assumption is that the observed object is not changing. Thus, a contiguous data set has to be recorded within a few seconds. This interval becomes even shorter when observing faster features like in the chromosphere or with larger telescopes offering higher spatial resolution and consequently smaller "diffraction-limited" pixels. These timescales affect instrument design, observing modes and strategies, and, in the end, data management and archiving.

Providing high-resolution data to the solar community is an ongoing process; i.e., the GREGOR GFPI and HiFI archive and the sTools data-processing pipeline are not static. At present, we are working on the polarimetric calibration routines, and they will be added to the sTools package once they are tested. The CRE implemented for the GREGOR telescope takes this into account and also contains provisions for future data access beyond the consortium and collaboration. Currently, our efforts are focused on characterizing image quality to identify the best data sets of HiFI images and GFPI spectral scans, which are both affected by the varying seeing conditions (see Denker et al. 2018). The algorithms and routines (e.g., with the MFGS method) developed for this purpose additionally allow us to monitor long-term trends in data quality, as well as to establish a database of seeing conditions at Observatorio del Teide.

Synoptic solar images, magnetograms, and Doppler maps with a typical spatial resolution of 1'' serve very successfully as input for feature identification and pattern recognition algorithms. In particular, in the context of space weather research and forecasting, tools were developed to enable easy access to such data products, e.g., the Heliophysics Event Knowledgebase (HEK; Hurlburt et al. 2012).

Even though the focus shifts from applications to more fundamental physics, a knowledge base for high-resolution data (spatial resolution below 0farcs1) is potentially very beneficial, bringing order to the plethora of small-scale features observed in the quiet Sun (e.g., G-band bright points, filigree, blinkers, etc.) and active regions (e.g., umbral dots, penumbral grains, Ellerman bombs, microflares, etc.). Morphological characteristics, photometric properties, and spectral/polarimetric features provide a wide parameter range, which can be stored in relational databases. A summary of image-processing techniques and various ways of performing feature tracking is given in Aschwanden (2010), and Turmon et al. (2010) demonstrated the capability of multidimensional feature identification. Database research can reveal relationships among solar small-scale features such as those of the photospheric network, which would otherwise be missed in case studies, and can track changes with the solar activity cycle, when the contents of the database grows over time (e.g., Muller & Roudier 1994; Jin et al. 2011; McIntosh et al. 2014; Roudier et al. 2017).

The GREGOR GFPI and HiFI archive and future associated relational databases are an attractive starting point for data-mining and machine-learning applications. In general, the huge amount of astronomical and astrophysical data has stimulated interest in effectively exploring them (see Ball & Brunner 2010; Ivezić et al. 2014 for current research in this field). Machine-learning algorithms gain knowledge from experience. A training data set teaches the underlying models to the machine, and new data sets can then be classified with the results from the initial training set. Furthermore, machine learning can be used in time-series and wavelet analysis. Data mining often employs machine-learning techniques when data sets become overwhelmingly large, extracting useful information from raw data and detecting new relations or anomalies. Furthermore, data mining and machine learning help validate model assumptions and ensure consistency. The performance of machine-learning models is mostly influenced by the amount and quality of the training data sets. A central repository with immediate access to calibrated high-resolution solar data, such as ours, speeds up the process of training neural networks or data mining.

As an example, DeepVel (Asensio Ramos et al. 2017) is a deep-learning neuronal network, estimating horizontal velocities at three different atmospheric heights from time series of high-resolution images. The training data were, in this case, numerical simulations. The advantage of DeepVel is not only its extremely fast execution as compared to other optical flow techniques but also its ability to infer velocity fields from subphotospheric layers. Machine learning will become increasingly helpful in image restoration and spectral inversions—both applications requiring significant computational resources.

The 1.5 m GREGOR solar telescope was built by a German consortium under the leadership of the Kiepenheuer Institute for Solar Physics in Freiburg with the Leibniz Institute for Astrophysics Potsdam, the Institute for Astrophysics Göttingen, and the Max Planck Institute for Solar System Research in Göttingen as partners and with contributions by the Instituto de Astrofísica de Canarias and the Astronomical Institute of the Academy of Sciences of the Czech Republic. We thank Dr. Michiel van Noort for his help in implementing the MOMFBD code at AIP. SJGM acknowledges the support of project VEGA 2/0004/16 and is grateful for financial support from the Leibniz Graduate School for Quantitative Spectroscopy in Astrophysics, a joint project of the Leibniz Institute for Astrophysics Potsdam and the Institute of Physics and Astronomy of the University of Potsdam. CD and REL were supported by grant DE 787/3-1 of the German Science Foundation (DFG). This study is supported by the European Commission's FP7 Capacities Program under grant agreement Number 312495. The AIP Almagest cluster and its CREs were partially funded by a European Regional Development Fund (ERDF) grant in 2012. Development of VO interfaces and facilities has been supported by the German Federal Ministry of Education and Research (BMBF) Collaborative Research Program for the German Astrophysical Virtual Observatory (GAVO). Development of Daiquiri has been partially supported by the BMBF grant "Survey-Competence." The Center of Excellence in Space Sciences India is funded by the Ministry of Human Resource Development, Government of India.

Facilities: GREGOR solar telescope (GFPI - , HiFI) - .

Software: DeepVel (Asensio Ramos et al. 2017), KISIP (Wöger & von der Lühe 2008), MOMFBD (Löfdahl 2002; van Noort et al. 2005), MPFIT (Markwardt 2009), SolarSoft (Bentley & Freeland 1998; Freeland & Handy 1998), and sTools (Kuckein et al. 2017a).

Footnotes

Please wait… references are loading.
10.3847/1538-4365/aab773