A publishing partnership

The following article is Open access

Identification of 4FGL Uncertain Sources at Higher Resolutions with Inverse Discrete Wavelet Transform

, , , , and

Published 2024 January 17 © 2024. The Author(s). Published by the American Astronomical Society.
, , Citation Haitao Cao et al 2024 ApJ 961 91 DOI 10.3847/1538-4357/ad0e6c

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

0004-637X/961/1/91

Abstract

In the forthcoming era of big astronomical data, it is a burden to find target sources from ground-based and space-based telescopes. Although machine-learning methods have been extensively utilized to address this issue, the incorporation of in-depth data analysis can significantly enhance the efficiency of identifying target sources when dealing with massive volumes of astronomical data. In this work, we focused on the task of finding active galactic nucleus (AGN) candidates and identifying BL Lacertae objects (BL Lac) or flat spectrum radio quasar (FSRQ) candidates from the 4FGL_DR3 uncertain sources. We studied the correlations among the attributes of the 4FGL_DR3 catalog and proposed a novel method, named fractal dimension–inverse discrete wavelet transform (FDIDWT), to transform the original data. The transformed data set is characterized as low-dimensional and feature-highlighted, with the estimation of correlation features by fractal dimension theory and the multi-resolution analysis by inverse discrete wavelet transform (IDWT). Combining the FDIDWT method with an improved lightweight MatchboxConv1D model, we accomplished two missions: (1) to distinguish the AGNs from others (non-AGNs) in the 4FGL_DR3 uncertain sources with an accuracy of 96.65% ± 1.32%, namely Mission A; and (2) to classify blazar candidates of uncertain type into BL Lacs or FSRQs with an accuracy of 92.03% ± 2.2%, namely Mission B. There are 1354 AGN candidates in Mission A, and 482 BL Lacs candidates and 128 FSRQ candidates were found in Mission B. The results show a high consistency of greater than 98% with the results in previous works. In addition, our method has the advantage of finding less variable and relatively faint sources than ordinary methods.

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 4.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

Active galactic nucleus (AGN) has been a hot topic for 60 yr in astronomy since its discovery in 1963 (Schmidt 1963). It is believed that AGNs centered a supermassive black hole, which is surrounded by an accretion disk, and the whole system provides energy for AGN radiation (Lynden-Bell 1969; Blandford & Znajek 1977; Blandford & Payne 1982). The emissions from AGNs are observed to span over the entire electromagnetic spectrum, and these emissions are found to be strong and variable. Based on the ratio of radio emission strength to optical emission strength, AGNs are divided into radio-loud and radio-quiet AGNs (Strittmatter et al. 1980; Kellermann et al. 1989). This method has recently been modified by a double-criterion method (Xiao et al. 2022a). Blazars, a subclass of radio-loud AGNs with jets pointing toward the observer, show multi-band high and fast variability, high and variable polarization, strong and variable γ-ray emissions, and apparent superluminal motion (Wills et al. 1992; Urry & Padovani 1995; Fan 2002; Villata et al. 2006; Fan et al. 2014, 2021; Gupta et al. 2016; Xiao et al. 2019, 2020a, 2022b; Abdollahi et al. 2020). Blazars consist of BL Lacertae objects (BL Lacs) and flat spectrum radio quasars (FSRQs), the former shows strong emission lines (rest-frame equivalent width, EW > 5 Å) and the latter demonstrates no or weak emission features (EW < 5 Å) (Urry & Padovani 1995; Scarpa & Falomo 1997).

Until the launch of the Large Area Telescope on board the Fermi Gamma-Ray Space Observatory (Fermi-LAT) in 2008, the study of blazars has been severely limited due to the small sample size. The Large Area Telescope has an unprecedented performance, a better energy resolution, angular resolution, and a wider effective area in both low-energy and high-energy bands 6 than its predecessor EGRET (Thompson et al. 1993), at energy ranges from 20 MeV to 300 GeV. Fermi-LAT collaboration has released 5 main γ-ray sources catalogs, namely 0FGL, 1FGL, 2FGL, 3FGL, and 4FGL (Abdo et al. 2009, 2010; Nolan et al. 2012; Acero et al. 2015; Abdollahi et al. 2020). However, there are still significant Fermi sources that are unrelated to any known class, e.g., 1010 unassociated sources in 3FGL and 2291 uncertain sources (2157 unassociated sources + 134 unknown sources) in the latest 4FGL_DR3, 573 blazar candidates of uncertain types (BCUs) in the 3FGL, and 1493 BCUs in the latest 4FGL_DR3.

It is time consuming to verify these uncertain sources individually through optical observation, thus more efficient methods must be explored. Many machine-learning (ML)-based algorithms have been employed for this issue. For instance, Saz Parkinson et al. (2016) applied two different ML methods—Random Forest (RF) and Logistic Regression (LR)—to identify 1008 unassociated sources in 3FGL. Among them, 334 sources were predicted as pulsars (PSRs), and 559 sources were predicted as AGNs. Chiaro et al. (2016) utilized blazar flaring patterns (B-FlaP) as an identification approach on BCUs. Since variability is one of the characterizing properties of blazars (Paggi et al. 2011), the light curve of blazars was used by an artificial neural network (ANN) to identify 573 BCUs in 3FGL. These BCUs were associated with 342 BL Lacs and 154 FSRQs, and 77 sources remained uncertain. Xiao et al. (2020b) carried out an ensemble ML method, picked out 748 AGNs candidates from 1010 3FGL unassociated sources, and identified 573 BCUs to be 326 BL Lac candidates and 247 FSRQ candidates. Moreover, Kang et al. (2019) studied the classification of 1312 BCUs from 4FGL_DR1 via three supervised ML methods and obtained 724 BL Lac and 332 FSRQ candidates.

The task of Fermi source classification can be seen as feature extraction with ML methods due to their ability to learn patterns from data and provide valuable insights, decisions, and predictions (Jordan & Mitchell 2015; Zhou et al. 2017). However, traditional ML methods are limited when dealing with the increasing volume of big astronomical data with the successful launch of more telescopes and detectors. In recent years, the popularity of graphics processing units had led to the research of deep learning (DL) to learn features from massive-scale data using deep neural networks (DNNs). This has become a major research focus in ML (Yu & Deng 2010; Liu et al. 2017). DNNs have proven to be successful in various real-world applications (Chen et al. 2018; Alemany et al. 2019; Jifara et al. 2019; Lam et al. 2019; Zyner et al. 2019). Furthermore, it has been shown that more complex problems require deeper networks (Bengio et al. 2009; He et al. 2016), which has led to the development of sophisticated networks such as VGG (Simonyan & Zisserman 2014), ResNet (He et al. 2016), and ChatGPT (Brown et al. 2020).

However, rather than solely focusing on the deep structure of networks that helps to learn intrinsic features, attribute analysis should also be emphasized. We propose that there exists a correlation feature among the attributes of raw data, and it will further improve the learning performance. In addition, numerous studies have shown that real-world data often contain highly redundant and unimportant attributes (Bakshi & Stephanopoulos 1993; Bengio et al. 2009; Glorot et al. 2011). This redundancy can lead to sparsity in high-dimensional space, where most samples in the data set are far away from each other. In classification tasks, this sparsity can result in less reliable predictions than in low dimensions because predictions are based on larger extrapolations (Géron 2017). Therefore, we believe that attribute analysis presents an opportunity for better results through additional correlation features and dimension reduction.

In this paper, we focus on two missions for 4FGL_DR3, i.e., classifying 2291 uncertain sources into AGN or non-AGN and associating 1493 BCUs to BL Lac or FSRQ, which are named Mission A and Mission B, respectively. First, we study some popular attribute analysis methods and briefly examine the attributes of 4FGL_DR3 sources. We then find the core attributes based on the fractal dimension (FD) theory and step into the research of multi-attribute analysis from the perspective of the whole data set. Based on the results, we propose a novel method, called fractal dimension–inverse discrete wavelet transform (FDIDWT), which is a combination of FD and inverse discrete wavelet transform (IDWT), to extract correlation features at a higher resolution. With FDIDWT, the original data set is transformed into a low-dimensional and feature-highlighted set, which benefits the subsequent learning process. Finally, we combine the FDIDWT method with a lightweight convolutional neural network (CNN) model as our contribution to accomplish the classification missions.

This paper is organized as follows. Section 2 describes the data sets in two missions and presents some commonly used attribute analysis methods. Based on that, Section 3 interprets our proposed method in detail. The experiments and results are reported in Section 4. Further discussions and conclusions are presented in Sections 5 and 6, respectively.

2. Data Sets and Attribute Analysis

2.1. Samples of 4FGL_DR3

The Fermi-LAT collaboration has recently released the incremental version of the 12 yr Fermi-LAT Gamma-ray Source Catalog (4FGL_DR3; Abdollahi et al. 2022), which contains 6659 sources, among which 3809 sources are AGNs, 559 sources are associated with non-AGNs (including pulsars, high-mass binaries, and supernova remnants, etc.), and 2291 uncertain sources (134 sources are associated with counterparts of unknown nature and 2157 unassociated sources). Within the AGNs, there are 3743 sources confirmed with blazars, among which 1456 blazars are associated with BL Lacs, 794 blazars are associated with FSRQs, and the other 1493 Blazars are BCUs that have not been tagged as BL Lacs or FSRQ.

To accomplish Mission A and B, we need to select features that can distinguish one from another. Variability is one of the widely known characteristics of AGNs with respect to others that are detected by Fermi-LAT (Abdollahi et al. 2020, 2022). Therefore, those features that contain variability information (Flux1000, Flux_Band, Variability_Index, and Frac_Variability) should be included for accomplishing Mission A. While FSRQs and BL Lacs demonstrate significant γ-ray spectra, the GeV γ-ray regime situates at the different regimes of the higher hump of blazar spectral energy distribution (Fan et al. 2016; Yang et al. 2022, 2023). Those features that contain spectral information (Flux_Band, Pivot_Energy, PL_Index) should be included for Mission B.

In this case, we compile the data of 13 attributes from 4FGL_DR3, as listed in Table 1.

Table 1. The 13 Selected Attributes from 4FGL_DR3

IndexAttributeDescription
a1 Pivot_EnergyEnergy at which error on differential flux is minimal
a2 Flux1000Integral photon flux from 1 to 100 GeV
a3 PL_IndexBest fit power-law index
a4 Variability_IndexSum of 2A-log (Likelihood) difference between the flux fitted in each
  time interval and the average flux over the full catalog interval
a5 Frac_VariabilityFractional variability computed from the fluxes in each year
a6 Flux_Band1Integral photon flux in the spectral band 0.05–0.1 GeV
a7 Flux_Band2Integral photon flux in the spectral band 0.1–0.3 GeV
a8 Flux_Band3Integral photon flux in the spectral band 0.3–1 GeV
a9 Flux_Band4Integral photon flux in the spectral band 1–3 GeV
a10 Flux_Band5Integral photon flux in the spectral band 3–10 GeV
a11 Flux_Band6Integral photon flux in the spectral band 10–30 GeV
a12 Flux_Band7Integral photon flux in the spectral band 30–100 GeV
a13 Flux_Band8Integral photon flux in the spectral band 100–1000 GeV

Download table as:  ASCIITypeset image

For Mission A and Mission B, we randomly split the data into three subsets with a ratio of 8:1:1, as shown in Figure 1. The training set is used to fit the ML methods or DNN models, and the validation set aids in fine-tuning hyper-parameters for better performance. Finally, the generalization capability of the model is independently evaluated on the test set. Moreover, to explore the robustness of the model, this split policy was repeated 10 times and in the end 10 data sets were prepared for each mission. In this section, we analyze the attributes of the training set with the methods shown in Figure 2.

Figure 1.

Figure 1. Data Set A (upper panel) and Data Set B (lower panel) for Mission A and Mission B, respectively.

Standard image High-resolution image
Figure 2.

Figure 2. The commonly used three attribute analysis methods and our proposed method (gray color).

Standard image High-resolution image

2.2. Attribute Importance

As mentioned earlier, it is found that real-world data often contain highly redundant and unimportant attributes. The decision tree (DT) is a popular technique to estimate the importance of attributes. The attributes that appear in the tree are considered important, with the frequency of their appearance being their attribute importance. An attribute that appears less frequently is assumed to be less important. RF is composed of multiple DTs and reduces the bias in estimating attribute importance (Breiman 2001). It is always the practice to remove the unimportant attributes and perform the so-called attribute selection (AS) for dimension reduction.

We employ 50,000 DTs to build the RF and take the entropy as the criterion. Other hyper-parameters align to the default values of scikit-learn (Pedregosa et al. 2011). The results of attribute importance on the training sets are averaged over 10 data sets and are shown in Figure 3. Interestingly, from the results, we find that the four most important attributes are Pivot_Energy, PL_Index, Variability_Index, and Frac_Variability in both Data Set A and B.

Figure 3.

Figure 3. The average attribute importance of the training sets in Data Set A (left) and Data Set B (right).

Standard image High-resolution image

2.3. Principal Component Analysis

Another method to study attributes is to project the data from the attribute space to a new space. In this new space, the direction with maximum variance is considered to contribute most to the data, while the directions with small variances can be removed without sacrificing crucial information. This method is called principal component analysis (PCA) (Jolliffe & Cadima 2016). For each split data set, we normalize the samples and perform PCA on training sets. The variance ratio is averaged and depicted in Figure 4. We find that most of the information of the training data is concentrated in the first three components (variance ratio larger than 0.1) in both Data Set A and B.

Figure 4.

Figure 4. The average variance ratio of the training sets in Data Set A (left-hand panel) and Data Set B (right-hand panel).

Standard image High-resolution image

2.4. Attribute Significance Estimator based on the Fractal Dimension (FDASE)

Even though the RF model aggregates the attribute importance values from all the DTs (see Section 2.2), it may result in redundancy because it does not consider the correlations among important attributes. To address this issue, the FD theory is applied to estimate the potential contribution of each attribute to the data set and measure the correlations among attributes (Belussi & Faloutsos 1995, 1998). This method is based on the observation that independent attributes contribute more to the data set, while correlated attributes contribute less.

Mathematically, for a data set A with E attributes: ${\mathbb{A}}=\{{a}_{1},{a}_{2},\ldots ,{a}_{E}\}$, the FD theory supposes that the potential existence of correlated attributes leads the set of points in the original E-dimensional space to describe one spatial object in a dimension that is lower than or equal to E. The dimension of the object represented by the data set is called the intrinsic dimension (ID), denoted by D, $D\in {{\mathbb{R}}}^{+}$. The ceiling of the ID ⌈D⌉ is the minimum number of attributes that must be retained to keep the essential characteristics of a data set (de Sousa et al. 2007).

However, the ID of a data set is difficult to obtain. Alternatively, we consider the ID projecting the data set on a subset C A , where C is defined by an attribute subspace ${\mathbb{C}}\subset {\mathbb{A}}$. It is named partial intrinsic dimension (PID) on ${\mathbb{C}}$: ${pD}({\mathbb{C}})$. Based on the definitions, the attribute ${a}_{i}\in ({\mathbb{A}}-{\mathbb{C}})$ increases ${pD}({\mathbb{C}})$ by at most its individual contribution (IC) according to the degree of the correlation between ai and the attributes in ${\mathbb{C}}$. The IC of ai , i.e., iC(ai ), is the maximum potential contribution of the attribute ai to ${pD}({\mathbb{C}})$. The greater the correlation between ai and the attributes in ${\mathbb{C}}$, the lower its contribution to ${pD}({\mathbb{C}})$.

In addition, iC(ai ) can be measured by pD({ai }) and it ranges in [0, 1]. A more independent distribution of the values of ai leads to iC(ai ) closer to one, while a more structured distribution brings iC(ai ) closer to zero (de Sousa et al. 2007). Thus, the E-dimensional data set can be seen as formed by adding the attributes with different contributions to the D-dimensional sub-data set.

Moreover, the degree of correlations among attributes can be measured by a threshold ξ: a sub-data set B A is said to be ξ-correlated to another sub-data set C A (their attribute spaces ${\mathbb{B}}\cap {\mathbb{C}}=\varnothing $) if every attribute ${a}_{i}\in {\mathbb{B}}$ does not contribute more than ξiC(ai ) to ${pD}({\mathbb{C}})$. The threshold ξ ∈ [0, 1) tunes how strong the correlation between attributes in ${\mathbb{B}}$ and attributes in ${\mathbb{C}}$ should be to be detected.

A greedy algorithm FDASE was developed by de Sousa et al. (2007) to find a subset of attributes whose PID approaches the ID of the whole data set. The resulting subset is called the attribute set core ξ C (ASC), with the given correlation threshold ξ and scale range n. Therefore, the ratio pD(ξ C)/ID normalizes the contribution of ASC to the whole data set. Similarly, for each split data set, we utilize the FDASE algorithm on the training sets with scanning correlation threshold ξ and compare the value of pD(ξ C)/ID with the number of ASC in Figure 5. We fix the scale range n = 50 and plot the IC for each attribute in Figure 6 under a suitable ξ. From the figures, we find that when ξ = 0.5 two attributes (PL_Index and Pivot_Energy) contribute around 98% information to Data Set A. While for Data Set B, four attributes (PL_Index, Pivot_Energy, Flux_Band1, and Flux_Band7) contribute around 92% information when ξ = 0.5.

Figure 5.

Figure 5. The average ratio of the PID of ASC to ID (magenta color) and the average number of ASC (blue color) with different correlation threshold ξ for the training sets of Data Set A (left-hand panel) and Data Set B (right-hand panel).

Standard image High-resolution image
Figure 6.

Figure 6. The IC for each attribute when ξ = 0.5 for the training sets of Data Set A (left-hand panel) and Data Set B (right-hand panel).

Standard image High-resolution image

3. Proposed Method

3.1. Inverse Discrete Wavelet Transform

The attribute analysis methods introduced in Section 2 are used to find the important attributes or components that can be retained for the learning process, while other attributes or components are removed for dimension reduction. However, the crude removal may cause the loss of correlation features that degrade the performance.

We propose to retain all of the attributes but perform IDWT on the original samples for representations at higher resolutions. After IDWT, the correlation features are supposed to be highlighted, while the dimension is possibly reduced, which is important when dealing with big data.

The well-known discrete wavelet transform (DWT) is usually implemented with filtering operations for high efficiency. The filters are designed based on the standpoint of multiresolution: the difference of information between the approximation of a signal at the resolutions 2m+1 and 2m (where m is an integer) can be extracted by decomposing this signal on an orthonormal basis of wavelets (Mallat 1989). The pyramidal structure of the wavelet filter bank makes it possible to infer information at a low resolution from information at a high resolution. IDWT is the converse process of DWT and provides representations at high resolutions for DL.

However, in practical applications, a finite signal should be considered. The length of the signals varies at different resolutions, which is due to the operations of downsampling, upsampling, and filtering. In the IDWT process, if we denote p and s as the length of the signal at low resolution and high resolution, respectively, then Rajmic & Prusa (2014) give

Equation (1)

where u is the length of the filters.

To perform IDWT on the original data set, the attribute data of one sample will be seen as the signal at a lower resolution. As a result, the attributes should be rearranged in some order based on wavelet theory. This can be achieved with the alignment of the information in the wavelet domain and attribute space, which will be discussed next.

3.2. Information in Wavelet Domain: Ic

The process of DWT is to analyze the original signal from a fine scale to a coarse scale. The representation of a finite signal f(t) in the wavelet domain after DWT is a collection of vectors:

Equation (2)

where $J\in {\mathbb{N}}$ is called decomposition level; c J contains the approximation coefficients at level J, i.e., the lowest resolution; and d j contains the detail coefficients at level j, i.e., the high resolutions. An example of the three-level pyramid transform is illustrated in Figure 7. We find that the original signal f(t) can be seen as the approximation coefficients at the highest resolution, i.e., c 0.

Figure 7.

Figure 7. Wavelet coefficients for a three-level pyramid transform.

Standard image High-resolution image

However, based on prior knowledge, it is often observed that at a specific level, the majority of information in a natural signal is typically presented in the approximation coefficients. Additionally, the information is typically reduced by half during the decomposition of the signal from level j to j + 1 in DWT, due to the downsampling operation. Hence, it can be found that the information of coefficients at different levels roughly respect

Equation (3)

As an example, the three-level DWT is performed on three images to intuitively illustrate the relationship, as shown in Figure 8. These figures suggest that the approximation coefficients at each level retain the image contour and contain a significant amount of information, while the objects in the image cannot be easily identified through the detail coefficients. However, to some extent, the detail coefficients at a low level appear to provide more information than those at a high level, i.e., Ic ( d j+1) < Ic ( d j ).

Figure 8.

Figure 8. Representations of images of Lena, Goldhill, and Peppers in (a)–(c) spacial domain, and (d)–(f) wavelet domain, respectively. For display performance, the images of coefficients have been mapped into pink color. The detail coefficients at a particular level drawn in the figures are the sum of the corresponding horizontal, vertical, and diagonal detail coefficients.

Standard image High-resolution image

To transform the coefficients from lower resolutions to higher resolutions in the IDWT process, the wavelet decomposition vector (WDV) c and bookkeeping vector (BV) l are required according to multiresolution analysis. The WDV includes the coefficients shown in Equation (2):

Equation (4)

while the BV is made up of the number of coefficients in c :

Equation (5)

where len(·) indicates the number of the elements in a vector.

3.3. Information in Attribute Space: Ia

In the field of big data, an attribute's information reflects its contribution to the data set. As explained in Section 2.4, the IC estimates the potential contribution of an individual attribute to the data set. However, the presence of correlations among attributes means that the actual contribution of an attribute ai , i.e., its information Ia ({ai }), cannot be precisely determined if ai involves other attributes.

Moreover, as discussed, ASC is the smallest subset of attributes that can fully characterize the entire data set. Given a correlation threshold ξ, the ASC ξ C can be found with the FDASE algorithm (de Sousa et al. 2007). The attributes in ASC are not ξ-correlated with each other and they contain the most information in the data set.

Based on the analysis, if ${\mathbb{A}}=\{{a}_{1},{a}_{2},\ldots ,{a}_{E}\}$ is denoted as the universal attribute set of the data set that has E attributes, then the real contribution of attribute ${a}_{i}\in ({\mathbb{A}}-\xi C)$ to the data set can be seen as the degree of correlation between ai and the attributes in ξ C. The weaker the correlation between ai and ξ C, the higher the contribution of ai to the data set, i.e., attribute ai contains more information. Therefore, the information of attribute ai can be estimated by

Equation (6)

where pD(·) is the PID of a sub-data set.

By estimating the information for each attribute ${a}_{i}\in ({\mathbb{A}}-\xi C)$, we can obtain a rough ordering of attributes based on their information. In this case, the data under the attributes can be treated as wavelet coefficients and arranged to form the WDV c and BV l for IDWT. The data set is then transformed into representations at a higher resolution where the correlation features are presented. This is the main idea of our method, which will be detailed next.

3.4. FDIDWT

The J-level IDWT aims to transform the original data into representations at the J-higher resolution. To achieve this, the WDV and BV must be constructed with the attribute data of each sample. However, the positions of the data in the WDV are determined by the estimation of the attribute information with Equation (6). As a result, this approach is referred to as FDIDWT.

As discussed earlier, the attributes in ASC $\xi C\subset {\mathbb{A}}$ contain the most information of the data set defined in ${\mathbb{A}}=\{{a}_{1},{a}_{2},\ldots ,{a}_{E}\}$. It is assumed that the information of attributes in ${\mathbb{A}}-\xi C$ respect an ascending order:

Equation (7)

where ${a}_{i}\in ({\mathbb{A}}-\xi C)$ and 1 ≤ iQ. If the number of attributes in ξ C is P, then P + Q = E. Contrasting with the order of the information in the wavelet domain shown in Equation (3), we group the attributes in ${\mathbb{A}}-\xi C$ and obtain a similar order of the information in the attribute space:

Equation (8)

where DL j is the attribute group containing L attributes whose information respects to Equation (7), i.e.,

Equation (9)

Thus, the data under the attributes of DL j and ξ C can be taken as the detail coefficients d j and approximation coefficients c J , respectively. Specifically, if denote x j and x ξ C as the data under the attributes in DL j and ξ C, respectively, for one sample of the original data set, then the WDV c and BV l for the IDWT process can be constructed as

Equation (10)

where O is the number of attributes in the transformed data set, i.e., the transformed dimension in the new attribute space, which can be derived with Equation (1). Table 2 shows the procedure of the proposed FDIDWT method. To better describe how the algorithm works, we fix the decomposition level J = 3. Moreover, as in the common case, the length of filters u for the pyramid transform is set to be even. Then:

  • 1.  
    Steps 1–3: Given the correlation threshold ξ and scale range n, the FDASE algorithm (de Sousa et al. 2007) is used to find out the attribute set core ξ C. The information Ia ({ai }) of the remaining attributes ${a}_{i}\in ({\mathbb{A}}-\xi C)$ is then calculated. The attributes of ${\mathbb{A}}$ are arranged into a set ${\mathbb{A}}^{\prime} $ according to the ascending order of Ia ({ai }). 7 For instance, if the relationship in Equation (7) holds, then
    Equation (11)
    In our experiment, we scanned the correlation threshold ξ in the range [0.05, 1.0] with step 0.05, and set scale range n = 50. We got the ASC {a3, a1} and {a3, a1, a5} for Data Set A and B, respectively, as shown in Figures 5 and 6. Then, the arranged attribute set ${\mathbb{A}}^{\prime} $ for the two data sets will be {a3, a1, a5, a13, a10, a2, a9, a12, a4, a8, a11, a7, a6} and {a3, a1, a5, a13, a4, a12, a11, a7, a8, a6, a9, a2, a10}.
  • 2.  
    Steps 4–5: P is the number of attributes in ξ C. In the proposed method, the length of the approximation coefficients at level J is assumed to be not shorter than P. If this length equals P, then Figure 9 shows other groups divided from ${\mathbb{A}}^{\prime} $. ${C}_{3}^{P}$ denotes the approximation coefficient group of attributes in ξ C. Note that the original dimension E, i.e., the number of attributes of the original data set, may be smaller than the length of WDV c due to the given decomposition level J and filter length u. In this case, some necessary placeholder attributes "∗" are inserted following the ASC group ${C}_{3}^{P}$. The role of the placeholder attributes is to ensure that IDWT works while not changing the information of the data. Therefore, the values under the placeholder attributes are set to zeros so that they contribute nothing to the data set. In the signal processing field, signals are always transformed into the wavelet domain with DWT for processing, and the lengths of the wavelet coefficients at each level are recorded for the reconstruction with IDWT. However, there is no DWT in the proposed method. Hence, many cases exist for the lengths of the divided groups based on Equation (1):
    Equation (12)
    The length of the placeholder attributes L can be derived based on the equality of the length of WDV, i.e., P + M + N + K = E + L. Then, the WDV and BV are constituted as follows:
    Equation (13)
    where O is the transformed dimension computed with:
    Equation (14)
    We used the Daubechies filter, which can be indicated by its length u. For example, the filter length u = 8 indicates that the db4 wavelet filter was used. Under the assumption of the maximum decomposition level J = 3, we built the WDV and BV for two missions in Table 3.
  • 3.  
    Steps 6–7: Finally, IDWT is performed on each sample of the normalized data set, and a new data set defined on ${\mathbb{B}}=\{{b}_{1},{b}_{2},\,\ldots ,\,{b}_{i},\,\ldots ,\,{b}_{O}\}$ is generated. bi is the ith new attribute.

Figure 9.

Figure 9. The groups divided from sorted attributes, taking decomposition level J = 3 as an example.

Standard image High-resolution image

Table 2. The Proposed FDIDWT Method

Input: original data set defined on attribute space ${\mathbb{A}}$
 decomposition level J
 correlation threshold ξ
 scale range n
Output: new data set defined on attribute space ${\mathbb{B}}$
1:run FDASE algorithm to find out the ASC ξ C, denote its length as P;
2:calculate Ia (ai ), ${a}_{i}\in ({\mathbb{A}}-\xi C)$, with Equation (6);
3:arrange the attributes of ${\mathbb{A}}$ into ${\mathbb{A}}^{\prime} $ according to the order of Ia (ai ) as shown in Equation (7);
4:based on P, calculate the length of other groups, insert placeholder attributes if needed;
5:construct the WDV c and BV l ;
6:perform IDWT on each normalized sample with c and l ;
7:output the new data set defined on ${\mathbb{B}};$

Download table as:  ASCIITypeset image

Table 3. The WDV and BV for the IDWT in Two Missions

Decomposition LevelWDVBVWavelet NameCase
1A: ({x3, x1, x5, x13, x10, x2, x9}, {0, x12, x4, x8, x11, x7, x6})(7,7,13)db1#1
 B: ({x3,x1,x5,x13,x4,x12,x11}, {0,x7,x8,x6,x9,x2,x10})(7,7,12)db2#2
  (7,7,11)db2#3
  (7,7,10)db3#4
  (7,7,9)db3#5
  (7,7,8)db4#6
  (7,7,7)db4#7
  (7,7,6)db5#8
  (7,7,5)db5#9
  (7,7,4)db6#10
  (7,7,3)db6#11
  (7,7,2)db7#12
  (7,7,1)db7#13
3A: ({x3, x1, x5},{x13, x10, x2},{x9, x12, x4},{x8, x11, x7, x6})(3,3,3,4,5)db2#14
 B: ({x3,x1,x5},{x13,x4,x12},{x11,x7,x8},{x6,x9,x2,x10})(3,3,3,4,6)db2#15

Note. Value 0 in the WDV means one placeholder attribute is inserted.

Download table as:  ASCIITypeset image

3.5. A Lightweight CNN

Since the new attribute space ${\mathbb{B}}$ is considered to respect some natural order after FDIDWT, we believe that the convolution operation with a kernel will learn more features and achieve better performance. The kernel of CNN determines a visual field on the adjacent attributes in DL. Thus, we utilized a lightweight CNN model modified from the Matchbox net (Majumdar & Ginsburg 2020) but with a smaller size of only 94k parameters for high efficiency, which we refer to as MatchboxConv1D. Its structure is depicted in Figure 10.

Figure 10.

Figure 10. The proposed lightweight MatchboxConv1D model.

Standard image High-resolution image

4. Experiments and Results

With the three attribute analysis methods introduced in Section 2, we independently used four ML classifiers, i.e., RF, support vector machine (SVM), AdaBoost, and multilayer perceptron (MLP) to carry out Mission A and Mission B. The hyper-parameters are listed in Table 4. The FDIDWT method is designed to work with the MatchboxConv1D model proposed in Section 3.5 and the hyper-parameters are shown in Table 5. After training the models with the training set, fine-tuning with the validation set, and evaluating with the test set, the uncertain samples, i.e., the uncertain sources in Mission A and the BCUs in Mission B, are finally predicted with the model achieving the highest test accuracy. This flowchart is illustrated in Figure 11.

Figure 11.

Figure 11. The flowchart of attribute analysis and learning process.

Standard image High-resolution image

Table 4. The Hyper-parameters of Attribute Analysis Methods and Classifiers

Attribute Analysis MethodsHyper-parametersClassifiersHyper-parameters
RF Attribute Selection (AS)#DTs: 50,000 criterion: entropyRF#DTs: 700 criterion: entropy
PCA Component Reduction (CR)normalization: StandardScalerSVMnormalization: StandardScaler C = 5.0 degree = 1
FDASE Attribute Selection (FDASE)scale range: 50 correlation threshold: 0.5AdaBoostnormalization: MinMaxScaler #estimators: 300 learning rate: 0.5
  MLPhidden units: [32, 32, 64, 32, 32] dropout: 0.5 activation function: Leaky ReLU batch size: 64 epochs: 500 learning rate: 0.0005

Download table as:  ASCIITypeset image

Table 5. The Hyper-parameters of the Proposed Method and Model

Proposed MethodHyper-parametersProposed ModelHyper-parameters
FDIDWTwavelet type: Daubechies max decomposition level: 3MatchboxConv1Ddropout: 0.9 activation function: tanh batch size: 64 epochs: 500 learning rate: 0.0005 more can be found in Figure 10

Download table as:  ASCIITypeset image

We performed the experiments on each split data set and the accuracy results were averaged over 10 test sets. In the experiments, we extensively evaluated the performance of an increasing number of the most important attributes or principal components. These attributes and components were sorted in descending order of importance and variance ratio, respectively, as shown in Figures 3, 4, and 6. The ML methods were implemented with scikit-learn (Pedregosa et al. 2011) and the training process of DNNs was accelerated by TensorFlow (Abadi et al. 2016) on NVIDIA GeForce RTX 3060. The code that was used in this experiment has been uploaded to GitHub. 8

The average accuracy results of the commonly used attribute analysis methods and classifiers on test sets are compared in Tables 6 and 7 for two missions. These results are also illustrated in Figure 12. From the results, we find that the highest test accuracy is 95.49% ± 1.05% in mission A, which comes from the AdaBoost classifier working with the FDASE attribute selection method in the case of reducing two dimensions. In mission B, the highest test accuracy is 91.19% ± 0.0%. This results from the MLP classifier working with different attribute analysis methods but reduces at most one dimension.

Figure 12.

Figure 12. A comparison of average test accuracy results in two missions. The results of the proposed methods in each mission are depicted in every panel.

Standard image High-resolution image

Table 6. The Average Test Accuracy Results of Mission A with the ML Methods

Attribute Analysis Methods#DimensionRFSVMAdaBoostMLP
RF Attribute Selection (AS)186.08% ± 1.15%87.02% ± 0.0%87.95% ± 0.41%87.02% ± 0.0%
 285.15% ± 1.34%87.02% ± 0.0%87.77% ± 0.61%87.02% ± 0.0%
 387.65% ± 1.21%87.02% ± 0.0%88.22% ± 1.0%87.02% ± 0.0%
 490.39% ± 1.18%89.38% ± 0.49%88.91% ± 1.58%86.56% ± 0.0%
 590.73% ± 1.16%91.12% ± 0.59%90.77% ± 1.0%89.98% ± 0.0%
 6 90.82% ± 1.18%90.93% ± 0.64%90.87% ± 0.91%90.2% ± 0.0%
 790.55% ± 0.98%92.23% ± 0.88%93.35% ± 0.93%91.34% ± 0.0%
 890.0% ± 0.92%93.14% ± 0.81%94.4% ± 1.03%92.03% ± 0.0%
 990.73% ± 1.03%93.26% ± 0.78%94.69% ± 1.02%92.71% ± 0.0%
 1090.64% ± 0.97%93.3% ± 0.83%94.81% ± 1.14% 93.39% ± 0.0%
 1190.27% ± 0.79%93.35% ± 0.78%95.33% ± 0.88%91.57% ± 0.0%
 1289.32% ± 0.83% 93.42% ± 0.76%95.4% ± 0.77%91.57% ± 0.0%
 1388.38% ± 0.59%93.37% ± 1.04% 95.47% ± 0.86%90.43% ± 0.0%
PCA Component Reduction (CR)180.75% ± 2.17%87.15% ± 0.75%86.97% ± 0.14%87.93% ± 0.0%
 285.97% ± 1.5%87.49% ± 0.81%86.63% ± 0.82%87.47% ± 0.0%
 391.85% ± 1.05%90.84% ± 0.53%91.5% ± 1.16%88.61% ± 0.0%
 492.69% ± 0.75%91.41% ± 0.61%91.64% ± 1.0%89.07% ± 0.0%
 592.78% ± 0.77%91.75% ± 0.64%92.71% ± 1.08%90.43% ± 0.0%
 693.37% ± 0.61%92.71% ± 0.6%93.3% ± 0.88%91.34% ± 0.0%
 793.33% ± 0.95%92.78% ± 0.59%93.39% ± 0.9% 91.8% ± 0.0%
 894.03% ± 0.97%93.12% ± 0.67%93.9% ± 0.58%91.8% ± 0.0%
 993.96% ± 0.89%93.28% ± 0.78%94.03% ± 0.52%91.57% ± 0.0%
 1094.03% ± 1.05%93.35% ± 0.9% 94.17% ± 0.83%91.57% ± 0.0%
 1194.37% ± 0.89% 93.37% ± 1.04%94.12% ± 0.74%91.57% ± 0.0%
 1294.76% ± 0.91%93.37% ± 1.04%94.15% ± 1.04%91.57% ± 0.0%
 13 94.87% ± 0.93%93.37% ± 1.04%94.17% ± 0.85%91.34% ± 0.0%
FDASE Attribute Selection (FDASE)287.88% ± 1.36%88.88% ± 0.45%87.38% ± 0.37%87.7% ± 0.0%
 387.43% ± 0.45%89.16% ± 0.49%88.2% ± 0.76%88.84% ± 0.0%
 489.18% ± 0.85%89.86% ± 0.61%89.43% ± 0.95%88.84% ± 0.0%
 588.2% ± 0.34%90.14% ± 0.79%89.61% ± 0.8%88.84% ± 0.0%
 690.73% ± 1.05%91.28% ± 0.57%90.93% ± 1.03%89.75% ± 0.0%
 789.09% ± 0.61%92.6% ± 0.8%93.14% ± 1.12%90.66% ± 0.0%
 887.93% ± 0.59%92.57% ± 0.75%93.39% ± 1.27%91.12% ± 0.0%
 9 90.91% ± 0.87%92.32% ± 0.61%93.83% ± 0.89%91.34% ± 0.0%
 1090.48% ± 0.88%93.12% ± 0.69%95.4% ± 1.02% 91.8% ± 0.0%
 1189.95% ± 0.93%93.23% ± 0.79% 95.49% ± 1.05%91.34% ± 0.0%
 1289.64% ± 0.89% 93.42% ± 0.76%95.4% ± 0.77%91.8% ± 0.0%
 1388.34% ± 0.69%93.37% ± 1.04%95.47% ± 0.86%91.57% ± 0.0%

Note. The highest test accuracy for each combination of the attribute analysis method and the classifier is highlighted in bold face.

Download table as:  ASCIITypeset image

Table 7. The Average Test Accuracy Results of Mission B with the ML Methods

Attribute Analysis Methods#DimensionRFSVMAdaBoostMLP
RF Attribute Selection (AS)180.31% ± 3.08%87.05% ± 1.85%86.61% ± 1.69%86.34% ± 0.0%
 287.97% ± 1.97%88.9% ± 1.67%87.89% ± 1.97%88.55% ± 0.0%
 389.65% ± 2.12% 90.04% ± 0.89% 89.82% ± 1.45%89.87% ± 0.0%
 489.69% ± 1.46%89.69% ± 1.02%89.74% ± 1.86%89.43% ± 0.0%
 589.65% ± 1.79%89.34% ± 1.05%89.12% ± 2.08%89.43% ± 0.0%
 689.82% ± 1.93%89.38% ± 1.13%89.52% ± 1.83%89.87% ± 0.0%
 789.91% ± 1.55%89.52% ± 1.03%89.38% ± 2.19%89.87% ± 0.0%
 8 90.04% ± 1.62%89.74% ± 0.95%89.3% ± 2.04%89.87% ± 0.0%
 989.96% ± 1.54%89.74% ± 1.3%89.07% ± 2.32%89.87% ± 0.0%
 1089.91% ± 1.69%89.82% ± 1.35%89.16% ± 1.82%90.31% ± 0.0%
 1189.87% ± 1.38%90.0% ± 1.21%89.12% ± 2.29%90.75% ± 0.0%
 1289.52% ± 1.39%90.0% ± 1.14%89.34% ± 1.81% 91.19% ± 0.0%
 1389.69% ± 1.78%90.0% ± 1.3%89.43% ± 1.83%90.75% ± 0.0%
PCA Component Reduction (CR)161.15% ± 2.41%65.42% ± 1.85%89.03% ± 1.16%63.44% ± 0.0%
 289.07% ± 1.11%90.4% ± 1.54%88.81% ± 1.65%89.87% ± 0.0%
 388.94% ± 1.4%90.53% ± 1.44%89.3% ± 1.54%90.75% ± 0.0%
 488.9% ± 1.66% 90.57% ± 1.44%88.81% ± 2.08%90.75% ± 0.0%
 589.65% ± 1.64%90.13% ± 1.27%88.94% ± 1.72%90.75% ± 0.0%
 689.78% ± 1.47%89.96% ± 1.33%88.85% ± 1.64%90.75% ± 0.0%
 790.09% ± 1.43%90.09% ± 1.25%88.85% ± 2.08%90.75% ± 0.0%
 889.96% ± 1.58%90.09% ± 1.32% 89.52% ± 1.73%90.75% ± 0.0%
 989.56% ± 1.47%90.0% ± 1.18%89.3% ± 1.48%90.75% ± 0.0%
 1089.82% ± 1.46%89.87% ± 1.31%88.99% ± 2.13%90.75% ± 0.0%
 1189.87% ± 1.66%89.91% ± 1.38%89.08% ± 2.02%90.75% ± 0.0%
 1290.04% ± 1.6%90.0% ± 1.3%88.94% ± 1.67% 91.19% ± 0.0%
 13 90.26% ± 1.74%90.0% ± 1.3%88.59% ± 1.56%90.31% ± 0.0%
FDASE Attribute Selection (FDASE)288.06% ± 2.15%88.9% ± 1.67%87.89% ± 1.97%88.55% ± 0.0%
 388.63% ± 1.99%88.63% ± 1.97%87.71% ± 2.13%86.78% ± 0.0%
 488.28% ± 1.78%88.68% ± 2.0%87.97% ± 2.13%86.78% ± 0.0%
 589.69% ± 2.14%89.74% ± 0.95%89.12% ± 1.62%90.31% ± 0.0%
 689.82% ± 2.08%89.52% ± 0.74%88.9% ± 1.41%89.87% ± 0.0%
 790.0% ± 1.92%89.52% ± 0.9%89.3% ± 1.74%90.75% ± 0.0%
 890.09% ± 2.16%89.6% ± 0.86%89.16% ± 1.64%90.31% ± 0.0%
 989.96% ± 1.99%89.56% ± 0.91% 89.52% ± 1.8%90.31% ± 0.0%
 10 90.13% ± 2.09%89.6% ± 1.06%89.12% ± 1.9%90.75% ± 0.0%
 1190.13% ± 2.06%89.74% ± 1.18%89.03% ± 1.6%90.75% ± 0.0%
 1289.78% ± 1.61% 90.0% ± 1.14%89.34% ± 1.81%90.31% ± 0.0%
 1389.91% ± 1.37%90.0% ± 1.3%89.43% ± 1.83% 91.19% ± 0.0%

Note. The highest test accuracy for each combination of the attribute analysis method and the classifier is highlighted in bold face.

Download table as:  ASCIITypeset image

Table 8 shows the average test accuracy results of the proposed method with different transformed dimensions (see Table 3) for two missions. We find that case #6 achieves the highest test accuracy and outperforms the results of Tables 6 and 7, while reducing five dimensions. Finally, the prediction results of uncertain sources and BCUs in two missions are listed in Table 9.

Table 8. The Average Test Accuracy Results of Two Missions with the FDIDWT Method and MatchboxConv1D Model

#DimensionCaseTest Accuracy
  Mission AMission B
13#194.95% ± 2.47%91.1% ± 1.61%
12#293.53% ± 1.98%90.97% ± 2.35%
11#396.36% ± 1.95%91.94% ± 2.21%
10#495.24% ± 1.44%90.16% ± 1.54%
9#596.01% ± 1.96%87.89% ± 2.53%
8#6 96.65% ± 1.32% 92.03% ± 2.2%
7#792.61% ± 1.81%91.8% ± 2.28%
6#893.74% ± 0.99%88.39% ± 3.94%
5#994.29% ± 1.4%84.57% ± 2.53%
4#1093.0% ± 1.25%84.1% ± 2.78%
3#1192.11% ± 2.07%81.81% ± 3.38%
2#1290.67% ± 1.47%82.8% ± 2.51%
1#1388.97% ± 0.01%80.59% ± 2.31%
5#1492.22% ± 2.51%91.62% ± 2.17%
6#1592.31% ± 2.57%91.50% ± 3.76%

Note. The highest test accuracy for each combination of the attribute analysis method and the classifier is highlighted in bold face.

Download table as:  ASCIITypeset image

Table 9. The Prediction Results of Uncertain Sources in Mission A and BCUs in Mission B

MissionSource NamePredict ProbabilityPredict Class
A4FGL J0000.3-73550.9963 ± 0.0028AGN
 4FGL J0000.5+07430.9948 ± 0.0048AGN
 4FGL J0000.7+25300.9822 ± 0.0229AGN
 4FGL J0001.6+35030.9805 ± 0.0153AGN
 4FGL J0002.1+6721c0.9698 ± 0.0111non-AGN
B4FGL J0001.2+47410.9315 ± 0.0404BL Lac (BCU)
 4FGL J0001.6-41560.9887 ± 0.0159BL Lac
 4FGL J0001.8-21530.8635 ± 0.0966BL Lac (BCU)
 4FGL J0002.3-08150.9881 ± 0.0047BL Lac
 4FGL J0002.4-51560.8595 ± 0.1072BL Lac (BCU)

Only a portion of this table is shown here to demonstrate its form and content. A machine-readable version of the full table is available.

Download table as:  DataTypeset image

5. Discussion

5.1. The Proposed Method

Extracting correlation features among the attributes of Fermi sources is a promising research field. The commonly used attribute analysis methods, such as the estimation of attribute importance (see Section 2.2) or principal components (see Section 2.3), aim to find the significant attributes or components, and remove the unimportant ones for dimension reduction. However, the correlation features will be removed with these methods, which results in inferior classification performance (see the results of Tables 6 and 7).

The FDASE gives a global perspective of the whole data set and estimates the correlation features among all of the attributes based on FD theory. The ASC is the resulting attribute subset of the FDASE algorithm and its data contains most of the information of the data set under some correlation threshold and scale range. Nevertheless, the retention of ASC also suffers from the loss of correlation features (see the results of FDASE attribute selection in Tables 6 and 7). Thus, we propose to rearrange the original attributes to highlight correlation features at higher resolutions based on FDASE and IDWT, which is referred to as the proposed FDIDWT method.

The resulting attribute space is also considered to respect some natural order after FDIDWT. We believe that the convolution operation will extract more correlation features from the ordered attributes. Meanwhile, the structure of the CNN gives more potentiality for better classification performance. Therefore, by combining FDIDWT with the MatchboxConv1D model, we obtain the best test accuracy for both Mission A and B in case #6. In addition, the dimension is transferred to be 8, which is significant for reducing the computing burden in big astronomical data processing. It may also be concluded that most features of Data Set A and B exist at one higher resolution because case #6 corresponds to decomposition level 1 (see Table 3).

The prediction results have been carried out with an accuracy of 96.65% ± 1.32% and 92.03% ± 2.2% for Mission A and Mission B, respectively, using the proposed method. The results are listed in Table 9. A general comparison between the predicted AGNs of Mission A and the originally confirmed AGNs is shown in Figure 13. A comparison between the predicted BL Lacs of Mission B and the originally confirmed BL Lacs is shown in Figure 14. Finally, a comparison between the predicted FSRQs of Mission B and the originally confirmed FSRQs is shown in Figure 15. From the comparison, we can find that the distribution shapes of the 13 attributes of the predicted sources resemble the shapes of the original 4FGL_DR3 sources with the corresponding classification, in general. The results indicate that our predicted sources are correctly classified.

Figure 13.

Figure 13. A comparison of 13 attributes between 4FGL_DR3 AGNs and predicted AGNs. The histogram of 4FGL_DR3 AGNs is shown in red and the histogram of predicted AGNs is shown in blue; the area under the histogram integrates into one.

Standard image High-resolution image
Figure 14.

Figure 14. A comparison of 13 attributes between 4FGL_DR3 BL Lacs and predicted BL Lacs. The histogram of 4FGL_DR3 BL Lacs is shown in red and the histogram of predicted BL Lacs is shown in blue; the area under the histogram integrates into one.

Standard image High-resolution image
Figure 15.

Figure 15. A comparison of 13 attributes between 4FGL_DR3 FSRQs and predicted FSRQs. The histogram of 4FGL_DR3 FSRQs is shown in red and the histogram of predicted FSRQs is shown in blue; the area under the histogram integrates to one.

Standard image High-resolution image

We notice that the histograms of the attribute Variability_Index show a longer high-variation-index tail for the original 4FGL_DR3 sources than for the predicted sources (including predicted AGNs, predicted BL Lacs, and predicted FSRQs in Figures 1315, respectively), and the predicted sources have more contribution on the high-variation-index head than the original 4FGL_DR3 sources. This result suggests that our method has the advantage of finding less variable sources from the uncertain sources. We also notice that the original 4FGL_DR3 sources give more contribution to the histogram tails than the predicted sources, while the predicted sources give more contribution to the histogram head for the attributes of multi-band intensities (Flux1000, Flux_Band1, Flux_Band2, Flux_Band3, Flux_Band4, Flux_Band5, Flux_Band6, Flux_Band7, and Flux_Band8). This suggests that our method has the advantage of finding relatively faint γ-ray sources from the uncertain sources. It also encourages us to believe that our method should be able to make efforts to identify less variable and faint sources in the era of survey telescopes, e.g., the Large Synoptic Survey Telescope (LSST; LSST Science Collaboration et al. 2009), China Space Station Telescope (CSST; Zhan 2011), etc.

5.2. The Classification Results of Mission A and Mission B

By employing the proposed method, we managed to predict 2291 uncertain sources into 1731 AGNs and 560 non-AGNs, and predict 1493 BCUs into 948 BL Lacs and 545 FSRQs. The likelihood probabilities of these predicted uncertain sources and BCUs are shown in Table 9 and the distribution is displayed in Figure 16.

Figure 16.

Figure 16. The likelihood probability distribution of Mission A (upper panel) and Mission B (lower panel). The AGN and non-AGN candidates are shown in the upper histogram with the solid line and the dashed line, respectively. The BL Lac and FSRQ candidates are shown in the lower histogram with the solid line and the dashed line, respectively. The dashed red lines are the set boundary of likelihood probability equals 95%.

Standard image High-resolution image

We set a boundary of likelihood probability greater than 95%, which is shown as the dashed-red lines in Figure 16, to claim this source as a candidate of the corresponding class. In this case, we get further constrained results into 1354 AGN candidates in Mission A, 482 BL Lacs candidates and 128 FSRQ candidates in Mission B, as shown in the last column of Table 9.

Several methods have been employed to classify 4FGL BCUs as FSRQ candidates or BL Lac candidates, as Mission B of the present work, in the previous works. Kang et al. (2019) utilized three supervised ML methods (RF, SVM, and ANN) to make a classification for 1312 4FGL_DR1 BCUs, and carried out a combined classification result of 724 BL Lac candidates and 332 FSRQ candidates. By crossing matching the results from Kang et al. (2019) and our results, we found that there are 419 overlapping BCUs in these two works. Among the overlapping 419 BCUs, 324 BCUs are predicted as BL Lac candidates and 90 BCUs are predicted as FSRQ candidates in both works, which gives our results a consistency of 98.8% with Kang et al.'s work.

Similarly, five different supervised ML algorithms (RF, LR, XGBoost, CatBoost, and neural network) were applied to the 4LAC_DR3 BCUs in Agarwal (2023), and 610 BCUs were able to be classified as BL Lac candidates and 333 BCUs were classified as FSRQ candidates. A comparison between our work and Agarwal's work suggests that there are 481 overlapping BCUs in the two works, 392 BCUs are classified as BL Lac candidates and 87 BCUs are classified as FSRQ candidates in both works, which gives our results a consistency of 99.6% with their work.

Fan et al. (2022) employed three physical parameters and built diagrams among them (i.e., the photon spectrum index against the photon flux diagram, the photon spectrum index against the variability index diagram, and the variability index against the photon flux) to separate the known BL Lacs from known FSRQs. They then used the boundary to divide BCUs into BL Lacs and FSRQs. In their work, 751 BCUs were classified as BL Lac candidates and 210 BCUs were classified as FSRQ candidates. There are 492 overlapping BCUs in the two works, 409 BCUs are classified as BL Lac candidates and 83 BCUs are classified as FSRQ candidates in both works, which gives our results a consistency of 100% with their work.

6. Conclusion

In this paper, the correlation features of attribute space of the 4FGL_DR3 data set are highlighted by the proposed FDIDWT method and the intrinsic features hidden in data are further extracted by a lightweight MatchboxConv1D model. With the combination of the FDIDWT method and the MatchboxConv1D model, we have obtained results with an accuracy of 96.65% ± 1.32% for Mission A and an accuracy of 92.03% ± 2.2% for Mission B. For the likelihood probability boundary of 95%, we managed to classify 1354 AGN candidates in Mission A, and 482 BL Lacs candidates and 128 FSRQ candidates in Mission B. A high consistency of greater than 98% emerges by comparing our predicted candidates with those from previous works. More importantly, our method has the advantage of finding less variable and faint sources.

Acknowledgments

H.B.X acknowledges the support from the National Natural Science Foundation of China (NSFC) under grant No.12203034, from the Shanghai Science and Technology Fund under grant No. 22YF1431500, and from the science research grants from the China Manned Space Project. Z.J.L acknowledges the support from NSFC grant 12141302, the Shanghai Science and Technology Fund under grant No. 20070502400, and from the science research grants from the China Manned Space Project. J.H.F acknowledges the support of the NSFC U2031201, NSFC 11733001, the Scientific and Technological Cooperation Projects (2020–2023) between the People's Republic of China and the Republic of Bulgaria, the science research grants from the China Manned Space Project with No. CMS-CSST-2021-A06, and the support for Astrophysics Key Subjects of Guangdong Province and Guangzhou City.

Footnotes

Please wait… references are loading.
10.3847/1538-4357/ad0e6c