Next Article in Journal
Designing a Fruit Identification Algorithm in Orchard Conditions to Develop Robots Using Video Processing and Majority Voting Based on Hybrid Artificial Neural Network
Previous Article in Journal
Influence of Pairing Startling Acoustic Stimuli with Postural Responses Induced by Light Touch Displacement
Previous Article in Special Issue
UAV-Based GNSS-R for Water Detection as a Support to Flood Monitoring Operations: A Feasibility Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semi-Supervised GNSS Scintillations Detection Based on DeepInfomax

1
Department of Electronics and Telecommunications, Politecnico di Torino, 10129 Torino, Italy
2
Department of Navigation and Positioning, Finnish Geospatial Research Institute (FGI-NLS), 02430 Masala, Finland
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(1), 381; https://doi.org/10.3390/app10010381
Submission received: 30 November 2019 / Revised: 25 December 2019 / Accepted: 30 December 2019 / Published: 4 January 2020
(This article belongs to the Special Issue Signal Processing for Satellite Positioning Systems)

Abstract

:
This work focuses on a machine learning based detection of ionospheric scintillation events affecting Global Navigation Satellite System (GNSS) signals. We here extend the recent detection results based on Decision Trees, designing a semi-supervised detection system based on the DeepInfomax approach recently proposed. The paper shows that it is possible to achieve good classification accuracy while reducing the amount of time that human experts must spend manually labelling the datasets for the training of supervised algorithms. The proposed method is scalable and reduces the required percentage of annotated samples to achieve a given performance, making it a viable candidate for a realistic deployment of scintillation detection in software defined GNSS receivers.

1. Introduction

Ionospheric scintillations are defined as phase and amplitude variations of trans-ionospheric radio signals, caused by irregularities and rapid fluctuations of the electron content in the upper layers of atmosphere. In this work, we focus on the effect of scintillation on Global Navigation Satellite Systems (GNSS) signals. Whether the main purpose of the application under consideration is navigation, in which the user demands high quality measurements and scintillation mitigation or rejection, or a scientific application, in which the user is interested in selecting scintillation events within huge datasets, a clean, fast, and precise detection of scintillation events is of paramount importance. The scientific literature on the study of scintillations and their effect on GNSS systems is wide and complete, and therefore is outside the aim of this paper. We refer the reader to several sources to study in depth the scintillation topic [1,2,3,4], the receiver design, and signal processing techniques [5,6,7].
The idea of using machine learning techniques for scintillation detection is not new. The need for machine learning solutions is due to the fact that manual classification of scintillation events based on visual inspection of scintillation indices time series is nowadays unpractical, given the amount of data. Similarly, simple heuristic algorithms, while requiring barely any tuning time from a human expert, have limited performance. Examples are algorithms that decide whether scintillation is occurring or not based on a simple threshold crossings of quantities such as the amplitude scintillation index ( S 4 ), the carrier to noise power spectral density ratio ( C / N 0 ) or the satellite elevation angle. It has been proven that machine learning algorithms can successfully replace and overcome traditional detection techniques.
Such approach was originally proposed in [8,9], where the authors use an automatic scintillation detector based on a supervised machine learning technique called support vector machine (SVM), respectively for amplitude and phase scintillation. This approach provides good detection results but introduces a Fast Fourier Tranform (FFT) operation on the data over windows of 3 min, thus reducing the temporal resolution of the detection results. A second disadvantage is that the input data need to be pre-filtered at an elevation mask of 30 to reduce multipath false alarms, at the expenses of a potential waste of useful and important information. An updated version of the algorithm is presented in [10]. Other works propose using decision tree and random forest algorithms and low level signal observables, such as the in-phase and quadrature correlation outputs of the receiver tracking loop [11,12]. This approach is able to reduce the false alarms rate due to the ambiguity between scintillation and multipath typical of the classical approaches based on the analysis of amplitude scintillation index and provide an early run-time alert. Another advantage is that, by exploiting the high rate correlator outputs, a finer time resolution in detection is obtained [13,14]. A more generic survey of data mining techniques is given in [15]. However, this method does not only rely on GNSS observations, but also includes external data, such as sensors and online forecast services.
The biggest problem with machine learning based classification is that datasets must be labeled manually by experts, wasting precious resources in terms of time, and thus preventing large scale scalability of the method [8,11]. The techniques proposed in the literature and described above partially solve the problem as they are based on fully-supervised learning: a significant portion of input data must be manually labeled to assist the learning process of the algorithm. We here propose to switch to a paradigm centered around semi-supervised learning, in which in contrast to fully supervised learning the label information is not required for all the data, and some entries can be thus classified with an “unknown” label.
In recent years, deep learning based solutions [16] for classification tasks have seen tremendous improvements in terms of performance. Neural networks in particular, however, have been rarely used as detection algorithms, one exception being [17]. Nevertheless, they have been successfully deployed for complex modeling of electron content in the ionosphere [18,19].
Usually, deep learning based semi-supervised classifiers (see, for example, [20]), reduce the dimensionality of the unlabeled data X using unsupervised learning, then build a classifier using only the compressed labeled datapoints F for the datapoints where the corresponding label Y is available, as depicted in the scheme of Figure 1.
In this paper, we make use of the DeepInfomax [21] approach in which a variational representation of mutual information [22] is used to build an unsupervised, information theoretic, feature extractor. Differently from other unsupervised techniques such as Variational Autoencoders [20] that require both an encoder and a decoder, the considered method only needs and encoder, almost halving the complexity. Variational techniques powered by neural networks for mutual information estimation, and in general divergences between probabilities, have been widely used in the literature [22,23,24,25,26]. The complete pipeline we propose consists of firstly compressing the input time series into smaller cardinality vectors, and then feed these vectors to a binary classifier for scintillation detection. During training, the unsupervised compression is performed for all datapoints, whereas, obviously, the training of the classifier is performed only for the labeled datapoints.
The aim of this paper is then to develop a new scintillation detection strategy based on Convolutional Neural Networks and semi-supervised learning. In more detail, we extend the results and approaches presented in [11], in which classification of scintillation events is performed using supervised decision trees. While showing good performance, this method has two fundamental problems. First, it lacks the capability of considering as a whole a single scintillation event sampled at different time instants, as the time feature is not taken into account. Second, it requires to have a fully labeled dataset, where the process of annotation had to be done manually by a team of experts in the field. We instead propose a solution that mitigates both problems by using a system that naturally:
  • handles subsequent samples in time, by applying an overlapping time window to all features, thus including the information about the temporal correlation;
  • requires only a smaller portion of the dataset to be labeled.
We show that to reach the same level of accuracy, just about 20% of datapoints must be manually labeled. For a faithful comparison, we used the same dataset as in [11] and we performed statistical measurements of testing classification performance using the same splitting of datasets for training and testing. Moreover, when the amount of labeled datapoints is augmented, the proposed method reaches extremely high performance with a top accuracy on the test set of 0.9993.
The overview of the paper is then the following: in Section 2, we discuss the need for unsupervised compression; in Section 3, we give a brief overview of the DeepInfomax approach that we further expand in Appendix A; in Section 4, we describe the considered datasets and their properties; in Section 5, we give an overview of the system architecture and training and testing procedure; and, in Section 6, we show the results of the proposed classification method. Finally, in Section 8, we discuss on the results and suggest possible future work directions. We also include a complete Appendix B to briefly describe the neural network architectures and training technicalities.

2. The Need for Unsupervised Compression

As depicted in Figure 1, one possible semi-supervised classification scheme is composed of two main blocks: an unsupervised feature extractor and a supervised classifier that is trained using as input the features. Notice that for some features the label correspondence is missing. The attentive reader might wonder why the first portion of the deep learning based scheme, the unsupervised encoding part, depicted in Figure 1, is necessary at all and a simple training of a classifier on the labeled portion of the dataset is not sufficient. The reason is that, as explained in the introduction (Section 1), in this work, we improved on two weak points of the previous decision tree based classification algorithms: being able to consider a collection in time of the features and to reduce the number of labeled training points. Generally speaking, in machine learning, these two requirements are in conflict: when increasing the cardinality of the inputs, as a rule of thumb, to maintain a good level of performance the number of training points must be increased. This is typically solved when dealing with high-dimensional inputs either by performing some form of manual feature engineering and reducing the dimensionality of the input or by increasing the amount of labeled data by acquiring larger labeled datasets. The approach of semi-supervised deep learning is instead different. Given a large dataset in terms of number of samples, with a large input cardinality, with a small number of labeled datapoints, an automatic, unsupervised, end-to-end feature extraction (we use interchangeably the term feature extraction or compression.) is performed on the whole dataset. A classifier is then trained using as inputs the small feature vectors thus resolving the problem of having few labeled datapoints, having the input be a small cardinality, a small number of labeled samples is sufficient to maintain the required level of performance. It will be shown in Section 5 in this work we consider a compression ratio of 32:1. In this work, we decided to use an information theoretic based feature extraction scheme [21].

3. The DeepInfomax Approach

In this section, we give a brief overview of the DeepInfomax approach [21]. Appendix A provides mode details about the technique. Figure 2 schematically depicts the idea.
The main idea is to build a two stage compression, using a cascade of two neural networks e η 1 ( · ) and e β 2 ( · ) , of an input X, respectively generating a first equivalent feature H = e η 1 ( X ) and a second one F = e β 2 ( H ) , such that the Mutual Information (MI) between H , F is maximized. In practice, as also explained in [21], a proxy to the mutual information that worked better in practice, the Jensen Shannon Divergence, was used. The authors in [21] introduce, and also maximize, a new metric: the Local Mutual Information. It is defined as the sum of all mutual information between each element H ( z ) , z = 1 , | H | of the variable H, that is, a multidimensional random vector and the whole feature vector F. In equations, M I l o c = 1 | H | z = 1 | H | I H ( z ) ; F , where I ( A , B ) is the mutual information (MI) between A and B.
The sketch of the architecture we consider in our work is depicted in Figure 3. We found that, for this work, considering only the local Mutual Information was the best choice in terms of performance. Moreover, a Uniform Prior regularizer, similarly to the one described in [21], was included to ensure that the features F cover evenly the latent space.
It is possible to show ([21,22], and also Appendix A) that a Local MI estimator can be built as depicted in Figure 4.
Couples extracted from the joint and marginal distributions of H , F are built (respectively p ( H , F ) , p ( H ) p ( F ) ). These couples are fed to two copies of the same network g θ l o c ( · ) whose output is the input of a cost function described by Equation (A7). Notice that, to build couples sampled from the joint distribution p ( H , F ) , it is sufficient to observe joint realizations of variables H , F , while, to build samples whose distribution is the product of the marginals p ( H ) p ( F ) , we can instead randomly shuffle one of the two sets of variables H or F and build new couples with the scrambled versions.
To perform regularization instead, always Figure 4, a discriminator network is fed with both the random variables uniformly generated and the features F, and trained to distinguish among the two. The rest of the system is trained adversarially [27] to fool the discriminator, and it is possible to show that, in an equilibrium regime, the distribution of F is the multivariate uniform distribution.
The whole process serves the purpose of extracting high quality and informative feature vectors F in an unsupervised way (i.e., no external labels are needed), which can be used for other downstream tasks such as classification or clustering.

4. The Data Collections

In this work, we use the same datasets as in [11]. The dataset is composed by a multivariate 10-dimensional time series composed of the following features: I and Q correlator outputs, sometimes referred as GNSS raw data; Signal Intensity ( S I ), computed as S I = I 2 + Q 2 ; the raw phase ζ of the received signal; the amplitude scintillation index S 4 ; the C / N 0 ; the satellite elevation and azimuth angles; the de-trended measure of carrier raw phase ϕ ; and the phase scintillation index σ ϕ . I, Q, the raw phase and consequently S I are provided at 50 Hz; the other observables are provided at 1 Hz. The overall dataset is then built by performing a moving average with integration time of 60 s and downsampling the high-rate features at 1 Hz. Before processing the data with neural networks a standard normalization, mean removal and division by standard deviation feature by feature has been performed.
Data have been collected in Hanoi, Vietnam (11 20’ N geo-magnetic latitude on 26 March and 2 April 2015. The measurements have been recorded by means of a Software Defined Radio (SDR) receiver and data grabber, as detailed in [28,29]. L1 C/A signals of 20 Global Positioning System (GPS) satellites were considered, corresponding to 169955 entries, recorded over a time interval of approximately 6 h, with a scintillation rate of 1:4. The ground truth label has been determined manually, based on the visual inspection of the observables.
To avoid confusion, it is important to underline that all the datasets used in this work have been manually annotated, as a ground truth is needed to measure classification accuracy performance. In our experiments, we artificially remove some of the labels to simulate a semi-supervised learning scenario.

5. Overview of Functioning and System Description

As mentioned earlier, one of our goals is to be able to harvest the information contained in the temporal correlation among features. For this reason, we consider as input of the system sliding windows of a multivariate serie X n built as described in Section 4. The input multivariate time series has thus cardinality equal to 10, or, in neural network language, the number of input channels is 10.
If we denote the sliding window size as N w i n , we can formally define the input as the [ 10 × N w i n ] matrix :
X n = X n N w i n + 1 1 X n N w i n + 2 1 X n 1 1 X n 1 X n N w i n + 1 10 X n N w i n + 2 10 X n 1 10 X n 10 T ,
where n is the last time instant of the window and X j is the j t h feature We decided to use a causal time window such that the method could be applied to a real-time detection scenario. We allow for a delay, such as in a post processing application, and define the window as
X n = X n N w i n 1 2 1 X n 1 X n N w i n + 1 2 1 X n N w i n 1 2 10 X n 10 X n N w i n + 1 2 10 T .
The performance will improve. Notice that in terms of coding the change is trivial with respect to what we implemented, a simple offset in the indexing is sufficient to switch between the two modalities. Notice also that for this second implementation N w i n must be odd. In this work, we consider the first variant and select N w i n = 64 . To test the performance of the proposed method, the entire dataset is split into a training part (90%) and a testing part (10%), using the same procedure as in [11].
During training, the sliding windows are processed and compressed according to a variation of DeepInfomax scheme [21] (highlighted in Section 3 and expanded in Appendix A) that produces a feature vector. As previously introduced, this first stage of training is completely unsupervised and, for our application, the purpose of this process is to create the lower cardinality equivalent feature vector that will be used as input of the classifier. To make the picture clearer, the cardinality of the raw input is the size to the 10-dimensional sliding windows, i.e., 10 × N w i n = 640 , while, in this work, we found that a well performing feature vector cardinality ( | F | ) is 20. Thus, for all the time instants for which the scintillation label is available, a classifier is trained with the corresponding feature vector. Notice that the training of compression scheme and classifier is jointly performed end to end using backpropagation.
Figure 5 depicts the global architecture of the system used during training. In the spirit of focusing on the important high level description of the method, all the details related to actual networks architectures and optimizer details are relegated to Appendix B. As anticipated, the first part of the scheme is related to the unsupervised feature extraction, while the second part is devoted to classification of labeled datapoints. The whole procedure can be summarized as follows, considering a batch based training:
  • A batch of input time windows { X r } is is sampled from the training set.
  • The inputs are processed by a first convolutional neural network e η 1 ( · ) and a first level compression is performed, generating variables { H r } .
  • Through e β 2 ( · ) , a second neural network with both convolutional and fully connected layers, the final features { F r } are built.
  • Starting from this point of the computation chain, three different parallel processes are performed:
    (i)
    classification of features { F r } to determine whether a given time instant r corresponds to scintillation. To be completely precise, as explained in Appendix B, the actual input to the classifier is the concatenation of feature vector F and the result of skip connections taken from the layers of the first convolutional network e η 1 ( · ) . This is done only for the time instants for which there is a reference. For all time instants for which a reference signal exists, the corresponding feature F r is fed to fully connected classifier c ψ ( · ) that receives as reference signal Y r , the corresponding label scintillation/no scintillation for time instant r.
    (ii)
    local Mutual Information (MI) computation: starting from the collections of { H r , F r } couples extracted from the joint and marginal distributions are built. These couples are fed to the network g θ l o c ( · ) whose output is the input of the C M I , l o c a l cost function, defined in (A7).
    (iii)
    adversarial matching of features F distribution to a target uniform distribution. Random variables F ˜ U ( 1 , 1 ) are generated. A discriminator is fed with both the random variables generated and the F (the rest of the system is trained adversarially [27] to fool the discriminator). This is used as a regularized to ensure that the features F cover evenly the latent space.
  • Standard backpropagation is performed, the cost function considered is C = C M I , l o c a l + C a d v e r s a r i a l + C c l a s s i f i c a t i o n , where the subscripts represent the various component of the cost function.
During testing, the only portion of the network that is used is the one related to classification, for simplicity depicted in Figure 6 that is clearly much simpler and lighter. In a realistic scenario, the training could be then performed offline on powerful machines while the lighter scintillation detection system could be subsequently uploaded on SDR receivers where the computational capabilities and the maximum power consumption are stricter.

6. Experiments

In this section, we present experimental results of the DeepInfomax method, compared with the Decision Tree proposed in [11], varying the number of labeled datapoints, by comparing the classification accuracy with respect to the baseline.The dataset used for the experiments is the same as in [11], but we neither train or test the method on the first 63 s of each data collection, due to the necessity of having a full window of samples. For each of the experiments, the labeled datapoints are randomly extracted from the training set. In the following, we present results in terms of standard machine learning performance indicators that we hereafter define. Denote as C the true class to which a datapoint belongs (0 for non scintillation, 1 for scintillation) and as C ^ the estimated class, i.e., the output of the classifier. We can define four joint probabilities:
  • true positive: T p = P C ^ = 1 , C = 1 ,
  • false positive: F p = P C ^ = 1 , C = 0 ,
  • true negative: T n = P C ^ = 0 , C = 0 ,
  • false negative: F n = P C ^ = 0 , C = 1 .
Derived from these four quantities are the following ones:
  • accuracy: A = T p + T n ,
  • precision: P = T p T p + F p ,
  • recall: R = T p T p + F n ,
  • F-score: F s c = 2 R P R + P .
Figure 7 reports the main result of this paper: a comparison between the proposed method and the decision tree based one, varying the number of labeled datapoints. For sake of completeness, it is furthermore necessary to clarify that, when dealing with the decision tree classification algorithm, the number of labeled datapoints (x-axis of Figure 7) coincides with the size of the dataset used for training, since there is no possibility to include unlabeled datapoints in the learning pipeline. Instead, with our approach, the size of the training dataset is always the same (90% of the complete dataset, roughly 150,000 datapoints) but only a portion of the datapoints is labeled.
The improvement in terms of performance is evident. First, it is worth noticing that the proposed DeepInfomax algorithm outperforms the Decision Tree technique, both in terms of accuracy and of F-score, for the same number of labeled elements. Second, a much lower number of labeled datapoints is required when using DeepInfomax to attain similar levels of accuracy and F-score. As the number of labeled points increases, the accuracy reaches a top value on the test set of 0.9993 for DeepInfomax, compared to the value 0.9824 for the Decision Tree.
It is important to notice that, even if from a performance point of view, the results are excellent, to attest to the usefulness of the method in a realistic application tests on many different datasets should be performed. It is possible, in fact, that bias is present in the considered data collection on which the networks base their decisions. However, this problem is complex and outside the scope of this paper; see [30].
Table 1 reports the results of the method for all the performance metrics introduced above. As expected, the performance increases with the increase of number of labeled datapoints, rapidly saturating to excellent levels. It is interesting to notice that the amount of false positives and false negatives is almost equivalent.

7. A Qualitative View on Features Importance

Neural networks are known for being black box models, in which obtaining human understandable information from the inner functioning is extremely complex, if not impossible. Several techniques have been proposed to determine and inspect feature importance on networks decisions, usually with specific methods tailored to particular experiments and particular network architectures.
In this subsection, we present results on features importance based on the simple but powerful features random permutation technique (similarly to the method introduced in [31]). The idea is to consider a trained classifier, randomly shuffle the variables one by one, and record the decrease in terms of performance. The measured decrease for all the shuffling experiments with respect to the baseline (the unshuffled dataset) is used as a proxy to determine the shuffled feature importance. Notice that this investigation strategy somehow lacks the capability of measuring joint higher order importance of features. In fact, two or more features can be extremely correlated and important for the considered task, but shuffling just one of them possibly does not impact performance in a substantial way, producing thus an underestimation of the feature importance.
In Figure 8, we report the results of permutation tests on the classifier trained with percentage of labeled examples equal to 10%. Despite the possible limitations of the considered measurement technique, we obtain results that are consistent with the physical knowledge of the problem as well as previous results [11].
The features in order of importance (computed as the decrease in terms of accuracy with respect to the baseline) are: S I , θ e l , S 4 , C / N 0 , ζ , θ a z , I , Q , σ ϕ , ϕ . In the considered data collections, amplitude scintillations’ events are observed. It is thus reasonable to assess that Signal Intensity, S I , is the most important feature, with a drop in accuracy of more than 40% when the feature is randomly permuted. The three features θ e l , S 4 , C / N 0 when permuted induces basically the same drop in terms of accuracy. Elevation angle θ e l is informative for classification: signal statistics for low elevation satellites greatly vary from the high in sky ones. The amplitude scintillation index S 4 is clearly an important feature since amplitude scintillations events are considered. Finally, the C / N 0 is yet another indirect measurement of signal intensity. Notice, moreover, that θ e l , S 4 , C / N 0 are the three features used for the so-called semihard detection rule [11]. The raw phase and the azimuth angle appear to contain less information about the scintillation events and the features I , Q , σ ϕ , ϕ have the lowest impact when randomized. It is possible that the relevant information contained in I , Q is already summarized in S I , while being the scintillation events linked to amplitude fluctuations the scintillation phase information is not that important.
A complete analysis of feature importance would require to retrain from scratch different versions of the classifier by completely excluding some of the features from the beginning of training, but, since this is computationally expensive and outside the scope of this paper, we leave this analysis for future works.

8. Conclusions

In this paper, we propose a new semi-supervised architecture for GNSS scintillations detection based on the DeepInfomax approach. We showed that it is possible to reduce by a large amount the number of labeled datapoints to achieve a target performance, decreasing the amount of time human researcher must spend manually labelling datasets. Thanks to the considered architecture, it is possible to process temporal windows of features empowering the capabilities of the model. Furthermore, the overall accuracy and F-score are improved when compared to other machine learning techniques. Future directions include testing of the trained system on datasets as different as possible, in order to verify the generalization capabilities of the proposed method. Moreover, we will optimize the system to further reduce the amount of supervision in the view of switching to an almost fully unsupervised algorithm.

Author Contributions

Conceptualization, G.F., N.L. and F.D.; methodology, G.F., N.L. and F.D.; software, G.F., N.L. and F.D.; validation, G.F., N.L. and F.D.; formal analysis, G.F., N.L. and F.D.; investigation, G.F., N.L. and F.D.; resources, G.F., N.L. and F.D.; data curation, G.F., N.L. and F.D.; writing–original draft preparation, G.F., N.L. and F.D.; writing–review and editing, G.F., N.L. and F.D.; visualization, G.F., N.L. and F.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors would like to thank Rayan Imam for the technical help with the preprocessing of the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Variational Representation of Divergence and Deep Infomax

In this appendix, we briefly review the concept of convex divergence between probability distributions, the particular case of mutual information (MI), and its usage for the construction of a feature extractor.
The class of f-convex divergences between distributions of two random variables X 1 p ( x ) , X 2 q ( x ) is defined as
D ( p | | q ) = p ( x ) f q ( x ) p ( x ) d x .
Besides from very special cases, such as multivariate Gaussian distributions, and simple f functions, no closed form exists for the computation. Moreover, in general, not even a parametric model about the distributions p , q is known, but all the information we have about the density functions is related to same empirical samples { x 1 i } i = 1 N 1 , { x 2 j } j = 1 N 2 drawn from the corresponding distributions.
The main idea of variational representation is that, roughly speaking, thanks to convex analysis, we can write the divergence as
D ( p | | q ) = sup g ( x ) G p ( x ) g ( x ) d x h q ( x ) t ( g ( x ) ) d x ,
where h , t are analytically known functions that can be computed once f is fixed. Thanks to this representation, we can estimate the divergence without knowing the actual distributions p , q using Monte Carlo methods:
D ( p | | q ) sup g ( x ) G i = 1 N 1 g ( x 1 i ) h j = 1 N 2 t ( g ( x 2 j ) ) .
An attentive reader might notice that since there is the problem of extremization sup g ( x ) G , we just moved the complexity of the problem to another point. The main idea discussed in [22], is that, however, what holds however is that
D ( p | | q ) = sup g ( x ) G · sup g ( x ) N · ,
where the space N G is chosen such that we can perform “easily” optimization (such as the space of Neural Networks, i.e., g θ ( x ) is a neural network with parameters θ ). We can finally estimate the divergence with the following approximation by optimizing (i.e., training) the network and using Monte Carlo techniques
D ( p | | q ) sup θ i = 1 N 1 g θ ( x 1 i ) h j = 1 N 2 t ( g θ ( x 2 j ) ) .
If g θ is sufficiently flexible, we can get arbitrarily close to the true value of the divergence. Notice that, since it is possible to compute the gradients of the functions g , h , t , end to end training is possible. In the particular case where the considered divergence is the mutual information (MI) between two random variables X 1 and X 2 , and that thus the dataset is composed { x 1 , x 2 i } i = 1 N , this corresponds to considering
I X 1 ; X 2 = D ( p ( x 1 , x 2 ) | | p ( x 1 ) p ( x 2 ) ) = p ( x 1 , x 2 ) log p ( x 1 , x 2 ) p ( x 1 ) p ( x 2 ) d x 1 d x 2 .
That can be approximated as
I X 1 ; X 2 sup θ i = 1 N g θ ( x 1 , x 2 i ) h j = 1 N t ( g θ ( x 1 , x 2 s c r a m b l e d j ) ) ,
where x 1 , x 2 i are simply samples taken from the dataset of joint variables (and thus extracted from p ( x 1 , x 2 ) ) while x 1 , x 2 s c r a m b l e d j are derived by randomly scrambling the indices of one of the two random variables in the dataset (thus having from a practical point of view statistical distribution equal to p ( x 1 ) p ( x 2 ) ). Remember that the functions h ( · ) , t ( · ) are known. Actually, in practice [21], maximizing instead the Jensen–Shannon Divergence (JSD) between p ( x 1 ) p ( x 2 ) and p ( x 1 ) p ( x 2 ) works better. After some manipulations, the final result is that the divergence can be simply written as:
JSD X 1 ; X 2 sup θ i = 1 N s p g θ ( x 1 , x 2 i ) j = 1 N s p g θ ( x 1 , x 2 s c r a m b l e d j ) ,
where s p ( z ) = log ( 1 + exp ( z ) ) is called soft-plus function.
Upon this idea in [21], they built an unsupervised compression tool of structured high dimensional random variables X (such as images or time series such as voice, etc...), called by the authors Deep Infomax. The (oversimplified) description of the system, schematically depicted in Figure A1 is that, using two neural network encoders e η 1 ( · ) , e β 2 ( · ) and two discriminator networks g θ g l o b ( · ) , g θ l o c ( · ) , it is possible to produce two latent representations H , F , respectively computed as H = e η 1 ( X ) , F = e β 2 ( H ) , such that both the MI between H , F and the Local Mutual Information M I l o c = 1 | H | z = 1 | H | I H ( z ) ; F (a new metric introduced by the authors of [21]) are maximized. We moreover clarify the role of r ζ ( · ) that is a discriminator network that tries to distinguish between samples drawn from F and samples drawn from a multivariate uniform distributions ( F ˜ U ( 1 , 1 ) ). The system is trained adversarially [27], where r ζ ( · ) and the rest of the network are the two competing agents (at equilibrium F has uniform distribution and this property is used to ensure that the latent variable covers evenly the latent space as a form of regularization). Notice that, in our implementation (see Section 5 and Appendix B), we only considered local mutual information and did not implement the portion of the system linked to the global, i.e., classical, mutual information.
Figure A1. Deep Infomax architecture.
Figure A1. Deep Infomax architecture.
Applsci 10 00381 g0a1

Appendix B. Neural Networks’ Architectures

In this appendix, we describe the details of the various networks as well as the optimizer settings. All the software has been written using MATLAB® (R2019a, The MathWorks, Inc.). In our notation, a single line of a table represents a layer. The convention is the following: at the beginning of the row, we have the type of layer as well as the nonlinearity associated with the layer (if present) and the indication of presence of BatchNorm (if the BN acronym is present). We have five types of layers in our network:
  • Input layer. It simply indicates the input of the system, the matrix X n .
  • 1d convolutional layer (indicated as conv_1d). Its parameters are the number of taps of the filters (Ntaps), the number of filters (Nch) as well as the stride (the decimation period at the output of the layer).
  • residual 1d convolutional layers: as standard conv_1d layers but with a residual connection (another set of filters that bypass the nonlinearity and whose output is summed to the output of the nonlinearity).
  • flatten layer. It simply indicates that the input of this layer (a multivariate time serie) is flattened (vectorized).
  • fully connected layers, whose argument is the dimension of the output.
We consider as nonlinearities the rectified linear unit (ReLu), the sigmoid (Sigm), and the hyperbolic tangent (Tanh). In the first convolutional network ( e β 1 ( · ) ), we included skip connections of the last element of the various sequences at the two layers that are fed to the fully connected classifier c ψ ( · ) .
All networks are trained using Adam optimizer with default settings and learning rate 0.001 for 30 epochs with batch size equal to 64.
Table A1. Architecture of e β 1 ( · ) .
Table A1. Architecture of e β 1 ( · ) .
input layer (size 64 × 1 × 10)    (layer corresponding to the input X)
res_conv1d (Ntaps = 4, Nch = 64, stride = 2), ReLu, BN
res_conv1d (Ntaps = 4, Nch = 128, stride = 2), ReLu, BN    (the output is H)
Table A2. Architecture of e β 2 ( · ) .
Table A2. Architecture of e β 2 ( · ) .
res_conv1d (Ntaps = 4, Nch = 64, stride = 2), Relu, BN
res_conv1d (Ntaps = 4, Nch = 128, stride = 2), Relu, BN
flatten layer
fully conncected (512), Relu, BN
fully conncected (512), Relu, BN
fully conncected (20),Tanh    (the output is F)
Table A3. Architecture of g θ l o c ( · ) .
Table A3. Architecture of g θ l o c ( · ) .
conv1d (Ntaps = 1, Nch = 64, stride = 1), Relu
conv1d (Ntaps = 1, Nch = 64, stride = 1), Relu
conv1d (Ntaps = 1, Nch = 1, stride = 1), Sigm
Table A4. r ζ ( · ) .
Table A4. r ζ ( · ) .
fully conncected (1024), Relu, BN
fully conncected (512), Relu, BN
fully connected (1), Sigm
Table A5. Architecture of c ψ ( · ) .
Table A5. Architecture of c ψ ( · ) .
fully conncected (512), ReLu, BN
fully conncected (512), ReLu, BN
fully conncected (1), Sigm    (the output is C ^ )

References

  1. Moreno, B.; Radicella, S.; De Lacy, M.; Herraiz, M.; Rodriguez-Caderot, G. On the effects of the ionospheric disturbances on precise point positioning at equatorial latitudes. GPS Solut. 2011, 15, 381–390. [Google Scholar] [CrossRef]
  2. Skone, S.; Knudsen, K.; De Jong, M. Limitations in GPS receiver tracking performance under ionospheric scintillation conditions. Phys. Chem. Earth Part A Solid Earth Geod. 2001, 26, 613–621. [Google Scholar] [CrossRef]
  3. Yeh, K.C.; Liu, C.H. Radio wave scintillations in the ionosphere. Proc. IEEE 1982, 70, 324–360. [Google Scholar]
  4. Kintner, P.; Ledvina, B.; De Paula, E. GPS and ionospheric scintillations. Space Weather 2007, 5. [Google Scholar] [CrossRef]
  5. Lee, J.; Morton, Y.; Lee, J.; Moon, H.; Seo, J. Monitoring and Mitigation of Ionospheric Anomalies for GNSS-Based Safety Critical Systems: A review of up-to-date signal processing techniques. IEEE Signal Process. Mag. 2017, 34, 96–110. [Google Scholar] [CrossRef]
  6. Taylor, S.; Morton, Y.; Jiao, Y.; Triplett, J.; Pelgrum, W. An improved ionosphere scintillation event detection and automatic trigger for GNSS data collection systems. In Proceedings of the Proceedings of the 2012 International Technical Meeting of The Institute of Navigation, Newport Beach, CA, USA, 30–31 January 2012. [Google Scholar]
  7. Linty, N.; Dovis, F. An Open-Loop Receiver Architecture for Monitoring of Ionospheric Scintillations by Means of GNSS Signals. Appl. Sci. 2019, 9, 2482. [Google Scholar] [CrossRef] [Green Version]
  8. Jiao, Y.; Hall, J.J.; Morton, Y.T. Automatic equatorial GPS amplitude scintillation detection using a machine learning algorithm. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 405–418. [Google Scholar] [CrossRef]
  9. Jiao, Y.; Hall, J.J.; Morton, Y.T. Performance Evaluation of an Automatic GPS Ionospheric Phase Scintillation Detector Using a Machine-Learning Algorithm. Navig. J. Inst. Navig. 2017, 64, 391–402. [Google Scholar] [CrossRef]
  10. Liu, Y.L.; Morton, Y.J.; Jiao, Y.J. Application of machine learning to the characterization of GPS L1 ionospheric amplitude scintillation. In Proceedings of the 2018 IEEE/ION Position, Location and Navigation Symposium (PLANS), Monterey, CA, USA, 23–26 April 2018; pp. 1159–1166. [Google Scholar]
  11. Linty, N.; Farasin, A.; Favenza, A.; Dovis, F. Detection of GNSS Ionospheric Scintillations based on Machine Learning Decision Tree. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 303–317. [Google Scholar] [CrossRef]
  12. Favenza, A.; Farasin, A.; Linty, N.; Dovis, F. A machine learning approach to GNSS scintillation detection: automatic soft inspection of the events. In Proceedings of the 30th International Technical Meeting of the Satellite Division Of the Institute of Navigation (ION GNSS+ 2017), Portland, OR, USA, 25–29 September 2017; pp. 4103–4111. [Google Scholar]
  13. Savas, C.; Dovis, F. Comparative performance study of linear and Gaussian kernel SVM implementations for phase scintillation detection. In Proceedings of the IEEE International Conference on Localization and GNSS (ICL-GNSS), Nuremberg, Germany, 4–6 June 2019; pp. 1–6. [Google Scholar] [CrossRef]
  14. Savas, C.; Dovis, F. The Impact of Different Kernel Functions on the Performance of Scintillation Detection Based on Support Vector Machines. Sensors 2019, 19, 5219. [Google Scholar] [CrossRef] [Green Version]
  15. Rezende, L.; De Paula, E.; Stephany, S.; Kantor, I.; Muella, M.; de Siqueira, P.; Correa, K. Survey and prediction of the ionospheric scintillation using data mining techniques. Space Weather 2010, 8, 1–10. [Google Scholar] [CrossRef]
  16. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  17. de Lima, G.; Stephany, S.; de Paula, E.; Batista, I.; Abdu, M. Prediction of the level of ionospheric scintillation at equatorial latitudes in Brazil using a neural network. Space Weather 2015, 13, 446–457. [Google Scholar] [CrossRef]
  18. Uwamahoro, J.C.; Habarulema, J.B. modeling total electron content during geomagnetic storm conditions using empirical orthogonal functions and neural networks. J. Geophys. Res. Space Phys. 2015, 120, 11000–11012. [Google Scholar] [CrossRef]
  19. Habarulema, J.B.; McKinnell, L.A.; Opperman, B.D. Regional GPS TEC modeling; Attempted spatial and temporal extrapolation of TEC using neural networks. J. Geophys. Res. Space Phys. 2011, 116. [Google Scholar] [CrossRef]
  20. Kingma, D.P.; Mohamed, S.; Rezende, D.J.; Welling, M. Semi-supervised learning with deep generative models. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2014; pp. 3581–3589. [Google Scholar]
  21. Hjelm, R.D.; Fedorov, A.; Lavoie-Marchildon, S.; Grewal, K.; Trischler, A.; Bengio, Y. Learning deep representations by mutual information estimation and maximization. arXiv 2018, arXiv:1808.06670. [Google Scholar]
  22. Belghazi, M.I.; Baratin, A.; Rajeswar, S.; Ozair, S.; Bengio, Y.; Courville, A.; Hjelm, R.D. Mine: Mutual information neural estimation. arXiv 2018, arXiv:1801.04062. [Google Scholar]
  23. Nguyen, X.; Wainwright, M.J.; Jordan, M.I. Estimating divergence functionals and the likelihood ratio by convex risk minimization. IEEE Trans. Inf. Theory 2010, 56, 5847–5861. [Google Scholar] [CrossRef] [Green Version]
  24. Ruderman, A.; Reid, M.; García-García, D.; Petterson, J. Tighter variational representations of f-divergences via restriction to probability measures. arXiv 2012, arXiv:1206.4664. [Google Scholar]
  25. Nowozin, S.; Cseke, B.; Tomioka, R. f-gan: Training generative neural samplers using variational divergence minimization. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2016; pp. 271–279. [Google Scholar]
  26. Poole, B.; Ozair, S.; Oord, A.v.d.; Alemi, A.A.; Tucker, G. On variational bounds of mutual information. arXiv 2019, arXiv:1905.06922. [Google Scholar]
  27. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2014; pp. 2672–2680. [Google Scholar]
  28. Curran, J.T.; Bavaro, M.; Morrison, A.; Fortuny, J. Developing a multi-frequency for GNSS-based scintillation monitoring receiver. In Proceedings of the 27th International Technical Meeting of the Satellite Division of the Institute of Navigation (ION GNSS+), Tampa, FL, USA, 8–12 September 2014; pp. 1142–1152. [Google Scholar]
  29. Cristodaro, C.; Dovis, F.; Linty, N.; Romero, R. Design of a configurable monitoring station for scintillations by means of a GNSS software radio receiver. IEEE Geosci. Remote. Sens. Lett. 2018, 15, 325–329. [Google Scholar] [CrossRef]
  30. Ribeiro, M.T.; Singh, S.; Guestrin, C. Why should i trust you?: Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  31. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Semi-supervised learning. Feature extraction is performed in an unsupervised fashion, while classification is based on a supervised approach (labels).
Figure 1. Semi-supervised learning. Feature extraction is performed in an unsupervised fashion, while classification is based on a supervised approach (labels).
Applsci 10 00381 g001
Figure 2. Deep Infomax—high level scheme.
Figure 2. Deep Infomax—high level scheme.
Applsci 10 00381 g002
Figure 3. High level scheme of architecture used in this work.
Figure 3. High level scheme of architecture used in this work.
Applsci 10 00381 g003
Figure 4. Detailed representation of high level scheme depicted in Figure 3.
Figure 4. Detailed representation of high level scheme depicted in Figure 3.
Applsci 10 00381 g004
Figure 5. Neural Network architecture (Training). The dashed red box contains the unsupervised feature extraction part, while the orange one contains the supervised classification.
Figure 5. Neural Network architecture (Training). The dashed red box contains the unsupervised feature extraction part, while the orange one contains the supervised classification.
Applsci 10 00381 g005
Figure 6. Neural Network architecture (Testing).
Figure 6. Neural Network architecture (Testing).
Applsci 10 00381 g006
Figure 7. Accuracy and F-score of Deep Infomax (DIM) method and Decision Tree (DT) [11] method.
Figure 7. Accuracy and F-score of Deep Infomax (DIM) method and Decision Tree (DT) [11] method.
Applsci 10 00381 g007
Figure 8. Feature importance analysis using random permutation method. The lower the accuracy with respect to the baseline, the higher the feature importance.
Figure 8. Feature importance analysis using random permutation method. The lower the accuracy with respect to the baseline, the higher the feature importance.
Applsci 10 00381 g008
Table 1. Varying of the considered performance metrics versus percentage of labeled datapoints (% labeled).
Table 1. Varying of the considered performance metrics versus percentage of labeled datapoints (% labeled).
% LabeledAccuracyPrecisionRecallF-Score T n F p F n T p
100.96330.92540.92670.92600.73390.01850.01820.2294
200.97980.97570.94440.95980.73890.00600.01420.2409
300.99290.98580.98600.98590.74630.00360.00350.2466
400.99390.98250.99320.98780.74530.00440.00170.2486
500.99630.99260.99280.99270.74590.00190.00180.2504
600.99640.99240.99290.99270.75110.00190.00180.2452
700.99750.99510.99490.99500.74850.00120.00130.2490
800.99890.99840.99720.99780.74940.00040.00070.2494
900.99850.99570.99840.99700.74380.00110.00040.2546
1000.99930.99860.99860.99860.75210.00040.00040.2472

Share and Cite

MDPI and ACS Style

Franzese, G.; Linty, N.; Dovis, F. Semi-Supervised GNSS Scintillations Detection Based on DeepInfomax. Appl. Sci. 2020, 10, 381. https://doi.org/10.3390/app10010381

AMA Style

Franzese G, Linty N, Dovis F. Semi-Supervised GNSS Scintillations Detection Based on DeepInfomax. Applied Sciences. 2020; 10(1):381. https://doi.org/10.3390/app10010381

Chicago/Turabian Style

Franzese, Giulio, Nicola Linty, and Fabio Dovis. 2020. "Semi-Supervised GNSS Scintillations Detection Based on DeepInfomax" Applied Sciences 10, no. 1: 381. https://doi.org/10.3390/app10010381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop