Next Article in Journal
Laser Spot Centering Algorithm of Double-Area Shrinking Iteration Based on Baseline Method
Next Article in Special Issue
Predicting the Porosity in Selective Laser Melting Parts Using Hybrid Regression Convolutional Neural Network
Previous Article in Journal
CFD Body Force Propeller Model with Blade Rotational Effect
Previous Article in Special Issue
Application of Neural Networks for Water Meter Body Assembly Process Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Generalized Hybrid Probability Convolutional Neural Network

1
College of Physics and Electronic Information, Gannan Normal University, Ganzhou 341000, China
2
College of Computer, Chengdu University, Chengdu 610106, China
3
New Engineering Industry College, Putian University, Putian 351100, China
4
Engineering Research Center of Big Data Application in Private Health Medicine, Fujian Province University, Putian 351100, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(21), 11301; https://doi.org/10.3390/app122111301
Submission received: 29 September 2022 / Revised: 1 November 2022 / Accepted: 2 November 2022 / Published: 7 November 2022
(This article belongs to the Special Issue Deep Convolutional Neural Networks)

Abstract

:
This paper first studies the generalization ability of the convolutional layer as a feature mapper (CFM) for extracting image features and the classification ability of the multilayer perception (MLP) in a CNN. Then, a novel generalized hybrid probability convolutional neural network (GHP-CNN) is proposed to solve abstract feature classification with an unknown distribution form. To measure the generalization ability of the CFM, a new index is defined and the positive correlation between it and the CFM is researched. Generally, a fully trained CFM can extract features that are beneficial to classification, regardless of whether the data participate in training the CFM. In the CNN, the fully connected layer in the MLP is not always optimal, and the extracted abstract feature has an unknown distribution. Thus, an improved classifier called the structure-optimized probabilistic neural network (SOPNN) is used for abstract feature classification in the GHP-CNN. In the SOPNN, the separability information is not lost in the normalization process, and the final classification surface is close to the optimal classification surface under the Bayesian criterion. The proposed GHP-CNN utilizes the generalization ability of the CFM and the classification ability of the SOPNN. Experiments show that the proposed network has better classification ability than the existing hybrid neural networks.

1. Introduction

The effect of pattern classification is mainly determined by classification features and classifiers. For an image classification task, the classification model needs to extract image features that are easy to distinguish, and then design a suitable classifier that matches the features. Traditional artificial feature extraction methods have achieved good results, such as the local binary pattern (LBP) [1,2], the speeded-up robust features (SURF) [3,4] and the scale-invariant features (SIFT) [5,6]. However, these features are single features without any hierarchical structure, sometimes lacking good generalization abilities. In recent years, the convolutional neural network (CNN) has attracted the extensive attention of scholars in many fields [7,8,9,10], because CNNs can extract various shallow features and deep features of images.
Based on the MLP training algorithm, LeCun et al. [11] designed a convolutional neural network called LeNet-5, which was an early classic convolutional neural network model. It was mainly used for handwritten digit recognition on checks. With a gradual increase in the number of convolutional layers, the CNN’s feature extraction classification abilities are becoming stronger and stronger. Similarly, graph convolution networks (Graph CNNs) have good feature extraction ability [12,13,14], because a Graph CNN is also very effective for non-Euclidean data, and is relatively more complex. In this paper, we only discuss CNNs. Krizhevsky [15] proposed the extended deep convolutional neural network AlexNet, which used the ReLU function instead of the saturated nonlinear function tanh. It also used the dropout technique to avoid overfitting in the training process. In 2014, Google proposed a convolutional neural network with more than 20 layers, GoogLeNet [16], to resolve the gradient disappearance and two loss functions were designed at different depths of the network. The VGG [17] network model proposed by Oxford University adopted a layer-by-layer training method. The trained VGG network showed a good ability to extract image features. The residual network ResNet proposed in [18] allowed the original input information to be passed directly to the subsequent network layer by adding connected channels directly. ResNet could train hundreds or even thousands of layers while achieving a good classification accuracy. In general, the above various types of convolutional neural networks have achieved good results in different scenarios, mainly because the convolutional layer of a CNN can extract abstract features with good separability. However, because the distribution of these abstract features is unpredictable, designing a suitable classifier for these features is a problem worthy of further study.
The above CNNs can extract all kinds of image features that are good for image classification, but the MLP classifiers in CNNs sometimes fall into local optimal [19,20,21,22,23,24,25,26,27]. To solve this problem, Niu et al. [19] proposed using a pretrained CNN as the feature extractor to extract the hierarchical features of an image, and then using an SVM as the classifier. The beginning of the method mainly focused on the binary classification problem. When applied to multiclass classification problems, it mainly adopted multiple two-classifier voting methods. In a multiple-two-classification problem, if a sample has the highest number of votes in two different categories, the SVM divides the sample into either category, which may lead to classification errors. Duan et al. [20] proposed using the pretrained CNN as a feature extractor to extract facial features, and then using an extreme learning machine (ELM) as a classifier. However, it was generally necessary to reset the number of hidden nodes when using ELM classifiers on different dataset. Tien et al. [26] used depth features and artificial features to merge into new hybrid features, and then used an SVM as a classifier. The new hybrid features combined the characteristics of deep features and artificial features and had a high separability. Guo et al. [27] proposed a multinetwork feature fusion structure that used merged features to classify features. Merged features have more information than a single feature, which is beneficial in classification tasks. Although studies have shown that convolutional neural networks have a good ability to extract image features and have good classification capabilities, training a deep convolutional neural network takes a long time and requires a large quantity of data. When a new dataset needs to be classified, the network needs to be retrained.
The above literature review shows that the convolutional layer as a CFM, which is fully trained by a multicategory source dataset, can extract image features that are beneficial to the classification, but there are no studies on the relationship between the generalization ability of the CFM-extracted features, the number of source dataset and target dataset categories. In this paper, the trained CNN on the multiclass source dataset does not need retraining. It directly extracts image features from new target tasks and then cascades the appropriate classifier. This can not only use the ability of the CNN convolutional layer to extract image features, but also makes use of the advantages of different classifiers.
The main contributions of this paper include the following three points. First, the general rule between the generalization ability of the CNN’s convolutional layer as a CFM, and the number of categories of the source dataset and target dataset are summarized. Then, the interpretability of the generalization ability of the CFM is given. Second, a new GHP-CNN is proposed to solve the problem of a target image classification. The GHP-CNN mainly includes a CFM that extracts the features of the target set and a classifier called the structure-optimized probability neural network (SOPNN). The SOPNN classifier is not restricted by the distribution form of the target feature. The GHP-CNN avoids the problem of MLP classifiers in CNN networks easily falling into local optima and improves the classification ability.
The organizational structure of the rest of this paper is as follows. In Section 2, the generalization ability of the CFM to extract target image features is studied. In Section 3, a structure-optimized probabilistic neural network classifier is designed. In Section 4, the GHP-CNN is proposed, and its principle and characteristics are analyzed. Section 6 presents the analytical and experimental results. The last section summarizes the conclusions of the study. Some important symbols used in this article are shown in Table 1.

2. Research on the Generalization Ability of the CNN Feature Mapper (CFM) to Extract Image Features

In this paper, the dataset S * for training the CNN is called the source dataset, and the dataset T * for testing the CFM generalization ability is called the target dataset. The number of sample categories in S * is s ( s = s 1 , , s m ) and the number of sample categories in T * is t ( t = t 1 , , t m ) . In general, the sample categories in the source dataset S * and the target dataset T * are not the same. This paper comprehensively studies the relationship between category reduction, unknown category expansion and the generalization ability of CNNs to extract target image features. In this paper, the category reduction refers to the source dataset S * containing the target dataset T * ( T * S * ). An unknown category extension means that there is no sample that is common to both the source dataset and the target dataset ( S * T * = ). The network used in this section was VGG-16 [17] (stride =1, padding = 1, MaxPooling), and the dataset used was CIFAR-100 [28].
The convolution neural network trained on the source dataset S * can be simply divided into two parts: the convolutional layer used to extract the feature and the MLP layer as the classifier. The convolutional network is fully trained on the source dataset until it converged, and the convolutional layer migrates as the CFM of the target set. As shown in Figure 1, after the source dataset S * is trained by the convolutional network and the MLP layer, the corresponding classification accuracy P s is obtained, the feature mapper corresponding to the source dataset S * of the s class is CFM s , and the target dataset T * of the t class is passed through the CFM s . The features are extracted, and then classification training is performed with the MLP classifier, and the corresponding accuracy P s t is obtained
For the different affiliations between the source dataset S * and the target dataset T * , and to make the experiment statistically valid, when training multiple CFM s , s different categories of the source dataset and t categories of the target dataset were selected, where each s and t category samples satisfied the uniform distribution in the CIFAR-100 dataset. Multiple source datasets S s 1 * , S s 2 * , , S s m * were selected, where S s 1 * S s 2 * , , S s m * . When T * S * , multiple target datasets T t 1 * , T t 2 * , , T t m * were selected, where T t 1 * T t 2 * , , T t m * .
For each t-class target dataset, there are an average classification accuracy P ˜ s t and a classification accuracy P s of the s-class source dataset corresponding to the CFM s . Comparisons between the magnitudes of P ˜ s t and P s show the generalization ability of the CFM s in the t-class target dataset.
r s t = P ˜ s t P s P s
The size of r s t can measure the strength of the CFM s ’s generalization ability. That is, the greater the value of r s t is, the stronger the generalization ability of the CFM s . When r s t > 0 , the CFM s can extract features with better separability for the t-class targets and can obtain a higher accuracy than P s under subsequent classifiers. That is, a CFM trained by the s-class source dataset S * has a better generalization ability for a t-class target dataset T * . When r s t < 0 , the CFM s fails to extract features with better separability in the t-class target dataset, which results in a lower accuracy than the original P s under subsequent classifiers. The corresponding classification accuracy can be obtained by changing the membership relationship between S * and T * , and the generalization ability of the CFM s under different conditions can be further reflected by increasing or decreasing the magnitude of the classification accuracy compared with the original accuracy.

2.1. Analysis of the Generalization Ability of the Feature Mapper to the Target Set after Category Reduction ( T * S * )

When T * S * , the trained CFM s on multiclass source dataset S * extracts the features of target dataset T * and trains the classification by the MLP classifier. In the CIFAR-100 dataset, we selected the source dataset with the number of categories for s ( s = 15 , 20 , 25 , , 100 ) to be fully trained with the VGG-16 algorithm, we obtained different original precisions P s and the corresponding 15 CFM s s, and then randomly extracted the target dataset T * with the number of categories t ( t = 5 , 10 , , 30 ) from the source dataset S * . The features of these target dataset T * were extracted by the CFM s , and then the training classification was performed by the MLP classifier. To make the results more universal, the target dataset of each category are randomly extracted eight times. The classification accuracy P s t was counted each time, and the final result was the average value of the multiple classification results, where t represented the number of categories in the target dataset T * , and s represented the number of categories in the source dataset S * .
As shown in Figure 2, it can be seen that for the determined t categories from target dataset T * , the value of r s t was positively correlated with s: as the corresponding s of the CFM s increased, r s t increased monotonously. That is, for a certain target dataset with t categories, the larger the value of s was, the better the generalization ability of the CFM s trained by the source dataset S * . For a CFM trained on the determined s source dataset S * , the value of r s t was negatively correlated with T * and with the number of t categories in the target dataset; r s t decreased monotonously as t increased. That is, for a CFM s that determined the number of categories, a target dataset T * with a smaller number of t had a better the generalization ability than a target dataset with larger values of t. When r s t was a certain value, that is, after the feature was extracted under the CFM s for t different categories, and the magnitude of the classification promotion was the same, the number of CFM s for different values of t corresponding to s was different.

2.2. Analysis of the Generalization Ability of the Feature Mapper to Unknown Target Categories ( S * T * = )

When S * T * = , the trained CFM s on a multiclass source dataset S * extracts the features of target dataset T * and the features are used for training the MLP classifier. In the CIFAR-100 dataset, 90 categories were randomly selected as a candidate source dataset, and the remaining 10 categories were used as a candidate target dataset. The sample sets with number of categories s ( s = 15 , 20 , 25 , , 90 ) were selected from the 90 candidate source dataset to be fully trained by VGG-16, and the relatively best classification results and corresponding 14 CFM s were obtained and randomly selected from the 10 candidate dataset. The unknown target dataset with a number of categories t ( t = 2 , 3 , , 8 ) was selected, and the features of the randomly selected unknown target dataset were extracted by the CFM and then trained by the MLP classifier. To make the results more universal, each number of unknown sample subsets was extracted three times, the classification results were counted each time, and the final result was the average of three results.
As shown in Figure 3, for the determined t category target dataset T * , the value of r s t was positively correlated with s; as the corresponding s of the CFM s increased, r s t increased monotonously. That is, for a certain t category target dataset, the larger s was, the better the generalization ability of the CFM s trained by the source dataset S * . For the CFM s trained from the determined s category source dataset S * , the value of r s t was negatively correlated with the number of t categories in the target dataset T * ; as t increased, r s t decreased monotonously. That is, for a CFM s that determined the number of categories, the generalization ability of the target dataset T * with a smaller t was better. When r s t was a certain value, that is, after the feature was extracted under the CFM s for different t categories from the target dataset, and the magnitude of the classification promotion was the same, then the number of CFM s for t categories corresponding to s categories was different.

2.3. Analysis of the CFM Generalization Ability

Through the above experiments, which were from different angles, the generalization ability of CFM was analyzed in a large number of ways. From the analysis, the generalization ability of a trained CNN as a CFM was related to the number of categories in the source dataset(t). The more categories in the source dataset were trained by the CFM, the better the generalization ability of the CFM, and the closer it was to global optimum characteristics.
In the above experiments, the features extracted by the CFM were input to the MLP for intensive training, which could be regarded as a fine-tuning of the MLP classifier after parameter migration. Although the network after fine-tuning the MLP classifier also had a better classification accuracy, it might also fall into a local optimum, and the internal function mapping relationship would be difficult to explain. Since the image features extracted by CFM s are generally abstract, the feature distribution is unknown. The classifier based on the Bayesian probability density is not restricted by the distribution of target features, and the final classification surface is close to the optimal classification surface under the Bayesian criterion. We design an optimized Bayesian classifier in the next section.

3. SOPNN Classifier in the GHP-CNN

3.1. The Overall Structure of the SOPNN Classifier

As shown in Figure 4, the SOPNN classifier consists of four layers. The first layer is the input layer for receiving the input feature vectors. A traditional PNN adopts a unit hypersphere normalization for the pattern vectors and input samples, which ensures that the distance between different vectors in a network can be expressed in a simple inner product form. That is, the squared distance between x and y can be expressed as x y = 2 2 x T y if x and y are normalized. Note that such a simple inner form can be easily realized with a neuron structure. However, with a hypersphere normalization, all sample points that are collinear with the sphere center overlap. To avoid the separability problem in the hypersphere normalization process, the SOPNN used the unit hyperspherical crown normalization x ^ = x ¯ / x ¯ from reference [29], where x ¯ = [ x , h ] R n + 1 and x are n-dimensional features of the target image extracted by the CFM. h is the additional dimension.
The second layer is the pattern layer, and the pattern nodes in this layer receives the vector x ^ from the input layer. Let a pattern node u ^ m j be the mth center in the jth class ( 1 j c , 1 m M j ). Its output can be given by
ϕ m j ( x ^ ) = exp x ^ u ^ m j T x ^ u ^ m j 2 σ m j 2 ( 2 π ) ( n + 1 ) / 2 σ m j ( n + 1 )
where σ m j denotes the kernel width, M j denotes the number of pattern nodes in the jth class and the total number of pattern nodes is M. That is, M = j = 1 c M j .
The third layer is the summation layer, which includes c neurons and one neuron per class. Each neuron calculates a conditional probability density function y j ( x ^ ) for each class using the following weighted summation form.
y j ( x ^ ) = m = 1 M j λ m j ϕ m j ( x ^ ) , 1 j c
where λ m j denotes the prior distribution information of the training samples of the mth center in the jth class and satisfies m = 1 M j λ m j = 1 ( 1 j c ) .
The fourth layer is the output layer, which makes decisions according to the following equation.
C * ( x ) = arg max { p j y j ( x ^ ) } , j = 1 , 2 , , c
where c is the class number and C * ( x ) is the predicted class of x . p j denotes the prior probability of the jth class. Obviously, all the parameters have an interpretable physical meaning in our classifier.

3.2. The Hidden Parameter Optimization in the SOPNN Using the EM Algorithm

The EM algorithm solves the parameter estimation problem using iterative methods in various probability models. Each iteration consists of two steps, called the E step and the M step. The E step uses the existing estimated parameters and the observed data to calculate the expectation of the hidden parameters. The M step uses the expectation obtained in the E step and maximum likelihood estimation to estimate the hidden parameter.
Let a set X that is used for training the SOPNN contain L samples from c classes, and let the jth class contain L j training samples. That is, X = x l l = 1 L = x l j l = 1 L j j = 1 c . With a hyperspherical crown mapping, X can be projected as X ^ = x ^ l l = 1 L = x ^ l j l = 1 L j j = 1 c . According to the independence of the sample, the corresponding log-likelihood function can be obtained:
L o g = j = 1 c l = 1 L j log y j x ^ l j = j = 1 c l = 1 L j log m = 1 M j λ m j ϕ m j x ^ l j
Using the EM algorithm to solve for the maximum value of the likelihood function, the following iterative formulas can be obtained:
u ^ m j = l = 1 L j w m j x ^ l j x ^ l j l = 1 L j w m j x ^ l j
σ m j 2 = l = 1 L j w m j x ^ l j x ^ l j u ^ m j T x ^ l j u ^ m j ( n + 1 ) l = 1 L j w m j x ^ l j
λ m j = l = 1 L j w m j x ^ l j L j
w m j x ^ l j = λ m j ϕ m j x ^ l j m = 1 M j λ m j ϕ m j x ^ l j
If the initial pattern node and kernel width are known, we can obtain the corresponding parameter value λ m j in Equation (4) by using the above iterative formulas. The termination criterion of the EM algorithm is that the absolute value of the difference between the hidden parameter θ ( i + 1 ) of the ( i + 1 ) th order and the hidden parameter θ ( i ) of the ( i ) th order is less than a given positive ϵ , where ϵ = 0.01 .
There are three differences between the used SOPNN classifier in this paper and the traditional probability neural network (PNN) classifier. First, the hyperspherical crown normalization method in the SOPNN is a map without separable information loss, and the hypersphere normalization in a traditional PNN is a map with separable information loss. The conclusion that the hyperspherical crown normalization is superior to the traditional hypersphere normalization in terms of separability was proven in [29]. Second, when the training set is large, the number of pattern nodes in the traditional PNN is equal to the number of training samples, which leads to a complex network structure. In this paper, the kernel coverage method based on the potential function in [30] is used to adaptively generate pattern nodes, which makes the number of pattern nodes far fewer than the number of training samples. This ensures that the network structure of the SOPNN is simpler than that of the traditional PNN. Third, the SOPNN assigns a local intraclass weight coefficient to each pattern node, which reflects the prior distribution information of the local region of the pattern node. Therefore, the probability density estimation in the SOPNN is more accurate than in the traditional PNN, and its classification surface can better approximate the optimal classification surface under the Bayesian criterion. The hidden parameters λ m j and σ m j in the SOPNN can be optimized by using the expectation maximization algorithm.

4. GHP-CNN Neural Network

Structure and Characteristics of the GHP-CNN

The network structure of the GHP-CNN in this paper is shown in Figure 5. The GHP-CNN mainly consists of three steps in the training process. First, the CNN is adequately trained with the source data. The source data information is propagated forward and backward until the CNN converges. Second, the parameters of the convolutional layer can be obtained in the first step of training, and the convolutional layer with fixed parameters is transferred as a separate CFM to extract the image features of the training samples in the target set. Third, the SOPNN is trained on the target dataset. The image features of the target training set are input into the SOPNN for training, and a well-trained SOPNN classifier can be obtained. SOPNN training only requires forward propagation once and does not require backpropagation. After completing the above three steps, we obtain the trained GHP-CNN network. If the test image in the target dataset is input directly into the CFM, then the SOPNN classifier in the GHP-CNN can be used for target classification. The proposed GHP-CNN has the following three characteristics.
First, the GHP-CNN makes full use of the generalization ability and the parameter transferability of the CFM to extract image features. It only needs to transfer the convolutional layer as a CFM, and we do not need to transfer the entire CNN network.
Second, there is no need to retrain the CNN on the target set, only the SOPNN needs to be trained. The GHP-CNN can still be trained and work well even if the target dataset is small. Since training the SOPNN does not require backpropagation, the efficiency of training the SOPNN on the target set is significantly better than that of retraining the CNN.
Third, the distribution of the features extracted by the CFM is unknown. The SOPNN classifier in the GHP-CNN is based on probability density estimation and it is not limited by the distribution of the extracted features. Moreover, the interface of the SOPNN is close to the optimal classification surface under the Bayesian criterion, and there is no local optimal problem with the MLP classifier in the CNN.
Given the target set and initial network. The main steps of the GHP-CNN learning algorithm can be summarized as follows:
Step 1. According to the relationship between d e l t a , s and t (as shown in Figure 2 and Figure 3), we can determine the number of categories in the source dataset and train various CFMs with different generality abilities.
Step 2. The CFM trained in the source dataset are transferred to extract the features of the target training set with the determined category number.
Step 3. The extracted features in step 2 are input into the SOPNN and are normalized by a hyperspherical crown normalization mapping. The appropriate parameter h should guarantee that the normalized data have the optimal separability under the Fisher criterion.
Step 4. The initial pattern node set of the SOPNN is generated by the adaptive kernel coverage method [30]. Because the number of generated pattern nodes is far less than the number of training samples, the SOPNN classifier with the optimized structure is obtained. A potential value is used to measure the distribution density of samples in the sample space, and the sample with the highest potential value is selected as the pattern node each time. Then, the potential value of the selected pattern region is iteratively eliminated, and the remaining samples are reordered according to the potential value.
Step 5. The parameters of the SOPNN are optimized by the expectation maximization method (Equations (7)–(10)), and then the well-trained GHP-CNN is obtained.
Step 6. The extracted features of the target test set are input into the SOPNN classifier in the GHP-CNN for classification.

5. Experiment and Results

5.1. Research on the GHP-CNN’s Classification Capabilities after Category Reduction

Through the analysis of the CFM’s generalization ability in the second part, the more source dataset categories used for training the CFM, the better the generalization ability. In the experiment, we used the CIFAR-100 and Caltech-256 datasets and the VGG and ResNet networks. After fully training VGG-16 with the CIFAR-100 dataset, the average accuracy of the 100-class test was 69.58%, and the fully trained convolutional layer was used as the CFM. Four, six, eight, twelve and sixteen subcategories were randomly selected from the CIFAR-100 data, and the corresponding classification accuracy of the subclasses was obtained after training with VGG-16. The features of these subclasses were extracted by the CFM and trained and tested with hybrid networks such as VGG-16-E, CNN-PNN, CNN-SVM, CNN-ELM and GHP-CNN. VGG-16-E is the result of training the MLP using the CFM in VGG-16. The results are shown in Table 2. After fully training ResNet-152 with Caltech-256, ResNet-152-E was the result of training the MLP using the CFM in ResNet-152. The results are shown in Table 3.
In Table 2, when s = 100 , the average classification accuracy of the corresponding 100 classes was determined, but for each different subclass t, the corresponding classification accuracy was different. When the number of source data categories was determined ( s = 100 ), the corresponding CFMs could extract different categories of target features ( t = 4 , 6 , 8 , 12 , 16 ) and have significantly improved classification effects in different classifiers. Overall, the CFM trained by a multiclass dataset could extract features with a high separability and had a good generalization ability of feature mapping. When subclasses were reclassified, the results of the comparison algorithm were significantly higher than the original accuracy. Our GHP-CNN method had the highest classification accuracy. In Table 3, when s = 256 , the average classification accuracy of the corresponding 256 classes was determined. When subclasses were reclassified, the results of the comparison algorithm were significantly higher than the original accuracy. Our GHP-CNN method had the highest classification accuracy.

5.2. Research on the Classification Ability of the GHP-CNN with Unknown Categories

The CFM trained by the multiclass dataset extracted the features of unknown classes and verified the classification performance of the VGG-16-E, the CNN-PNN, the CNN-SVM, the CNN-ELM and the GHP-CNN classifiers. Ninety categories were randomly selected from the CIFAR-100 dataset as a candidate source dataset for training the CFM, and the remaining 10 categories were used as the unknown target dataset. The classification accuracies for the unknown categories are shown in Table 4, when we trained ResNet-152 with Caltech-256. The classification accuracies for the unknown categories are shown in Table 5.
In Table 4, the CFM s trained from the multiclass source dataset also had a good generalization ability for unknown subclasses. The accuracy of the proposed GHP-CNN was significantly higher than that of the comparison algorithms. In Table 5, the accuracy of the proposed GHP-CNN was significantly higher than that of the comparison algorithms.

5.3. Result Analysis

In the comparison algorithm, the MLP classifier performed intensive training on the mapped subclass features, and the accuracy was higher than the original accuracy. However, because the MLP classifier could easily fall into a local optimum during the training process, the overall performance was not good, and the CNN-PNN, CNN-SVM, CNN-ELM and GHP-CNN classifiers had better results. The GHP-CNN avoided some loss of separability information caused by the hypersphere normalization in the traditional PNN and retained the optimal classification surface under the Bayesian criterion. The classification ability of the GHP-CNN was obviously better than the other algorithms. In addition, the accuracy of the classification of unknown categories did not reach the classification accuracy of subclasses, because the generalization ability of the mapping of unknown subclass features was not as good as the generalization ability of the mapping of subcategory features. The separability of mapping features was not good enough, and the features of different categories were mutually integrated.

5.4. Parameter Analysis

In the experiment of the generalizability to unknown categories, the parameter A of a potential function [30] in the GHP-CNN represents the distance weighting factor between the vectors. By adjusting this parameter, the potential value between the two vectors can be changed. When the potential value between two vectors is assumed to be a fixed value, the larger the distance between the two vectors, the smaller A is. When the distance between two vectors is assumed to be a fixed value, the larger A is, the smaller the potential value between the sample vectors. Figure 6 shows that when parameter A changed in a small range, the change in classification accuracy was relatively stable, which showed that the algorithm had a good robustness to parameter A.
The size of the initial kernel width σ affects the coverage of pattern nodes of the GHP-CNN, thereby affecting the number of mode layer nodes. If the value of σ is too small, then the coverage region of the generated pattern node does not contain other sample points, and the PNN classifier is close to the nearest neighbor classifier, which is equivalent to using a training sample to represent a pattern node, and the structure of the network pattern layer cannot be simplified in this case. If the value of σ is too large, the selected pattern node cannot reflect the real sample distribution and may increase the number of heterogeneous samples in the kernel, which seriously affects the classification accuracy. In the algorithm of this paper, the input sample was a vector after the hyperspherical mapping, so that σ could be kept in a small range for proper selection. As shown in Figure 7, when σ was between 0.05 and 0.25, the GHP-CNN algorithm could achieve a relatively good classification accuracy.
Note that different h values lead to differences in the separability of the normalized samples. We can select h according to the following separability criterion.
h = arg max h [ t r ( SM b ) t r ( SM t ) ]
where SM b and SM t denote the between-class scatter matrix and total scatter matrix of the normalized training samples, respectively. However, as a new dimension, the value of h should be matched with the other dimensions. Therefore, the initial value of h can be selected as follows.
h = a ( w = 1 n Δ w ) 1 n + ( 1 a ) ( det ( SM t ) ) 1 n
where a [ 0 , 1 ] . w = 1 n Δ w denotes the volume of an n-dimension hypercube for the target sample space, and det ( SM t ) measures the volume of an n-dimension superellipsoid of the sample scatter for the target set. We compute the two volumes differently because the actual distribution shape of the features of the target set is generally irregular.
To ensure fairness in the comparison, all algorithms were tested using the same CPU. The unit of time consumed was in seconds. As shown in Table 6, the computational efficiency of the CNN-ELM and CNN-SVM algorithms was significantly higher than that of the other algorithms. The VGG-16-E algorithm took the most time because the node parameters of the MLP layer were much larger than in the other classifiers. The time required for the GHP-CNN algorithm proposed in this paper was much lower than the VGG-16-E test time.

6. Conclusions

In this paper, the generalization ability of a CFM to extract image features was studied and a novel quantization index for the CFM was defined. The analysis showed that the trained CFM could obtain abstract features with good separability. Since the distribution of the extracted features was unknown and the MLP classifier in the fully connected layer sometimes fell into a local optimum, an improved classifier named the structure-optimized probabilistic neural network (SOPNN) was used for various abstract feature classifications. Utilizing the characteristics of the CFM and the classification ability of the SOPNN, a novel generalized hybrid probability convolutional neural network (GHP-CNN) was proposed. Experiments showed that, for the same abstract features, the proposed GHP-CNN was superior to the existing hybrid neural networks in terms of classification ability.

Author Contributions

Conceptualization, writing—original draft preparation, W.Z. and J.Z.; methodology, H.F. and J.Z.; validation, H.F. and J.Z. and H.W.; formal analysis, H.W.; investigation, Y.X.; writing—review and editing, visualization, W.Z. and H.F.; supervision, project administration, funding acquisition, H.W. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the PHD Research Foundation of Gannan Normal University (grant no. BSJJ202261); the National Natural Science Foundation of China (grant no. 62276146); the Open Foundation of Engineering Research Center of Big Data Application in Private Health Medicine, Fujian Province University (grant no. MKF202201); the Fujian Provincial Natural Science Foundation Projects (grant no. 2022J011171).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to copyright.

Conflicts of Interest

The authors declare no conflict of interest.

Sample Availability

Not applicable.

References

  1. Ahonen, T.; Hadid, A.; Pietikainen, M. Face description with local binary patterns: Application to face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 2037–2041. [Google Scholar] [CrossRef] [PubMed]
  2. Satpathy, A.; Jiang, X.; Eng, H.L. LBP-based edge-texture features for object recognition. IEEE Trans. Image Process. 2014, 23, 1953–1964. [Google Scholar] [CrossRef] [PubMed]
  3. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  4. Li, J.; Zhang, Y. Learning surf cascade for fast and accurate object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3468–3475. [Google Scholar]
  5. Dellinger, F.; Delon, J.; Gousseau, Y.; Michel, J.; Tupin, F. SAR-SIFT: A SIFT-like algorithm for SAR images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 453–466. [Google Scholar] [CrossRef] [Green Version]
  6. Zheng, L.; Yang, Y.; Tian, Q. SIFT meets CNN: A decade survey of instance retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1224–1244. [Google Scholar] [CrossRef] [Green Version]
  7. Song, L.; Liu, J.; Qian, B.; Sun, M.; Yang, K.; Sun, M.; Abbas, S. A deep multi-modal CNN for multi-instance multi-label image classification. IEEE Trans. Image Process. 2018, 27, 6025–6038. [Google Scholar] [CrossRef]
  8. Ferrari, C.; Lisanti, G.; Berretti, S.; Del Bimbo, A. Investigating nuisances in DCNN-based face recognition. IEEE Trans. Image Process. 2018, 27, 5638–5651. [Google Scholar] [CrossRef]
  9. Yin, X.; Liu, X. Multi-task convolutional neural network for pose-invariant face recognition. IEEE Trans. Image Process. 2018, 27, 964–975. [Google Scholar] [CrossRef] [Green Version]
  10. Basu, T.; Menzer, O.; Ward, J.; SenGupta, I. A Novel Implementation of Siamese Type Neural Networks in Predicting Rare Fluctuations in Financial Time Series. Risks 2022, 10, 39. [Google Scholar] [CrossRef]
  11. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  12. Rama-Maneiro, E.; Vidal, J.C.; Lama, M. Embedding Graph Convolutional Networks in Recurrent Neural Networks for Predictive Monitoring. arXiv 2021, arXiv:2112.09641. [Google Scholar]
  13. Nassar, M.; Wang, X.; Tumer, E. Fully Convolutional Graph Neural Networks using Bipartite Graph Convolutions. ICLR 2020. [Google Scholar]
  14. Hong, Y.; Liu, Y.; Yang, S.; Zhang, K.; Wen, A.; Hu, J. Improving graph convolutional networks based on relation-aware attention for end-to-end relation extraction. IEEE Access 2020, 8, 51315–51323. [Google Scholar] [CrossRef]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2012; pp. 1097–1105. [Google Scholar]
  16. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  17. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  19. Niu, X.X.; Suen, C.Y. A novel hybrid CNN–SVM classifier for recognizing handwritten digits. Pattern Recognit. 2012, 45, 1318–1325. [Google Scholar] [CrossRef]
  20. Duan, M.; Li, K.; Yang, C.; Li, K. A hybrid deep learning CNN–ELM for age and gender classification. Neurocomputing 2018, 275, 448–461. [Google Scholar] [CrossRef]
  21. Fu, M.y.; Liu, F.y.; Yang, Y.; Wang, M.l. Background pixels mutation detection and Hu invariant moments based traffic signs detection on autonomous vehicles. In Proceedings of the 33rd Chinese Control Conference, Nanjing, China, 28–30 July 2014; pp. 670–674. [Google Scholar]
  22. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  23. Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.; Wortman, J. Learning bounds for domain adaptation. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2008; pp. 129–136. [Google Scholar]
  24. Liu, F.; Lin, G.; Shen, C. CRF learning with CNN features for image segmentation. Pattern Recognit. 2015, 48, 2983–2992. [Google Scholar] [CrossRef] [Green Version]
  25. Xie, G.S.; Zhang, X.Y.; Yan, S.; Liu, C.L. Hybrid CNN and dictionary-based models for scene recognition and domain adaptation. IEEE Trans. Circuits Syst. Video Technol. 2015, 27, 1263–1274. [Google Scholar] [CrossRef] [Green Version]
  26. Nguyen, D.T.; Pham, T.D.; Baek, N.R.; Park, K.R. Combining deep and handcrafted image features for presentation attack detection in face recognition systems using visible-light camera sensors. Sensors 2018, 18, 699. [Google Scholar] [CrossRef] [Green Version]
  27. Guo, Z.; Chen, Q.; Wu, G.; Xu, Y.; Shibasaki, R.; Shao, X. Village building identification based on ensemble convolutional neural networks. Sensors 2017, 17, 2487. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images. Master’s Thesis, University of Toronto, Toronto, ON, Canada, 2009. [Google Scholar]
  29. Pei, J.; Fan, H.; Pu, L. Range space super spherical cap discriminant analysis. Neurocomputing 2017, 244, 112–122. [Google Scholar] [CrossRef] [Green Version]
  30. Wen, H.; Xie, W.; Pei, J.; Guan, L. An incremental learning algorithm for the hybrid RBF-BP network classifier. EURASIP J. Adv. Signal Process. 2016, 2016, 57. [Google Scholar] [CrossRef]
Figure 1. The experimental structure diagram of generalization ability of feature mapper.
Figure 1. The experimental structure diagram of generalization ability of feature mapper.
Applsci 12 11301 g001
Figure 2. The accuracy increase of different t unknown subclasses compared with the original accuracy.
Figure 2. The accuracy increase of different t unknown subclasses compared with the original accuracy.
Applsci 12 11301 g002
Figure 3. The accuracy increase of different t unknown subclasses compared with the original accuracy.
Figure 3. The accuracy increase of different t unknown subclasses compared with the original accuracy.
Applsci 12 11301 g003
Figure 4. The overall structure of the SOPNN classifier.
Figure 4. The overall structure of the SOPNN classifier.
Applsci 12 11301 g004
Figure 5. The network structure of GHP-CNN.
Figure 5. The network structure of GHP-CNN.
Applsci 12 11301 g005
Figure 6. Effect of parameter A on classification accuracy of GHP-CNN.
Figure 6. Effect of parameter A on classification accuracy of GHP-CNN.
Applsci 12 11301 g006
Figure 7. Effect of parameter σ on classification accuracy of GHP-CNN.
Figure 7. Effect of parameter σ on classification accuracy of GHP-CNN.
Applsci 12 11301 g007
Table 1. Important notations and variables used in this paper.
Table 1. Important notations and variables used in this paper.
S * The source dataset for training the CNN
sThe number of classes in the source dataset S *
T * The target dataset for classification
tThe number of classes in the target dataset T *
P s The original training accuracy of VGG-16 in the source dataset with s class samples
P t s The test accuracy of target sets with t class samples
P ˜ t s The average accuracy of multiple P t s
CFM Convolutional feature mapper
CFM s Convolutional feature mapper trained by the source dataset with s class samples
T * S * The samples in T * are selected from S *
S * T * = The number of common samples between T * and S * is zero
r t s The increased amplitude of the t class target dataset compared with the s class source dataset
t r ( A ) The trace of matrix A
d e t ( A ) The determinant of matrix A
Table 2. Classification accuracy corresponding to different algorithms after category reduction with the CIFAR-100 dataset.
Table 2. Classification accuracy corresponding to different algorithms after category reduction with the CIFAR-100 dataset.
s = 100 t = 4 t = 6 t = 8 t = 12 t = 16
VGG-16-E90.0087.5085.8682.5881.01
CNN-PNN94.090.6789.0085.2582.63
CNN-SVM94.7589.3388.7583.6783.00
CNN-ELM95.0090.8389.5085.7583.01
GHP-CNN98.5096.5094.3892.7591.13
Table 3. Classification accuracy corresponding to different algorithms after category reduction with the Caltech-256 dataset.
Table 3. Classification accuracy corresponding to different algorithms after category reduction with the Caltech-256 dataset.
s = 256 t = 4 t = 6 t = 8 t = 12 t = 16
ResNet-152-E89.6087.3080.4178.5475.81
CNN-PNN89.0382.3579.8475.6172.34
CNN-SVM90.0083.5380.7279.1877.50
CNN-ELM82.5073.3772.8371.6270.63
GHP-CNN95.0088.5185.3781.4278.72
Table 4. Classification accuracy corresponding to different algorithms after the unknown categories with the CIFAR-100 dataset.
Table 4. Classification accuracy corresponding to different algorithms after the unknown categories with the CIFAR-100 dataset.
s = 90 t = 3 t = 4 t = 5 t = 6 t = 8 t = 10
VGG-16-E90.0085.7577.2076.0075.8874.30
CNN-PNN88.3386.0078.0074.3374.2573.10
CNN-SVM91.3385.5078.2075.7074.6371.00
CNN-ELM92.6788.3079.0076.3075.3773.10
GHP-CNN95.0089.2580.6077.3376.3375.50
Table 5. Classification accuracy corresponding to different algorithms after the unknown categories with the Caltech-256 dataset.
Table 5. Classification accuracy corresponding to different algorithms after the unknown categories with the Caltech-256 dataset.
s = 246 t = 3 t = 4 t = 5 t = 6 t = 8 t = 10
ResNet-152-E90.1586.5184.8381.3279.4776.14
CNN-PNN90.6286.4784.4380.1577.8175.40
CNN-SVM89.4087.7485.4281.3678.6277.48
CNN-ELM84.4281.3075.3172.4571.1770.09
GHP-CNN92.3291.2587.9185.4182.7480.73
Table 6. Comparison of test time for each algorithm.
Table 6. Comparison of test time for each algorithm.
s = 90 t = 3 t = 4 t = 5 t = 6 t = 8 t = 10
VGG-16-E19.1528.3632.4241.5156.8471.91
CNN-PNN19.2128.3132.9341.3357.7272.64
CNN-SVM19.0128.0232.0241.0356.0371.04
CNN-ELM19.0028.0032.0041.0056.0471.08
GHP-CNN19.2528.2232.6941.2857.5872.45
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhou, W.; Fan, H.; Zhu, J.; Wen, H.; Xie, Y. Research on Generalized Hybrid Probability Convolutional Neural Network. Appl. Sci. 2022, 12, 11301. https://doi.org/10.3390/app122111301

AMA Style

Zhou W, Fan H, Zhu J, Wen H, Xie Y. Research on Generalized Hybrid Probability Convolutional Neural Network. Applied Sciences. 2022; 12(21):11301. https://doi.org/10.3390/app122111301

Chicago/Turabian Style

Zhou, Wenyi, Hongguang Fan, Jihong Zhu, Hui Wen, and Ying Xie. 2022. "Research on Generalized Hybrid Probability Convolutional Neural Network" Applied Sciences 12, no. 21: 11301. https://doi.org/10.3390/app122111301

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop