Research article

Relative mispricing and takeover likelihood

  • Received: 30 July 2021 Accepted: 12 December 2021 Published: 17 December 2021
  • JEL Codes: G1; G11; G34

  • This paper examines the effect of acquirer likelihood on future stock returns. In sharp contrast to prior findings, acquirer likelihood is a strong and negative predictor of cross-sectional future returns after controlling for target likelihood. If takeover exposure represents a risk premium, the effect on stock valuation should only present in either likelihood measure (acquirer or target likelihood). This evidence casts doubt on the rational risk explanation, but is consistent with a relative mispricing story. Investors take positions accordingly to explore profits from takeovers. Profits from trading strategy based on takeover probability are concentrated in stocks with high misvaluation characteristics, including small size, value, high momentum, high investment, and low turnover firms, as well as both high and low issuance (or accrual) firms.

    Citation: Keming Li. Relative mispricing and takeover likelihood[J]. Quantitative Finance and Economics, 2021, 5(4): 689-715. doi: 10.3934/QFE.2021031

    Related Papers:

    [1] Yufeng Li, Chengcheng Liu, Weiping Zhao, Yufeng Huang . Multi-spectral remote sensing images feature coverage classification based on improved convolutional neural network. Mathematical Biosciences and Engineering, 2020, 17(5): 4443-4456. doi: 10.3934/mbe.2020245
    [2] Haifeng Song, Weiwei Yang, Songsong Dai, Haiyan Yuan . Multi-source remote sensing image classification based on two-channel densely connected convolutional networks. Mathematical Biosciences and Engineering, 2020, 17(6): 7353-7377. doi: 10.3934/mbe.2020376
    [3] Hua Wang, Weiwei Li, Wei Huang, Jiqiang Niu, Ke Nie . Research on land use classification of hyperspectral images based on multiscale superpixels. Mathematical Biosciences and Engineering, 2020, 17(5): 5099-5119. doi: 10.3934/mbe.2020275
    [4] Eric Ke Wang, Nie Zhe, Yueping Li, Zuodong Liang, Xun Zhang, Juntao Yu, Yunming Ye . A sparse deep learning model for privacy attack on remote sensing images. Mathematical Biosciences and Engineering, 2019, 16(3): 1300-1312. doi: 10.3934/mbe.2019063
    [5] Haifeng Song, Weiwei Yang, Songsong Dai, Lei Du, Yongchen Sun . Using dual-channel CNN to classify hyperspectral image based on spatial-spectral information. Mathematical Biosciences and Engineering, 2020, 17(4): 3450-3477. doi: 10.3934/mbe.2020195
    [6] Qian Zhang, Haigang Li, Ming Li, Lei Ding . Feature extraction of face image based on LBP and 2-D Gabor wavelet transform. Mathematical Biosciences and Engineering, 2020, 17(2): 1578-1592. doi: 10.3934/mbe.2020082
    [7] Eric Ke Wang, Fan Wang, Ruipei Sun, Xi Liu . A new privacy attack network for remote sensing images classification with small training samples. Mathematical Biosciences and Engineering, 2019, 16(5): 4456-4476. doi: 10.3934/mbe.2019222
    [8] Qihan Feng, Xinzheng Xu, Zhixiao Wang . Deep learning-based small object detection: A survey. Mathematical Biosciences and Engineering, 2023, 20(4): 6551-6590. doi: 10.3934/mbe.2023282
    [9] Qing Yang, Jun Chen, Najla Al-Nabhan . Data representation using robust nonnegative matrix factorization for edge computing. Mathematical Biosciences and Engineering, 2022, 19(2): 2147-2178. doi: 10.3934/mbe.2022100
    [10] Lei Yang, Guowu Yuan, Hao Wu, Wenhua Qian . An ultra-lightweight detector with high accuracy and speed for aerial images. Mathematical Biosciences and Engineering, 2023, 20(8): 13947-13973. doi: 10.3934/mbe.2023621
  • This paper examines the effect of acquirer likelihood on future stock returns. In sharp contrast to prior findings, acquirer likelihood is a strong and negative predictor of cross-sectional future returns after controlling for target likelihood. If takeover exposure represents a risk premium, the effect on stock valuation should only present in either likelihood measure (acquirer or target likelihood). This evidence casts doubt on the rational risk explanation, but is consistent with a relative mispricing story. Investors take positions accordingly to explore profits from takeovers. Profits from trading strategy based on takeover probability are concentrated in stocks with high misvaluation characteristics, including small size, value, high momentum, high investment, and low turnover firms, as well as both high and low issuance (or accrual) firms.



    Hyperspectral Image (HSI) is the simultaneous imaging of target areas in dozens to hundreds of continuous spectral bands [1,2,3,4]. It effectively integrates the spatial and spectral information in the imaging scene, with strong target detection ability and better material identification ability [5,6,7,8]. It is widely used in agriculture and forestry, geological exploration, marine exploration, Environmental monitoring and other fields [9,10,11,12]. However, HSI is characterized by high data dimension, large information redundancy and high correlation between bands, which brings great difficulties to its processing and classification [13,14,15,16,17,18,19,20,21,22,23]. Therefore, how to reduce the redundant information of the data, extract the features of the hyperspectral images effectively, and realize the accurate classification of the hyperspectral images are the hot and difficult issues in the current hyperspectral image processing and classification research.

    Sample labeling of hyperspectral images often requires expert knowledge and experience, so the cost of sample labeling is high [24,25,26]. When the labeled samples are limited, semi-supervised learning can explore the useful information of the unlabeled samples to participate in the model training and reduce the labeling cost [27,28]. In the field of machine learning, semi-supervised learning acquires knowledge and experience from a small number of labeled samples. The usable information is mined from a large number of unlabeled samples, which can improve the classification accuracy [29,30,31]. Therefore, a large number of scholars have studied the semi-supervised learning for remote sensing images. Camps-Valls et al. [32] proposed a graph-based hyperspectral image classification method, and constructed the graph structure through the graph method. The data context information based on the composite kernel is integrated, and the Nystrom method is used to speed up classification. Yang et al. [33] proposed a semi-supervised band selection technique for hyperspectral image classification. A metric learning method has been used to measure the features of hyperspectral images, and a semi-supervised learning method has been used to select a subset of valid bands from the original bands. Tan et al. [34] proposed a hyperspectral image classification method based on segmentation integration and semi-supervised support vector machine. The spatial information of the tag samples is extracted by using a segmentation algorithm to filter the samples, and classified based on semi-supervised learning. Samiappan et al. [35] combined active learning and co-training to perform semi-supervised classification of hyperspectral images. The initial classification model is trained according to the labeled samples, and the heuristic active learning is performed on the unlabeled samples. Combined with the original data, the labeled samples are divided into views, and the unlabeled samples with high heuristic values are selected to join the training samples. Zhang et al. [36] used a semi-supervised classification method based on hierarchical segmentation and active learning to extract spatial information from hyperspectral images, the training set is updated iteratively by using the information of a large number of unlabeled samples to complete the hyperspectral image classification. In addition, some other methods are also proposed in recent years [37,38,39,40,41,42,43,44,45].

    In hyperspectral images, each pixel corresponds to a spectral curve that reflects its inherent physical, chemical and optical properties. The main basis of hyperspectral image classification is to use the feature information of different pixels to label the pixels belonging to different landmarks and obtain the corresponding classification maps [46,47]. Therefore, a large number of scholars have studied hyperspectral image classification. Melgani et al. [48] proposed a hyperspectral image classification method based on support vector machines (SVM). The kernel function is introduced to solve the nonlinear separable problem and avoid the curse of dimensionality. Ratle et al. [49] introduced neural networks into hyperspectral image classification. In the training phase, the loss function is optimized to avoid local optimization. Chen et al. [50] constructed a hyperspectral image classification model based on sparse representation, and compared the classification results of common machine learning methods. In order to improve the shortcomings of sparse representation in dealing with nonlinear problems, Chen et al. [51] proposed a kernel sparse representation technique. In addition, Cui et al. [52] proposed a multiscale sparse representation algorithm for robust hyperspectral image classification. Automatic and adaptive weight allocation schemes based on spectral angle ratio are incorporated into the multi-classifier framework to fuse sparse representation information at all scales. Tang et al. [53] proposed two sparse representation algorithms based on manifolds to solve the instability problem of sparse algorithms. The regularization and local invariance techniques are used, and two manifold-based regularization items are merged into the l1-based objective function. Wang et al. [54] applied the neighborhood-cutting technique to sparse representation, and combined the joint spatial and spectral sparse representation classification method. Wang and Celik [55] improved the classification accuracy of hyperspectral images by combining context information in the sparse coefficient domain. Hu et al. [56] proposed two weighted kernel joint sparse representation methods, which determine the calculation weight by calculating the kernel similarity between adjacent pixels. Xue et al. [57] presented two novel sparse graph regularization methods, SGR and SGR with total variation. Yang et al. [58] studied the effect of the p-norm distance metric on the minimum distance technique, and proposed a supervised-learning p-norm distance metric to optimize the value of p. Zhang et al. [59] proposed a multi-scale dense network for HSI classification that made full use of different scale information in the network structure and combined scale information throughout the network. Liu et al. [60] proposed a class-wise adversarial adaptation method in conjunction with the class-wise probability MMD as the class-wise distribution adaptation network. Wang et al. [61] proposed the graph-based semi-supervised learning with weighted features for HSI classification. In addition, some new other methods are also proposed by a lot of researchers [62,63,64,65,66,67,68,69,70,71].

    To sum up, hyperspectral images contain rich spectral and spatial information of earth surface features, which increases the processing and analysis difficulties. In addition, the training samples of actual hyperspectral images are small and there is a sample labeling problem. The local binary pattern, sparse representation and mixed logistic regression model are used in this paper. A new hyperspectral image feature extraction method based on local binary pattern is proposed to obtain texture features of hyperspectral images and enrich hyperspectral image sample information. A sample selection strategy based on active learning is designed to determine the unlabeled samples. Based on these, a new sample labeling method based on neighborhood information and priority classifier discrimination is deeply studied to expand the training samples. And a novel classification method based on texture features and semi-supervised learning is studied to improve the classification accuracy of remote sensing images.

    The main contributions of this paper are described as follows.

    (ⅰ) A novel a classification method of hyperspectral remote sensing images based on texture features and semi-supervised learning is proposed by using local binary pattern, sparse representation, hybrid logistic regression.

    (ⅱ) The local binary pattern is used to effectively extract the features of spatial texture information of remote sensing images and enrich the feature information of samples.

    (ⅲ) A multiple logistic regression model is used to optimally select unlabeled samples, which are labeled by using neighbourhood information and priority classifier discrimination to achieve pseudo-labeling of unlabeled samples.

    (ⅳ) A novel classification method of hyperspectral remote sensing images based on semi-supervised learning is proposed to effectively achieve accurate classification of hyperspectral images.

    In here, the local binary pattern (LBP) and sparse expression are introduced in order to clearly describe the basic theories of these methods.

    LBP is a feature extraction method, which can extract spatial texture information of images. Texture, which is widely used in image processing and image analysis, represents the slow change or periodic change of the surface structure of the object [72]. LBP is also widely used in feature extraction of hyperspectral images due to the simple structure and easy calculation. Give the center pixel gc(xc,yc) and the neighborhood pixel gp, which are described as follow.

    gp=(xc+Rcos(2ΠpP),ycRsin(2ΠpP)) (1)

    where, gp(p=0,1,...,P1) represents the coordinate values of P pixels uniformly distributed on the circular domain with gc as the centre and R as the radius. The quantized texture feature form of one region is shown in Figure 1.

    LBPgc=2p×P1p=0s(gpgc) (2)
    s(x)={1,0,x>0x0 (3)
    Figure 1.  The quantized texture feature form of one region.

    Sparse representation means that the signal can be approximately represented by using linear combination of the atoms in the dictionary. Now, X=[X1,X2,...,Xc]RD is given as the HSI pixel and D is the number of image bands. In here, Xi=[xi1,xi2,...,xiNc]RD, Nc represents the number of samples in the class i.

    For samples in the class i, it can be approximated as follow.

    yxi1α1+xi2α2+...+xiNcαNc = [xi1,xi2,...,xiNc][α1,α2,...,αNc]T=Xiαi (4)

    where, Xi represents a sparse sub-dictionary of the samples in the class i. αi represents the sparse vector of test samples y, which contains only a few non-zero values.

    In order to obtain the sparsest vector αi, the following formula is solved.

    ˜α=argminαi0,s.t.y=Aαi (5)

    where,  0 is a l0 norm, which represents the number of non-zero atoms in the vector, also known as sparsity. A is a sparse dictionary. It is a NP-hard problem to solve the Eq (5) directly. Under some conditions, the minimization solving problem (l0) is approximated by the minimization solving problem (l1), which can be relaxed.

    ˜α=argminαi1,s.t.y=Aαi (6)

    Furthermore, the solution can be converted to the following expression.

    ˜α=argminαi0,s.t.Aαiy22<ε (7)

    where, ε represents the refactoring error. Orthogonal matching pursuit (OMP) algorithm is used to solve the expression (7). After the sparsity coefficient is calculated, the reconstruction residual for each class of the test samples y can be calculated.

    ri(y)=yA˜αi2 (8)

    where, there is i{1,2,...,C}.

    Finally, the reconstruction residuals of all dictionaries are compared. The minimum residual is the classification y.

    class(y)=argmin(ri(y)),i=1,2,...C (9)

    In hyperspectral remote sensing image classification, the attribute characteristics of each ground feature are subordinated to a specific distribution. If the number of training samples does not reach the amount that depicts the distribution of the ground feature, it will affect the subsequent classification results. In practical applications, there are difficulties, such as fewer labeled samples, difficulty in manual labeling, and time-consuming and labor-intensive problems and so on. Therefore, a new sample labeling method based on neighborhood information and priority classifier discrimination is proposed to obtain the learned pseudo labeled samples in order to expand the labeled samples and improve the classification accuracy of model.

    Before the samples are labeled, sample selection is required. This is because if all unlabeled samples are labeled directly, all unlabeled samples need to be labeled, which will cost a lot of computation time. Moreover, due to the small number of initial labeled samples and the limited available information, it is difficult to label some samples with a certain accuracy. The mislabeled samples obviously affect the classification accuracy of model. The main objective of the sample selection strategy is to select the unlabeled samples with the largest amount of information. These unlabeled samples can construct a valuable training set after labeling, which can effectively promote the improvement of classification results. Therefore, a sample selection method based on multiple logistic regression is proposed to realize the selection of samples.

    The classified probability matrix p(yki|xi) of each sample by using multiple logistic regression model has a large amount of information that can be mined. The multiple logistic regression classifier is modeled by discriminant Bayesian decision model. According to the generalized linear model, it can be obtained as follow.

    P(y;δ)=b(y)exp(δTT(y)a(δ)) (10)

    The specific expression of multiple logistic regression is described as follow.

    p(yi=kxi,η)=exp(ηkg(xi))/Nk=1exp(ηkg(xi)) (11)

    where, g(x)=[g1(x),g2(x),,gf(x)]T is the feature vectors of the input, and η=[ηT1,ηT2,ηTk] represents the regression parameter vector of the classifier. It is worth noting that the feature vector is often represented by introducing the kernel, which is not only used to improve the indivisibility, but also obtain the better classifier by training samples. Generally, the kernel function is radial basis function (RBF), which is described as follow.

    K(xm,xn)=e||xmxn||22p2 (12)

    After the feature vector is determined, the regression parameter η of model is only determined, and the probability matrix p(yki|xi) of each unlabeled sample belonging to each class is determined. The sample amount of information is determined by the Breaking Ties (BT) and the Least Confidence (LC). In this paper, the BT is selected to determine the amount of information.

    The BT shows the similarity between the two classifications by comparing the difference between the maximum category probability and the sub-maximum category probability. The difference is smaller, the similarity between the two types of samples is greater. The uncertainty is greater, the amount of information is greater. Si is the similarity between classifications, which is described as follow.

    Si=maxp(yki|xi)secondmax(p(yki|xi)) (13)

    The Si is finally sorted in ascending order. p(yki|xi) is the obtained probability matrix of each sample by multiple logistic regression classifier. maxp(yki|xi) represents the largest value in the obtained probability matrix, secondmax(p(yki|xi)) represents the second largest value in the obtained probability matrix. k is the number of regression parameters.

    The features of hyperspectral remote sensing images have some correlation. The ground objects are closer, the correlation is stronger. In the research of sample labeling, spatial neighborhood information based on training samples is widely used. However, due to the unknown central pixel and the lack of sufficient information, the neighborhood information of unlabeled samples is relatively less in the research of sample labeling. Generally, each pixel label on hyperspectral image must be consistent with one pixel label in its neighborhood. This property can be used to label the unlabeled samples. The label information of training samples around the unlabeled samples can be used to discriminate the unlabeled samples. The labeling discrimination method based on neighborhood information centers on the sample to be labeled. The unlabeled samples are labeled with a block diagram. All the occurrences of sample labels are recorded and denoted as the neighborhood information. The labeled samples are used to train the classifier and classify the unlabeled samples. Determine whether the predicted sample label by the classifier appears in the neighborhood information of the unlabeled samples. If it appears, the predicted label by the classifier is the sample label. Otherwise, the samples are put into the unlabeled samples. One of the most important problems is whether the unlabeled samples which satisfy the neighborhood information can be reliably labeled by the classifier. At present, some studies use multiple classifiers to discriminate together and achieve good classification effect. However, how to determine the determination of labels, when the predicted labels by multiple classifiers are inconsistent, but all appear in the neighborhood information of unlabeled samples.

    Therefore, a sample labeling method based on neighborhood information and priority classifier discrimination is proposed in this paper. For unlabeled samples with the neighborhood information, the classifier with the highest priority is used for prediction. If the obtained prediction marker appears in the neighborhood information, its marker is determined. Otherwise, the classifier with the lowest priority is used for prediction. Determine whether the label can be determined until the sample labeling is ended. The sample labeling process based on neighborhood information and priority classifier discrimination is shown in Figure 2.

    Figure 2.  The sample labeling process based on neighborhood information and priority classifier discrimination.

    The sample labeling method is an iterative process. Although it cannot ensure enough training samples around all unlabeled samples at the initial stage of sample labeling, it can ensure that some unlabeled samples are sufficient. The unlabeled samples are labeled and extended to the training set. With each iteration, the training set grows. Those unlabeled samples whose neighborhood training samples are not sufficient may reach the label condition at a certain labeling. The sample labeling method with replacement ensures the sample labeling accuracy to a certain extent, and improves the performance of classifier.

    Hyperspectral remote sensing images consist of pairs of continuous spectral bands, which contain rich spectral and spatial information of earth surface features. Some objects that cannot be identified by conventional remote sensing means can be identified in hyperspectral images. However, the abundant data information increases the processing and analysis difficulties, and there are some problems of fewer labeled samples, difficulty in manual labeling and time-consuming and so on. In order to improve the classification accuracy of hyperspectral remote sensing images, a new classification method of hyperspectral remote sensing images based on texture features and semi-supervised learning is proposed in this paper. Firstly, aiming at the problems of high correlation between bands, information redundancy, high dimension and complex processing, LBP is employed to deal with the hyperspectral images. The texture features of hyperspectral images are effectively extracted to enrich the feature information. To solve the problem of limited label samples, a new sample labeling method based on neighborhood information and priority classifier discrimination is proposed in here. Secondly, a sample selection strategy is designed to find some samples from a large number of unlabeled samples. Finally, the sparse representation and mixed logistic regression are applied to achieve a new classification method to effectively achieve accurate classification of hyperspectral images.

    The model of hyperspectral image classification method based on texture features and semi-supervised learning is shown in Figure 3.

    Figure 3.  The flow of hyperspectral image classification method based on texture features and semi-supervised learning.

    The detailed implementation steps of hyperspectral image classification method based on texture features and semi-supervised learning are described as follows.

    Step 1. Normalize the hyperspectral images. Principal component analysis is used to perform dimensionality reduction.

    Step 2. Calculate the texture for each principal component by using LBP. The histogram statistics are performed according to the symmetric rotation invariant equivalence mode. D-dimensional feature data for each layer is used to obtain feature information for each pixel and retain pixel coordinates.

    Step 3. The original hyperspectral images are performed by linear discriminant analysis to obtain a projection matrix.

    Step 4. The hyperspectral images are mapped into low dimensional images for suitable Euclidean distance analysis by using a projection matrix.

    Step 5. Training sample sets and test sample sets are randomly selected from each type of sample according to a certain proportion of low-dimensional images.

    Step 6. Calculate the Euclidean distance from all training samples by using Lk=yiAk2 to maintain K training coordinates that are close to the Euclidean distance of each test sample as the sparse dictionary for the test samples.

    Step 7. The extracted feature information is loaded into the test samples and corresponding sparse dictionary according to the corresponding coordinates to obtain the test sample set and sparse dictionary.

    Step 8. The sparse representation is solved. Orthogonal matching pursuit (OMP) has been used to calculate the sparse coefficient, and the classifications of the test samples is obtained.

    1) Indian Pines data

    The images of Indian pines in northwest Indiana were collected by AVIRIS sensor. The images consist of 145 × 145 pixels and 224 spectral reflection bands with a wavelength range of 0.4~2.5 nm, including 16 types of feature elements. The false color map and real ground object distribution are shown in Figure 4.

    Figure 4.  Hyperspectral remote sensing images of Indian Pines.

    2) Salinas Scene data

    The AVIRIS spectrometer collects images of the Salinas Valley in California, USA, with a size of 512 × 217 pixels and a total of 224 bands. After the bands covering the water absorption area are removed, 204 bands are used, including 16 types of ground feature elements. The false color map and the real ground object distribution are shown in Figure 5.

    Figure 5.  Hyperspectral remote sensing image of Salinas Scene.

    3) Pavia University data

    Images of the Italian University of Pavia campus taken by the Rosis Spectrometer. It is 610 × 340 pixels in size and has a total of 115 wavebands. The 103 wavebands after the wavebands covering the water-absorbing region contain a total of 9 types of features are removed. The false color map and the real ground object distribution are shown in Figure 6.

    Figure 6.  Hyperspectral remote sensing image of Pavia University.

    In the experiment, 10% of each type of ground object of the three kinds of data is randomly selected as the training samples, and the rest is the test samples.

    Confusion Matrix (CM) is usually used in the classification and evaluation of hyperspectral images. A confusion matrix is generally defined as follow.

    P=[p11p21p12p22p1np2npn1pn2pnn] (14)

    where, n denotes the number of objects in the category, pij represents the number of samples belonging to the class i that are assigned to the class j. The total amount of data in each row denotes the true number of objects in the category. The total amount of data in each column represents the total number of samples.

    Based on the confusion matrix, three classification indexes can be obtained, which are Overall Accuracy (OA), Average Accuracy (AA) and Kappa coefficient.

    OA=ni=1piiN (15)

    where, N represents the total number of samples in the classification. pii represents the number of correctly classified samples of the class i. OA represents the probability that the classified result corresponds to its true label for each random sample.

    CAi=piiNi (16)

    where, Ni represents the total number of samples for the first category in the class i. CAi represents the probability that category i is correctly classified.

    Kappa(n(Ni=1pii)Ni=1(Nj=1pijNj=1pji))n2Ni=1(Nj=1pijNj=1pji) (17)

    The Kappa coefficient comprehensively considers the number of correctly classified objects and the error of misclassified objects on the confusion matrix.

    In the experiment, a large number of alternative values are tested, and some classical values are selected from literatures, these parameter values are experimentally modified until the most reasonable parameter values are determined. These selected parameter values have obtained the optimal solution, so that they can accurately and efficiently verify the effectiveness.

    The quality of sample selection directly affects the efficiency of the experiment, and also affects the performance of classifier. In order to select the best samples, the experimental results of Information Entropy (IE), Min Error (ME), Breaking Ties (BT) and Least Confidence (LC) on three kinds of hyperspectral images are compared. The experiment is 10 initial samples for each class, and all remaining samples are test samples. Two hundred unlabeled samples are selected by four sample selection methods in each iteration. The label samples are used to expand the training samples, train the classifier and classify the test samples. The quality of the selection samples is determined according to the classification results after each iteration is executed. The classification accuracy of different sample selection methods on different data sets is shown in Table 1.

    Table 1.  The classification accuracy of different methods on different data sets (%).
    Data Selection method 1 2 3 4 5 6 7 8 9 10
    Indian Pines IE 78.39 79.10 79.99 80.71 82.11 83.47 84.10 85.51 86.48 87.23
    ME 80.33 86.83 91.08 92.92 95.28 97.02 98.00 98.45 98.90 99.13
    BT 91.47 95.47 98.22 98.36 98.64 98.59 98.66 98.71 99.34 99.29
    LC 84.74 88.95 91.63 94.33 95.27 96.46 98.15 98.55 98.62 98.66
    Pavia University IE 69.87 71.31 71.75 71.74 72.04 72.40 73.25 73.36 73.76 74.45
    ME 73.01 75.23 78.74 82.78 89.14 92.82 95.14 96.10 97.13 97.82
    BT 87.54 92.63 94.51 95.24 95.84 96.10 96.39 96.50 96.69 96.71
    LC 74.91 76.33 80.76 84.45 87.23 89.89 90.27 90.67 90.56 91.47
    Salinas Scene IE 84.16 84.60 84.65 85.02 85.09 85.25 85.50 85.65 85.91 85.98
    ME 85.14 89.29 92.81 94.31 96.48 97.41 98.04 98.20 98.66 98.88
    BT 95.30 96.98 98.26 98.71 98.95 98.90 99.03 99.21 99.25 99.24
    LC 88.60 91.19 92.84 93.58 93.80 95.74 97.56 97.72 98.56 98.86

     | Show Table
    DownLoad: CSV

    From Table 1, it can be seen that the classification accuracies of ME, BT and LC are greatly improved on the three data sets with the increase of the number of iterations. The results of the BT method are significantly better than those of the other methods. The accuracy can be improved in the first few iterations, which indicate that the BT method can select samples with greater classification improvement. Therefore, the BT method is chosen as the sample selection method in this paper.

    In the labeling process, the samples are not completely labeled correctly. The more samples are screened, the more samples may be misclassified. This will make the training set noisier and affects the generalization ability of the classifier. If the number of samples is too small, the number of labeled samples will not improve the classification accuracy of the classifier or will reduce the classification efficiency. The results of classification accuracy under different sample sizes are shown in Table 2.

    Table 2.  The classification accuracy results under different sample sizes (%).
    Data Quantity 1 2 3 4 5 6 7 8 9 10
    Indian Pines 200 77.30 77.57 78.32 77.97 78.59 78.97 78.96 79.28 79.51 79.23
    400 77.54 78.60 79.80 80.80 82.13 82.20 83.03 83.77 83.83 83.85
    600 77.52 79.54 79.37 79.27 80.08 81.99 83.89 83.60 84.30 84.48
    800 77.75 79.84 80.87 80.22 82.11 83.91 84.41 84.57 85.12 85.94
    1000 77.85 80.49 80.28 82.74 81.35 82.60 84.18 85.88 86.77 87.84
    1200 77.85 79.95 79.79 80.38 81.59 84.41 84.72 85.74 87.48 88.85
    1400 78.18 80.20 80.09 83.96 85.00 85.78 87.93 89.90 91.07 91.20
    1600 78.55 80.56 80.59 84.12 86.90 87.82 89.02 90.87 91.49 91.83
    1800 78.34 79.76 79.23 82.64 85.39 87.32 88.81 89.97 90.69 91.46
    2000 78.01 79.16 80.24 82.55 86.02 87.48 88.66 89.63 90.02 90.46
    Pavia University 200 68.75 73.93 76.73 78.18 79.23 80.92 81.85 82.40 82.74 83.60
    400 66.41 73.11 75.45 78.20 81.18 82.13 82.57 83.39 84.07 83.98
    600 68.88 76.35 78.22 80.73 82.59 83.29 83.68 84.57 84.95 84.98
    800 69.89 77.50 80.31 81.91 83.22 84.85 84.99 84.79 85.16 85.31
    1000 70.28 76.35 79.92 82.68 83.83 84.48 84.84 85.21 85.37 85.04
    1200 70.24 75.18 80.32 83.13 84.14 85.04 85.30 85.55 85.04 84.90
    1400 70.34 76.23 80.57 82.46 83.92 84.71 85.59 85.77 85.87 86.47
    1600 70.40 75.87 80.64 83.06 83.93 84.69 85.28 86.02 86.19 85.83
    1800 69.77 76.12 80.18 82.99 85.19 85.04 84.89 85.26 85.68 85.68
    2000 69.71 75.90 82.29 83.40 84.41 85.23 85.59 85.77 85.86 85.87
    Salinas Scene 200 85.09 87.26 89.04 89.35 89.24 89.22 88.85 88.47 88.14 88.03
    400 84.94 88.34 89.88 89.26 89.12 88.88 88.26 87.68 87.40 87.25
    600 85.69 90.85 90.80 90.42 89.71 89.04 88.63 87.84 87.12 86.60
    800 85.36 89.17 88.87 87.96 87.46 86.85 86.42 85.69 85.13 84.58
    1000 85.35 88.93 89.88 89.04 88.34 87.58 87.03 85.49 85.08 85.08
    1200 85.04 89.66 90.08 88.98 87.67 87.14 86.13 85.30 85.09 84.85
    1400 85.07 88.46 89.39 88.63 88.10 88.06 87.34 86.81 86.69 86.58
    1600 85.47 89.54 89.88 88.87 87.68 86.21 85.00 84.34 83.89 83.70
    1800 85.50 90.16 90.15 89.45 88.58 87.71 86.79 86.52 86.31 85.93
    2000 85.48 89.40 89.38 88.60 87.73 85.98 85.67 85.12 85.39 85.34

     | Show Table
    DownLoad: CSV

    It can be seen from Table 2 that the sample selection screening quantity in different data sets presents different rules. Indian Pines datasets have the highest classification accuracy after 10 iterations are finished. Pavia University datasets have the highest classification accuracy after 8–10 iterations are finished. Salinas Scene datasets have the highest classification accuracy after 2–4 iterations are finished. The sample screening quantity with the highest accuracy is regarded as the experiment parameter, which are 1600 for Indian Pines, 1400 for Pavia University and 600 for Salinas Scene.

    The size of the block window determines the neighborhood information of the samples, which directly affect the accuracy of the pseudo-tagging method. Due to the different size of data, the optimal block window size is also determined by a large number of experiments. The classification accuracy under different block window sizes is shown in Table 3.

    Table 3.  The classification accuracy under different block window sizes (%).
    Data Block window size 1 2 3 4 5 6 7 8 9 10
    Indian Pines 3 77.62 77.86 78.35 78.65 78.91 79.14 79.09 79.12 79.56 79.56
    4 77.73 78.24 78.65 78.75 78.70 78.66 79.14 79.18 79.70 79.74
    5 78.41 79.61 79.41 81.28 81.11 81.73 82.67 82.84 84.18 84.68
    6 77.85 80.14 80.41 80.91 80.90 83.54 84.91 85.09 86.42 88.30
    7 78.86 80.58 82.58 84.90 86.55 86.99 88.47 88.87 89.58 90.98
    8 78.33 78.95 81.00 84.24 85.44 86.37 87.60 88.13 88.57 89.19
    9 79.61 81.00 83.39 85.51 85.71 86.34 86.78 86.86 87.07 87.86
    10 78.61 79.32 83.45 85.45 85.59 85.69 86.12 87.00 87.17 87.32
    Pavia University 5 68.39 68.38 68.38 68.38 68.38 68.38 68.38 68.38 68.38 68.38
    10 68.20 67.67 69.05 69.06 68.30 68.13 68.06 67.98 67.89 67.62
    15 70.91 70.59 70.96 70.50 71.89 73.54 73.80 74.58 75.25 75.32
    20 71.47 71.63 73.60 76.12 75.99 75.61 77.18 77.51 77.81 78.14
    25 71.25 74.57 78.52 79.15 81.95 82.32 83.66 84.88 85.45 85.51
    30 70.28 76.35 79.92 82.68 83.83 84.48 84.84 85.21 85.37 85.04
    35 71.48 77.43 80.15 81.75 83.18 83.62 84.11 83.65 83.23 83.17
    40 73.40 77.50 80.36 81.92 83.15 82.68 82.23 82.38 82.18 82.04
    Salinas Scene 5 83.94 83.94 83.94 83.94 83.94 83.94 83.94 83.94 83.94 83.94
    10 83.61 84.16 84.95 85.03 84.77 84.68 84.59 84.82 84.61 85.48
    15 83.86 83.17 85.51 87.04 87.70 88.44 89.77 89.60 91.01 91.09
    20 84.16 86.84 89.41 90.56 90.22 90.69 91.02 91.13 90.87 90.68
    25 84.11 88.22 89.46 89.68 89.35 88.77 87.86 87.49 86.71 86.78
    30 85.35 88.93 89.88 89.04 88.34 87.58 87.03 85.49 85.08 85.08
    35 86.17 88.80 88.67 87.67 86.63 85.84 85.26 84.87 83.95 83.06
    40 86.86 89.25 89.02 87.78 86.37 84.74 83.60 82.83 81.69 80.92

     | Show Table
    DownLoad: CSV

    As can be seen from Table 3, there have different changes in classification accuracy for different datasets. Compared with the other two data sets, the size of Indian Pines is the smallest, so its experimental block window side length values are from 3 to 10. With the increase of the number of iterations, the classification accuracy showed a trend of gradual increase, and the optimal accuracy is obtained when the side length is 7. When the side length of the block window for Pavia and Salinas datasets is too small, the classification accuracy will not improve with the increase of iterations. This indicates that the neighborhood information cannot distinguish the category at this time.

    With the increase of the side length of the block window, the number of iterations to achieve the optimal classification accuracy is advanced, but the optimal accuracy decreases. The block window size is larger, the more noise information will be introduced, and it will affect the sample labeling accuracy. Therefore, the block window size of the Indian Pines dataset is 7 × 7, and the block window sizes of the Pavia and Salinas datasets are 25 × 25 and 20 × 20, respectively.

    In fact, the determination of pseudo-tags of samples mainly depends on the determination of classifiers. K-Nearest Neighbor (KNN), Sparse Representation-Based Classifier (SRC), Neighbor Rough Set (NRS) and mixed logistic regression (MLR) are employed to determine the pseudo-tags. The experimental results of single classifier and different combination classifiers on different datasets are shown in Tables 46.

    Table 4.  The experimental results of different classifier combinations in Indian Pines data set.
    Classifier Index 1 2 3 4 5 6 7 8 9 10
    KNN NUM 314 673 1139 1630 2178 2680 3324 4063 4820 5474
    OA (%) 78.69 79.93 81.20 82.05 82.49 84.38 85.70 86.52 87.33 87.77
    SRC NUM 311 648 1068 1525 2032 2606 3248 3899 4668 5522
    OA (%) 78.64 80.42 80.10 81.56 82.87 84.39 85.83 87.04 87.92 88.19
    NRS NUM 315 673 1116 1597 2136 2697 3265 3815 4437 5016
    OA (%) 78.82 81.02 80.96 83.71 84.94 85.57 86.87 88.01 89.25 89.42
    MLR NUM 133 295 580 936 1296 1790 2396 3077 3858 4648
    OA (%) 77.55 79.31 82.21 83.91 83.99 85.75 87.36 87.68 88.27 88.41
    KNN + SRC NUM 317 706 1120 1702 2294 2967 3728 4494 5233 5950
    OA (%) 78.71 80.96 81.99 83.97 84.96 85.82 86.99 87.75 87.93 88.10
    KNN + NRS NUM 317 707 1198 1691 2308 2954 3625 4450 5374 6305
    OA (%) 78.78 80.27 80.73 82.51 83.82 85.45 87.56 89.06 89.95 90.19
    KNN + MLR NUM 318 712 1206 1794 2555 3398 4292 5072 5783 6678
    OA (%) 78.79 80.56 82.57 86.08 87.10 87.87 88.50 88.88 89.18 89.31
    SRC + KNN NUM 317 706 1134 1673 2223 2875 3658 4317 5011 5748
    OA (%) 78.71 81.09 81.51 83.17 84.27 85.84 86.94 87.54 88.26 88.64
    SRC + NRS NUM 318 730 1205 1778 2372 3091 3969 4813 5787 6641
    OA (%) 78.81 80.79 82.37 85.03 86.81 88.45 88.93 89.86 90.49 90.78
    SRC + MLR NUM 315 708 1153 1744 2457 3322 4231 5042 5800 6735
    OA (%) 78.80 80.92 82.16 85.16 86.24 87.61 88.39 88.71 89.07 89.19
    NRS + KNN NUM 317 707 1202 1712 2385 3021 3700 4538 5399 6333
    OA (%) 78.74 80.67 81.33 83.36 84.46 86.35 87.93 89.23 90.03 90.43
    NRS + SRC NUM 318 734 1207 1728 2378 3061 3959 4799 5691 6739
    OA (%) 78.77 81.01 82.56 84.52 86.37 87.43 88.95 90.11 90.61 90.90
    NRS + MLR NUM 318 694 1148 1690 2446 3194 4102 4950 5768 6792
    OA (%) 78.91 81.39 81.62 85.51 87.21 89.63 90.28 90.92 91.28 91.88
    MLR + KNN NUM 318 677 1166 1689 2429 3246 4018 4737 5677 6672
    OA (%) 78.30 81.13 81.85 85.94 87.22 88.28 88.77 88.98 89.20 89.29
    MLR + SRC NUM 315 704 1258 1741 2439 3286 4097 4867 5775 6760
    OA (%) 78.31 81.81 82.55 85.09 86.86 88.30 89.01 89.44 90.21 90.71
    MLR + NRS NUM 318 683 1154 1701 2458 3301 4219 5057 5997 6889
    OA (%) 78.46 81.49 82.11 86.14 87.86 89.70 90.68 91.58 92.15 92.42

     | Show Table
    DownLoad: CSV
    Table 5.  The experimental results of different classifiers for Pavia University data.
    Classifier Index 1 2 3 4 5 6 7 8 9 10
    KNN NUM 236 444 661 965 1210 1466 1812 2254 2762 3251
    OA (%) 71.19 73.36 74.70 75.69 76.10 76.16 77.28 78.48 77.60 77.69
    SRC NUM 225 477 733 1080 1492 2047 2793 3654 4605 5631
    OA (%) 72.21 72.19 76.97 79.12 80.08 81.77 83.09 84.07 84.95 85.27
    NRS NUM 225 438 763 1007 1256 1487 1726 1893 1991 2179
    OA (%) 71.22 73.74 75.14 76.59 76.15 76.47 76.41 76.20 75.68 75.40
    MLR NUM 100 244 396 605 829 1035 1304 1587 1872 2183
    OA (%) 72.36 76.37 77.18 79.04 79.40 81.16 82.21 82.18 83.36 85.85
    KNN + SRC NUM 248 473 787 1118 1509 1996 2552 3111 3975 4837
    OA (%) 71.48 72.82 74.61 76.59 78.38 80.08 80.82 81.62 82.40 83.99
    KNN + NRS NUM 240 491 775 1067 1382 1850 2403 3040 3794 4415
    OA (%) 71.11 73.83 77.44 78.83 79.50 79.58 79.39 80.11 80.74 81.48
    KNN + MLR NUM 244 515 822 1176 1616 2162 2905 3698 4745 5682
    OA (%) 71.37 75.03 77.84 78.08 79.73 79.59 81.30 83.88 84.90 85.08
    SRC + KNN NUM 248 476 795 1215 1725 2298 3055 3967 4977 5913
    OA (%) 71.38 72.54 75.26 78.09 79.19 80.21 81.88 83.31 83.35 83.51
    SRC + NRS NUM 242 511 867 1385 1889 2513 3201 3988 4910 5794
    OA (%) 71.40 74.12 77.56 80.80 82.19 82.55 83.02 83.80 84.34 85.09
    SRC + MLR NUM 236 507 841 1289 1731 2261 3016 3928 4939 6015
    OA (%) 71.52 73.62 76.47 78.45 79.51 81.54 83.10 84.37 84.68 84.66
    NRS + KNN NUM 240 486 803 1119 1541 1992 2557 3190 3939 4611
    OA (%) 71.05 74.37 76.57 77.77 78.06 77.21 76.75 77.35 78.88 79.28
    NRS + SRC NUM 242 501 803 1296 1792 2433 3089 3839 4776 5798
    OA (%) 71.44 74.50 76.72 79.93 81.44 81.90 82.85 83.81 85.00 85.48
    NRS + MLR NUM 234 517 828 1237 1796 2446 3354 4220 5061 5857
    OA (%) 71.49 75.47 77.89 80.86 83.68 84.43 84.93 85.12 85.57 85.46
    MLR + KNN NUM 244 486 746 1170 1658 2300 3205 4208 5336 6514
    OA (%) 71.47 75.30 76.67 78.76 80.35 82.97 85.05 86.05 86.88 87.02
    MLR + SRC NUM 236 504 903 1310 1708 2207 3009 4024 5098 6234
    OA (%) 71.71 74.01 79.30 79.81 80.40 83.03 86.27 86.93 87.97 88.53
    MLR + NRS NUM 234 524 787 1205 1738 2374 3116 3953 4754 5586
    OA (%) 71.63 75.94 79.83 80.66 82.75 84.64 85.98 86.19 86.37 86.87

     | Show Table
    DownLoad: CSV
    Table 6.  The experimental results of different classifiers for Salinas Scene data.
    Classifier Index 1 2 3 4 5 6 7 8 9 10
    KNN NUM 133 251 443 684 963 1304 1672 2047 2425 2857
    OA (%) 83.38 86.46 86.46 86.65 87.26 87.31 87.79 87.92 87.66 87.15
    SRC NUM 144 275 441 666 937 1255 1606 1968 2391 2811
    OA (%) 83.10 83.56 85.11 85.83 85.81 86.63 86.50 86.64 86.95 86.96
    NRS NUM 148 271 423 668 952 1263 1569 1940 2351 2799
    OA (%) 84.04 86.06 87.62 87.63 87.10 87.48 88.44 88.43 88.32 87.92
    MLR NUM 102 177 302 451 621 867 1176 1518 1848 2217
    OA (%) 82.88 85.41 87.11 88.36 88.73 90.68 91.52 92.20 91.70 92.02
    KNN + SRC NUM 146 294 479 694 976 1330 1691 2106 2520 2985
    OA (%) 83.60 84.65 85.38 86.76 87.39 87.39 88.00 88.11 87.53 87.23
    KNN + NRS NUM 150 297 500 761 1108 1466 1891 2354 2834 3316
    OA (%) 83.53 86.43 87.14 87.49 87.64 87.37 88.06 88.03 87.95 88.03
    KNN + MLR NUM 143 285 508 768 1132 1546 2026 2526 3049 3590
    OA (%) 83.38 85.50 86.34 88.35 88.80 90.07 90.24 90.08 89.86 89.79
    SRC + KNN NUM 146 287 472 686 957 1281 1673 2061 2485 2892
    OA (%) 83.25 84.70 85.79 85.88 86.70 86.92 86.89 86.93 86.85 86.90
    SRC + NRS NUM 150 280 493 755 1017 1328 1708 2114 2539 2982
    OA (%) 83.07 85.55 86.28 85.47 86.66 86.97 87.52 87.05 87.19 87.16
    SRC + MLR NUM 150 271 483 720 1108 1499 1958 2451 2969 3487
    OA (%) 83.15 84.27 87.25 87.88 88.88 89.29 89.36 89.85 89.70 89.85
    NRS + KNN NUM 150 298 519 814 1148 1556 2037 2538 3027 3522
    OA (%) 84.01 87.12 87.79 87.30 87.54 88.38 88.09 88.25 87.95 87.88
    NRS + SRC NUM 150 284 488 803 1158 1539 1988 2482 2955 3423
    OA (%) 83.91 87.23 86.98 87.73 88.30 88.97 89.57 89.57 89.60 89.43
    NRS + MLR NUM 153 293 509 762 1104 1509 1940 2441 2934 3452
    OA (%) 83.87 85.82 88.25 89.93 90.40 90.51 90.75 90.48 90.12 89.61
    MLR + KNN NUM 143 299 521 825 1187 1602 2046 2514 2993 3407
    OA (%) 82.80 85.42 87.67 88.79 89.68 90.06 90.54 90.92 91.27 91.32
    MLR + SRC NUM 150 292 521 799 1123 1537 2007 2448 2929 3367
    OA (%) 82.91 85.45 88.89 90.12 90.21 90.93 90.99 91.83 92.08 92.64
    MLR + NRS NUM 153 315 564 841 1197 1605 2064 2561 3060 3500
    OA (%) 82.81 84.87 88.16 89.34 89.69 90.03 90.42 90.52 91.19 91.43

     | Show Table
    DownLoad: CSV

    As can be seen in Table 4, with the increase of iterations, the classification accuracy of each category samples increased gradually. Compared with the results of the single classifier, the SRC has the largest number of samples after 10 iterations are finished, but the classification effect is not the best value. The classifier with the best classification effect is NRS. The number of samples and classification accuracy of the combination classifiers are mostly better than that of single classifier. The experiment results are different for combination classifiers with different priority. The classifier with NRS can achieve more than 90% classification effect after 10 iterations are finished. The number of combinations with MLR is more than 6600 after 10 iterations are finished, and the best combination is MLR + NRS after 10 iterations are finished.

    As can be seen in Table 5, compared with the experimental results by single classifier, the number of labeled samples with SRC is the largest value after 10 iterations are finished, which is higher than the other three methods. However, the MLR obtained best classification results. After 10 iterations are finished, the number of labeled samples of two classifiers is more than that of the single classifier. KNN, SRC and NRS are considered as the first priority classifiers, the classification results of the sample set after 10 iterations are not as good as those of MLR. The combination of MLR as the first priority classifier has better classification effect than single MLR after 10 iterations are finished.

    For Salinas Scene data, the number of labeled samples by MLR after 10 iterations is the smallest result, but the classification accuracy of the labeled samples is the highest result. After 10 iterations are finished, the number of iterations of the two classifiers is also higher than that of the single classifier. However, from the perspective of the performance of labeled samples in classification, MLR+SRC has higher classification accuracy than MLR, which indicates that the addition of classifiers improves the classification accuracy.

    From three experiment results, it can be seen that the method with the largest number of labeled samples does not necessarily achieve the best classification results. The labeled samples are needed to improve the classification accuracy of the classifier, so the obtained labeled samples after 10 iterations are taken as the evaluation criteria. The Indian Pines dataset uses a combination of classifiers MLR + NRS. The Pavia University and Salinas Scene data sets use MLR + SRC.

    Based on the analysis, the settings of the related parameters are shown in Table 7.

    Table 7.  The settings of the related parameters.
    Data set Indian Pines Pavia University Salinas Scene
    Selection policy BT BT BT
    Number of selections 1600 1400 600
    Window size 7 × 7 25 × 25 20 × 20
    Combination of classifiers MLR + NRS MLR + SRC MLR + SRC
    Number of labeled samples 6889 6234 3367

     | Show Table
    DownLoad: CSV

    Firstly, the LBP is used to extract the features of spatial texture information of hyperspectral remote sensing images. Secondly, the sample labeling method based on neighborhood information and priority classifier discrimination is used to obtain the learned pseudo-labeled samples. The SRC classifier is trained with the labeled samples, and the test samples are predicted. The obtained classification results are compared with those of the SRC classifier on the training data, and the classification results of training models with different training dataset are shown in Table 8.

    Table 8.  The classification results of training models with different training dataset.
    Training samples Index Initial samples Labeling samples
    Indian Pines AA 67.93% 84.70%
    OA 77.38% 92.42%
    KAPPA 0.746 0.914
    Pavia University AA 60.53% 81.87%
    OA 69.00% 88.53%
    KAPPA 0.609 0.848
    Salinas Scene AA 82.59% 87.76%
    OA 84.00% 92.64%
    KAPPA 0.823 0.918

     | Show Table
    DownLoad: CSV

    It can be seen from Table 8, for Indian Pines dataset, the classification results of AA, OA and KAPPA are 84.7%, 94.42% and 0.914, respectively. For Pavia University dataset, the classification results of AA, OA and KAPPA are 81.87%, 88.53% and 0.848, respectively. For Salinas Scene dataset, the classification results of AA, OA and KAPPA are 87.76%, 92.64% and 0.918, respectively. Therefore, the proposed classification method obtains higher classification accuracy.

    The classification visualizations of the proposed classification method for the initial samples and labeled samples are shown in Figures 7 and 8.

    Figure 7.  The classification results of the initial samples.
    Figure 8.  The classification results of the labeling samples.

    By comparing with the results of the experiments, it can find that the classification results of the trained classifier with expanded samples on the three datasets are better than those of the trained classifier with initial samples. Moreover, from the classification visualization, it can also see that the obtained classification results by the classifier and the labeled samples is smoother and has fewer discrete points, which indicates that the generalization ability of the classifier is improved by labeling the samples.

    For the processing and analysis difficulties of hyperspectral images, a new sample labeling method based on neighborhood information and priority classifier discrimination is developed to implement a new classification method of hyperspectral remote sensing images based on texture features and semi-supervised learning by introducing LBP, sparse representation and mixed logistic regression. The LBP is employed to extract the texture features of the hyperspectral remote sensing images. The multivariate logistic regression model is used to select the unlabeled samples with the largest amount of information, and the unlabeled samples with neighborhood information and priority classifier tags are selected to obtain the pseudo-labeled samples after learning. The problem of limited labeled samples of hyperspectral images is solved. The Indian Pines dataset, Salinas scene dataset and Pavia University dataset are selected in here. The experiment results show that the block window of Indian Pines dataset is 7 × 7, and the block windows of Pavia University and Salinas scene are 25 × 25 and 20 × 20, respectively. The combination of MLR and SRC can obtain better classification results. The obtained classification results are smoother and have fewer discrete points, which indicate that the generalization ability of the classifier is improved by labeling the samples. For Indian Pines dataset, the classification results of AA, OA and KAPPA are 84.7%, 94.42% and 0.914, respectively. For Pavia University dataset, the classification results of AA, OA and KAPPA are 81.87%, 88.53% and 0.848, respectively. For Salinas Scene dataset, the classification results of AA, OA and KAPPA are 87.76%, 92.64% and 0.918, respectively. Therefore, the proposed classification method obtains higher classification accuracy by comparing with the other methods.

    However, the proposed classification method has the more computing time, so the next step should be more in-depth research to reduce the time complexity.

    This research was funded by the Sichuan Science and Technology Program, grant number 2021YFS0407, 2022YFS0593, 2023YFG0028; the Sichuan Provincial Transfer Payment Program, Chian under Grant R21ZYZF0006; the A Ba Achievements Transformation Program under Grant R21CGZH0001, R22CGZH0006, R22CGZH0007; the Chengdu Science and technology planning project, grant number 2021-YF05-00933-SN, the Research Foundation for Civil Aviation University of China grant number 2020KYQD123.

    The authors declare no conflict of interest.



    [1] Agrawal A, Jaffe JF, Mandelker GN (1992) The Post-Merger Performance of Acquiring Firms: A Re-examination of an Anomaly. J Financ 47: 1605–1621. doi: 10.1111/j.1540-6261.1992.tb04674.x
    [2] Aigner D, Lovell CAK, Schmidt P (1977) Formulation and Estimation of Stochastic Frontier Production Function Models. J Econom 6: 21–37. doi: 10.1016/0304-4076(77)90052-5
    [3] Ambrose BW, Megginson WL (1992) The Role of Asset Structure, Ownership Structure, and Takeover Defenses in Determining Acquisition Likelihood. J Financ Quant Anal 27: 575–89. doi: 10.2307/2331141
    [4] Amihud Y, Lev B, Travlos N (1990) Corporate Control and the Choice of Investment Financing: The Case of Corporate Acquisitions. J Financ 45: 603–616. doi: 10.1111/j.1540-6261.1990.tb03706.x
    [5] Anderson C, Mandelker G (1993) Long Run Return Anomalies and the Book-to-Market Effect: Evidence on Mergers and IPOs.
    [6] Andrade G, Mitchell M, Stafford E (2001) New Evidence and Perspectives on Mergers. J Econ Perspect 15: 103–120. doi: 10.1257/jep.15.2.103
    [7] Banerjee A, Owers WEE (1992) Wealth Reduction in White Knight Bids. Financ Manage 21: 48–57. doi: 10.2307/3666018
    [8] Barber B, Odean T (2008) All that Glitters: The Effect of Attention and News on the Buying Behavior of Individual and Institutional Investors. Rev Financ Stud 21: 785–818. doi: 10.1093/rfs/hhm079
    [9] Barberis N, Shleifer A (2003) Style Investing. J Financ Econ 68: 161–199. doi: 10.1016/S0304-405X(03)00064-3
    [10] Bates TW, Becher DA, Lemmon ML (2008) Board Classification and Managerial Entrenchment: Evidence from the Market for Corporate Control. J Financ Econ 87: 656–677. doi: 10.1016/j.jfineco.2007.03.007
    [11] Battese GE, Corra G (1977) Estimation of a Production Frontier Model: with Application to the Pastoral Zone of Eastern Australia. Aust J Agr Econ 21: 167–179.
    [12] Berkovitch E, Narayanan MP (1993) Motives for Takeovers: An Empirical Investigation. J Financ Quant Anal 28: 347–362. doi: 10.2307/2331418
    [13] Bradley M, Desai A, Kim EH (1988) Synergistic Gains from Corporate Acquisitions and Their Division between the Stockholders of Target and Acquiring Firms. J Financ Econ 21: 3–40. doi: 10.1016/0304-405X(88)90030-X
    [14] Bruner RF (2004) Applied Mergers and Acquisitions. New York: John Wiley & Sons.
    [15] Brunnermeier MK, Julliard C (2008) Money Illusion and Housing Frenzies. Rev Financ Stud 21: 135–180. doi: 10.1093/rfs/hhm043
    [16] Campbell JY, Shiller RJ (1988) Stock Prices, Earnings, and Expected Dividends. J Financ 43: 661–676. doi: 10.1111/j.1540-6261.1988.tb04598.x
    [17] Campbell JY, Vuolteenaho T (2004) Money Illusion and Stock Prices. Am Econ Rev Pap Proc 94: 19–23. doi: 10.1257/0002828041301533
    [18] Campbell JY, Hilscher J, Szilagyi J (2008) In Search of Distress Risk. J Financ 63, 2899–2939. doi: 10.1111/j.1540-6261.2008.01416.x
    [19] Campello M, Graham JR (2013) Do Stock Prices Influence Corporate Decisions? Evidence from the Technology Bubble. J Financ Econ 107: 89–110. doi: 10.1016/j.jfineco.2012.08.002
    [20] Carhart M (1997) On Persistence in Mutual Fund Performance. J Financ 52: 57–82. doi: 10.1111/j.1540-6261.1997.tb03808.x
    [21] Chakravarty S, McConnell JJ (1997) An analysis of prices, bid/ask spreads, and bid and ask depths surrounding Ivan Boesky's illegal trading in Carnation's stock. Financ Manage, 18–34. doi: 10.2307/3666164
    [22] Chakravarty S, McConnell JJ (1999) Does insider trading really move stock prices? J Financ Quant Anal 34: 191–209. doi: 10.2307/2676278
    [23] Chen C, Lung P, Wang A (2009) Stock Market Mispricing: Money Illusion or Resale Option. J Financ Quant Anal 44: 1125–1147. doi: 10.1017/S0022109009990238
    [24] Cohen R, Polk C, Voulteenaho T (2003) The Value Spread. J Financ 58: 609–642. doi: 10.1111/1540-6261.00539
    [25] Cornell B, Sirri ER (1992) The reaction of investors and stock prices to insider trading. J Financ 47: 1031–1059. doi: 10.1111/j.1540-6261.1992.tb04004.x
    [26] Cooper WE, Gulen H, Schill M (2008) Asset Growth and the Cross-Section of Stock Returns. J Financ 63: 1609–1652. doi: 10.1111/j.1540-6261.2008.01370.x
    [27] Corwin SA, Coughenour JF (2008) Limited Attention and the Allocation of Effort in Securities Trading. J Financ 63: 3031–3067. doi: 10.1111/j.1540-6261.2008.01420.x
    [28] Cremers KJM, Nair VB (2005) Governance Mechanisms and Equity Prices. J Financ 60: 2859–2894. doi: 10.1111/j.1540-6261.2005.00819.x
    [29] Cremers KJM, Nair VB, John K (2009) Takeovers and the Cross-Section of Returns. Rev Financ Stud 22: 1409–1445. doi: 10.1093/rfs/hhn032
    [30] Daniel K, Grinblatt M, Titman S, et al. (1997) Measuring Mutual Fund Performance with Characteristic based Benchmarks. J Financ 52: 1035–1058. doi: 10.1111/j.1540-6261.1997.tb02724.x
    [31] Datta S, Iskandar-Datta M, Raman K (2001) Executive Compensation and Corporate Acquisition Decisions. J Financ 56: 2299–2336. doi: 10.1111/0022-1082.00406
    [32] De Bondt WFM, Thaler RH (1985) Does the Stock Market Overreact. J Financ 40: 793–805. doi: 10.1111/j.1540-6261.1985.tb05004.x
    [33] DeLong G (2001) Stockholder Gains from Focusing versus Diversifying Bank Mergers. J Financ Econ 59: 221–252. doi: 10.1016/S0304-405X(00)00086-6
    [34] De Long JB, Shleifer A, Summers LH, et al. (1990) Noise Trader Risk in Financial Markets. J Polit Econ 98: 703–738. doi: 10.1086/261703
    [35] Dennis DK, McConnell JJ (1986) Corporate Mergers and Security Returns. J Financ Econ 16: 143–187. doi: 10.1016/0304-405X(86)90059-0
    [36] Dong M, Hirshleifer D, Richardson S, et al. (2006) Does Investor Misvaluation Drive the Takeover Market. J Financ 61: 725–762. doi: 10.1111/j.1540-6261.2006.00853.x
    [37] Dong M, Hirshleifer D, Teoh SH (2012) Overvalued Equity and Financing Decisions. Rev Financ Stud 25: 3645–3683. doi: 10.1093/rfs/hhs112
    [38] Dlugosz J, Fahlenbrach R, Gompers P, et al. (2006) Large Blocks of Stock: Prevalence, Size, and Measurement. J Corp Financ 12: 594–618. doi: 10.1016/j.jcorpfin.2005.04.002
    [39] Eckbo E, Thorburn K (2000) Gains to Bidder Firms Revisited: Domestic and Foreign Acquisitions in Canada. J Financ Quant Anal 35: 1–25. doi: 10.2307/2676236
    [40] Edmans A, Goladtein WE, Jiang W (2012) The Real Effects of Financial Markets: The Impact of Prices on Takeovers. J Financ 67: 933–971. doi: 10.1111/j.1540-6261.2012.01738.x
    [41] Fama EF, French KR (1992) The Cross Section of Expected Stock Returns. J Financ 47: 427–465. doi: 10.1111/j.1540-6261.1992.tb04398.x
    [42] Fama EF, French KR (1993) Common Risk Factors in the Returns on Stocks and Bonds. J Financ Econ 33: 3–56. doi: 10.1016/0304-405X(93)90023-5
    [43] Fama Eugene, MacBeth J (1973) Risk, Return and Equilibrium: Empirical Tests. J Polit Econ 81: 607–636. doi: 10.1086/260061
    [44] Fishe RP, Robe MA (2004) The impact of illegal insider trading in dealer and specialist markets: evidence from a natural experiment. J Financ Econ 71: 461–488. doi: 10.1016/S0304-405X(03)00188-0
    [45] Franks J, Harris R, Titman S (1991) The Post Merger Share Price Performance of Acquiring Firms. J Financ Econ 29: 81–96. doi: 10.1016/0304-405X(91)90014-B
    [46] Fuller K, Netter J, Stegemoller M (2002) What do Returns to Acquiring Firms Tell Us? Evidence from Firms that Make Many Acquisitions. J Financ 57: 1763–1794. doi: 10.1111/1540-6261.00477
    [47] Gao Y, Oler D (2012) Rumors and pre-announcement trading: why sell target stocks before acquisition announcements? Rev Quant Financ Account 39: 485–508. doi: 10.1007/s11156-011-0262-z
    [48] Gompers P, Ishii J, Metrick A (2003) Corporate Governance and Equity Prices. Q J Econ 118: 107–155. doi: 10.1162/00335530360535162
    [49] Gorton G, Kahl M, Rosen R (2009) Eat or Be Eaten: A Theory of Mergers and Firm Size. J Financ 64: 1291–1344. doi: 10.1111/j.1540-6261.2009.01465.x
    [50] Habib M, Ljungvist A (2005) Firm Value and Managerial Incentives: A Stochastic Frontier Approach. J Bus 78: 2053–2093 doi: 10.1086/497040
    [51] Harrison JM, Kreps D (1978) Speculative Investor Behavior in a Stock Market with Heterogeneous Expectations. Q J Econ 89: 323–336. doi: 10.2307/1884166
    [52] Hasbrouck J (1985) The Characteristics of Takeover Targets: q and Other Measures. J Bank Financ 9: 351–362. doi: 10.1016/0378-4266(85)90038-X
    [53] Healy PM, Palepu KG, Ruback RS (1992) Does Corporate Performance Improve after Mergers. J Financ Econ 31: 135–75. doi: 10.1016/0304-405X(92)90002-F
    [54] Hong H, Stein J (1999) A Unified Theory of Underreaction, Momentum Trading, and Overreaction in Asset Markets. J Financ 54: 2143–2184. doi: 10.1111/0022-1082.00184
    [55] Houston J, James C, Ryngaert M (2001) Where do Merger Gains Come from? Bank Mergers from the Perspectives of Insiders and Outsiders. J Financ Econ 60: 285–331. doi: 10.1016/S0304-405X(01)00046-0
    [56] Huberman G (2001) Familiarity Breeds Investment. Rev Financ Stud 14,659–680. doi: 10.1093/rfs/14.3.659
    [57] Huberman G, Regev T (2001) Contagious Speculation and a Cure for Cancer: A Nonevent that Made Stock Prices Soar. J Financ 56: 387–396. doi: 10.1111/0022-1082.00330
    [58] Jegadeesh N, Titman S (1993) Returns to Buying Winners and Selling Losers: Implications for Stock Market Efficiency. J Financ 48: 65–91. doi: 10.1111/j.1540-6261.1993.tb04702.x
    [59] Jarrell GA, Brickley J, Netter J (1988) The Market for Corporate Control: Empirical Evidence since 1980. J Econ Perspect 2: 49–68. doi: 10.1257/jep.2.1.49
    [60] Jarrell GA, Poulsen AB (1989) The Returns to Acquiring Firms in Tender Offers-Evidence from 3 Decades. Financ Manage 18: 12–19. doi: 10.2307/3665645
    [61] Jensen MC, Ruback SR (1983) The Market for Corporate Control: The Scientific Evidence. J Financ Econ 11: 5–50. doi: 10.1016/0304-405X(83)90004-1
    [62] Jovanovic B, Rousseau PL (2002) The Q-theory of Mergers. Am Econ Rev 92: 198–204. doi: 10.1257/000282802320189249
    [63] Kaplan S, Weisbach M (1992) The Success of Acquisitions, Evidence from Divestitures. J Financ 47: 107–139. doi: 10.1111/j.1540-6261.1992.tb03980.x
    [64] Kumbhakar SC, Lovell KCA (2000) Stochastic Frontier Analysis. Cambridge University Press, Cambridge.
    [65] Lakonishok J, Shleifer A, Vishny RW (1994) Contrarian Investment, Extrapolation, and Risk. J Financ 49: 1541–1578. doi: 10.1111/j.1540-6261.1994.tb04772.x
    [66] Lang L, Stulz R, Walkling R (1991) A Test of the Free Cash Flow Hypothesis: The Case of Bidder Returns. J Financ Econ 29: 315–335. doi: 10.1016/0304-405X(91)90005-5
    [67] Langetieg TC (1978) An Application of a Three-Factor Performance Index to Measure Stockholder Gains from Merger. J Financ Econ 6: 365–383. doi: 10.1016/0304-405X(78)90010-7
    [68] Lee CMC, Swaminathan B (2000) Price Momentum and Trading Volume. J Financ 55: 2017–2069. doi: 10.1111/0022-1082.00280
    [69] Leeth JD, Borg JR (2000) The Impact of Takeovers on Shareholder Wealth during the 1920s Merger Wave. J Financ Quant Anal 35: 217–238. doi: 10.2307/2676191
    [70] Loughran T, Vijh AM (1997) Do Long Term Shareholders Benefit from Corporate Acquisitions? J Financ 52: 1759–1790.
    [71] Maquieria C, Megginson W, Nail L (1998) Wealth Creation versus Wealth Redistributions in Pure Stock for Stock Mergers. J Financ Econ 48: 3–33. doi: 10.1016/S0304-405X(98)00002-6
    [72] Martin KJ (1996) The Method of Payment in Corporate Acquisitions, Investment Opportunities, and Management Ownership. J Financ 51: 1227–46. doi: 10.1111/j.1540-6261.1996.tb04068.x
    [73] Masulis R, Wang C, Xie F (2007) Corporate Governance and Acquirer Returns. J Financ 62: 1851–1889. doi: 10.1111/j.1540-6261.2007.01259.x
    [74] Meeusen W, van den Broeck J (1977) Efficiency Estimation from Cobb-Douglas Production Function with Composed Error. Int Econ Rev 18: 435–444. doi: 10.2307/2525757
    [75] Modigliani F, Cohn R (1979) Inflation, Rational Valuation, and the Market. Financ Anal J 37: 24–44. doi: 10.2469/faj.v35.n2.24
    [76] Moeller SB, Chlingemann FPS, Stulz RM (2005) Wealth Destruction on a Massive Scale? A Study of Acquiring-Firm Returns in the Recent Merger Wave. J Financ 60: 757–782. doi: 10.1111/j.1540-6261.2005.00745.x
    [77] Morck R, Shleifer A, Vishny R (1990) Do Managerial Incentives Drive Bad Acquisitions? J Financ 45: 31–48. doi: 10.1111/j.1540-6261.1990.tb05079.x
    [78] Meulbroek LK (1992) An empirical analysis of illegal insider trading. J Financ 47: 1661–1699. doi: 10.1111/j.1540-6261.1992.tb04679.x
    [79] Mulherin JH (2000) Incomplete Acquisitions and Organizational Efficiency. Working Paper, Penn State.
    [80] Mulherin JH, Boone AL (2000) Comparing Acquisitions and Divestitures. J Corp Financ 6: 117–139. doi: 10.1016/S0929-1199(00)00010-9
    [81] Meulbroek LK, Hart C (1997) The effect of illegal insider trading on takeover premia. Rev Financ 1: 51–80. doi: 10.1023/A:1009755324212
    [82] Newey W, West K (1987) A Simple, Positive Semi-Definite, Heteroscedastic and Autocorrelation Consistent Covariance Matrix. Econometrica 55: 703–708. doi: 10.2307/1913610
    [83] Nguyen G, Swanson P (2009) Firm Characteristics, Relative Efficiency, and Equity Returns. J Financ Quant Anal 44: 213–236. doi: 10.1017/S0022109009090012
    [84] Palepu KG (1986) Predicting Takeover Targets: A Methodological and Empirical Analysis. J Account Econ 8: 3–35. doi: 10.1016/0165-4101(86)90008-X
    [85] Pastor L, Stambaugh RF (2003) Liquidity Risk and Expected Stock Returns. J Polit Econ 111: 642–685. doi: 10.1086/374184
    [86] Peng L (2005) Learning with Information Capacity Constraints. J Financ Quant Anal 40: 307–329. doi: 10.1017/S0022109000002325
    [87] Peng L, Wei X (2006) Investor Attention, Overconfidence and Category Learning. J Financ Econ 80: 563–602. doi: 10.1016/j.jfineco.2005.05.003
    [88] Pontiff J, Woodgate A (2008) Share Issuance and Cross-Sectional Returns. J Financ 63: 921–945. doi: 10.1111/j.1540-6261.2008.01335.x
    [89] Ravenscraft D, Scherer FM (1987) Mergers, Sell-offs, and Economic Efficiency. Washington, D.C.: Brookings Institution Press.
    [90] Rhodes-Kropf M, Viswanathan S (2004) Market Valuation and Merger Waves. J Financ 59: 2685–2718. doi: 10.1111/j.1540-6261.2004.00713.x
    [91] Rhodes-Kropf M, Robinson DT, Viswanathan S (2005) Valuation Waves and Merger Activity: The Empirical Evidence. J Financ Econ 77: 561–603. doi: 10.1016/j.jfineco.2004.06.015
    [92] Savor PG, Lu Q (2009) Do Stock Mergers Create Value for Acquirers? J Financ 64: 1061–1097. doi: 10.1111/j.1540-6261.2009.01459.x
    [93] Scheinkman JA, Xiong W (2003) Overconfidence and Speculative Bubbles. J Polit Econ 111: 1183–1219. doi: 10.1086/378531
    [94] Schwert GW (1996) Markup Pricing in Mergers and Acquisitions. J Financ Econ 41: 153–192. doi: 10.1016/0304-405X(95)00865-C
    [95] Servaes H (1991) Tobin's q and the gains from takeovers. J Financ 46: 409–419. doi: 10.1111/j.1540-6261.1991.tb03758.x
    [96] Shiller RJ (1981) Do Stock Prices Move Too Much to Be Justified by Subsequent Changes in Dividends. Am Econ Rev 71: 421–436.
    [97] Shleifer A, Vishny RW (1997) The Limits of Arbitrage. J Financ 52: 35–55. doi: 10.1111/j.1540-6261.1997.tb03807.x
    [98] Shleifer A (2000) Inefficient Markets: An Introduction to Behavioral Finance. Oxford: Oxford University Press.
    [99] Shleifer A, Vishny RW (2003) Stock Market Driven Acquisitions. J Financ Econ 70: 295–311. doi: 10.1016/S0304-405X(03)00211-3
    [100] Sloan, RG (1996) Do Stock Prices Fully Reflect Information in Accruals and Cash Flows about Future Earnings? Account Rev, 289–315.
    [101] Smith RL, Kim J (1994) The Combined Effects of Free Cash Flow and Financial Slack on Bidder and Target Stock Returns. J Bus 67: 281–310. doi: 10.1086/296633
    [102] Titman S, Wei KJ, Xie F (2004) Capital Investments and Stock Returns. J Financ Quant Anal 39: 677–700. doi: 10.1017/S0022109000003173
    [103] Travlos N (1987) Corporate Takeover Bids, Method of Payment, and Bidding Firm's Stock Returns. J Financ 52: 943–963.
    [104] Zhang XF (2006) Information Uncertainty and Stock Returns. J Financ 61: 105–137. doi: 10.1111/j.1540-6261.2006.00831.x
  • This article has been cited by:

    1. Songyang Lyu, Ray C. C. Cheung, Efficient and Automatic Breast Cancer Early Diagnosis System Based on the Hierarchical Extreme Learning Machine, 2023, 23, 1424-8220, 7772, 10.3390/s23187772
    2. Yansong Du, 2023, Research on Photographic Image Classification Based on Multi-model Fusion and Data Augmentation, 979-8-3503-1545-5, 1, 10.1109/ICIICS59993.2023.10420882
    3. Fuyang Tian, Yinuo Zhang, Shakeel Ahmed Soomro, Qiang Wang, Shuaiyang Zhang, Ji Zhang, Qinglu Yang, Yunpeng Yan, Zhenwei Yu, Zhanhua Song, SSOD-MViT: A novel model for recognizing alfalfa seed pod maturity based on semi-supervised learning, 2025, 236, 01681699, 110439, 10.1016/j.compag.2025.110439
  • Reader Comments
  • © 2021 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2637) PDF downloads(108) Cited by(2)

Figures and Tables

Figures(1)  /  Tables(8)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog