Processing math: 100%
Research article

Extension of topological structures using lattices and rough sets

  • Received: 13 December 2023 Revised: 13 January 2024 Accepted: 22 January 2024 Published: 22 February 2024
  • MSC : 54A05, 54B10, 54D30, 54G99

  • This paper explores the application of rough set theory in analyzing ambiguous data within complete information systems. The study extends topological structures using equivalence relations, establishing an extension of topological lattice within lattices. Various relations on topological spaces generate different forms of exact and rough lattices. Building on Zhou's work, the research investigates rough sets within the extension topological lattice and explores the isomorphism between topology and its extension. Additionally, the paper investigates the integration of lattices and rough sets, essential mathematical tools widely used in problem-solving. Focusing on computer science's prominent lattices and Pawlak's rough sets, the study introduces extension lattices, emphasizing lower and upper extension approximations' adaptability for practical applications. These approximations enhance pattern recognition and model uncertain data with finer granularity. While acknowledging the benefits, the paper stresses the importance of empirical validations for domain-specific efficacy. It also highlights the isomorphism between topology and its extension, revealing implications for data representation, decision-making, and computational efficiency. This isomorphism facilitates accurate data representations and streamlines computations, contributing to improved efficiency. The study enhances the understanding of integrating lattices and rough sets, offering potential applications in data analysis, decision support systems, and computational modeling.

    Citation: Mostafa A. El-Gayar, Radwan Abu-Gdairi. Extension of topological structures using lattices and rough sets[J]. AIMS Mathematics, 2024, 9(3): 7552-7569. doi: 10.3934/math.2024366

    Related Papers:

    [1] Iftikhar Ahmad, Hira Ilyas, Muhammad Asif Zahoor Raja, Tahir Nawaz Cheema, Hasnain Sajid, Kottakkaran Sooppy Nisar, Muhammad Shoaib, Mohammed S. Alqahtani, C Ahamed Saleel, Mohamed Abbas . Intelligent computing based supervised learning for solving nonlinear system of malaria endemic model. AIMS Mathematics, 2022, 7(11): 20341-20369. doi: 10.3934/math.20221114
    [2] Peng Lai, Wenxin Tian, Yanqiu Zhou . Semi-supervised estimation for the varying coefficient regression model. AIMS Mathematics, 2024, 9(1): 55-72. doi: 10.3934/math.2024004
    [3] Jun Ma, Junjie Li, Jiachen Sun . A novel adaptive safe semi-supervised learning framework for pattern extraction and classification. AIMS Mathematics, 2024, 9(11): 31444-31469. doi: 10.3934/math.20241514
    [4] Yahia Said, Amjad A. Alsuwaylimi . AI-based outdoor moving object detection for smart city surveillance. AIMS Mathematics, 2024, 9(6): 16015-16030. doi: 10.3934/math.2024776
    [5] Mashael M Asiri, Abdelwahed Motwakel, Suhanda Drar . Robust sign language detection for hearing disabled persons by Improved Coyote Optimization Algorithm with deep learning. AIMS Mathematics, 2024, 9(6): 15911-15927. doi: 10.3934/math.2024769
    [6] Peng Zhong, Xuanlong Wu, Li Zhu, Aohao Yang . A new APSO-SPC method for parameter identification problem with uncertainty caused by random measurement errors. AIMS Mathematics, 2025, 10(2): 3848-3865. doi: 10.3934/math.2025179
    [7] Chaeyoung Lee, Sangkwon Kim, Soobin Kwak, Youngjin Hwang, Seokjun Ham, Seungyoon Kang, Junseok Kim . Semi-automatic fingerprint image restoration algorithm using a partial differential equation. AIMS Mathematics, 2023, 8(11): 27528-27541. doi: 10.3934/math.20231408
    [8] Jun Ma, Xiaolong Zhu . Robust safe semi-supervised learning framework for high-dimensional data classification. AIMS Mathematics, 2024, 9(9): 25705-25731. doi: 10.3934/math.20241256
    [9] Rehab Alharbi, S. E. Abbas, E. El-Sanowsy, H. M. Khiamy, K. A. Aldwoah, Ismail Ibedou . New soft rough approximations via ideals and its applications. AIMS Mathematics, 2024, 9(4): 9884-9910. doi: 10.3934/math.2024484
    [10] Hyam Abboud, Clara Al Kosseifi, Jean-Paul Chehab . Stabilized bi-grid projection methods in finite elements for the 2D incompressible Navier-Stokes equations. AIMS Mathematics, 2018, 3(4): 485-513. doi: 10.3934/Math.2018.4.485
  • This paper explores the application of rough set theory in analyzing ambiguous data within complete information systems. The study extends topological structures using equivalence relations, establishing an extension of topological lattice within lattices. Various relations on topological spaces generate different forms of exact and rough lattices. Building on Zhou's work, the research investigates rough sets within the extension topological lattice and explores the isomorphism between topology and its extension. Additionally, the paper investigates the integration of lattices and rough sets, essential mathematical tools widely used in problem-solving. Focusing on computer science's prominent lattices and Pawlak's rough sets, the study introduces extension lattices, emphasizing lower and upper extension approximations' adaptability for practical applications. These approximations enhance pattern recognition and model uncertain data with finer granularity. While acknowledging the benefits, the paper stresses the importance of empirical validations for domain-specific efficacy. It also highlights the isomorphism between topology and its extension, revealing implications for data representation, decision-making, and computational efficiency. This isomorphism facilitates accurate data representations and streamlines computations, contributing to improved efficiency. The study enhances the understanding of integrating lattices and rough sets, offering potential applications in data analysis, decision support systems, and computational modeling.



    Person re-identification (re-ID) aims to find the person-of-interest across different cameras from the gallery when the person-of-interest is given. With the development of surveillance equipment and the increasing demand for public safety, many camera networks have been installed in public places such as theme parks, airports, streets and university campuses. Therefore, there is an urgent need to develop an intelligent technology for monitoring image analysis. At present, re-ID has been widely used in the security field and has become the focus of academic research. However, person re-ID faces more challenges such as lighting, pose, viewpoint, camera variation, and so on. At the early work of person re-ID, there is only a few small dataset, and the texture feature, color feature, and hand-crafted feature are used in the field of person re-ID. In recent years, with the development of deep Convolutional Neural Networks (CNN) and the emergence of large-scale datasets, CNN-based deep learning models have made a great success in the field of re-ID. Most person re-ID methods focus on supervised learning [1,2,3,4,5,6,7]. These methods rely on labeled datasets, that is, each pedestrian in the training data has an identity label. However, due to the high cost of labeling large-scale datasets, the semi-supervised person re-ID has been proposed.

    The semi-supervised person re-ID methods use part of the labeled data and part of the unlabeled data for training. We focus on one-example setting, i.e., each identity has only one labeled example. The setting can use the labeled data to estimate the pseudo-label of unlabeled data by calculating the feature distance between the labeled data and the unlabeled data. Then the reliable pseudo-label data are selected and added to the label data to train the model together with unlabeled data. However, due to the difference in the camera frames, the lighting, and the location of the camera, the training dataset under one-example setting will be affected by the style variation between the cameras and reduce the model performance. This problem can be solved by style transfer.

    The style transfer method is to use the Generative Adversarial Network (GAN) to transfer the labeled pedestrian image style to the unlabeled test data and use it to train the model. It can better solve the problem of performance degradation when training and testing on different datasets or between different cameras in the same dataset. At the same time, this method also avoids the tedious work of labeling data on the unlabeled test data and reduces the expensive cost of labeling data on the unlabeled test data and it is widely used in supervised person re-ID or semi-supervised person re-ID. Figure 1 shows the style transfer from random cameras to other cameras in the same dataset and the label of the generated image is the same as the original image. Therefore, we propose a random style transfer strategy that effectively solves the problem of camera style variation in the one-example setting.

    Figure 1.  An image can be generated with the help of CycleGAN to generate images with different camera styles.

    In the one-example person re-ID task, Wu et al. [8] use part of the labeled person images and part of the unlabeled person images for training. During training, an image is randomly selected as the labeled data for each identity under the camera with the smallest number, and the rest is used as unlabeled data. Then these two kinds of data are used for training. However, the method of [8] cannot eliminate the domain difference between cameras in the training phase and the label evaluation phase. The random style transfer strategy we propose is to randomly transform the image style of the labeled data and unlabeled data to other camera styles without changing the person identity label during training. This can solve the problem of camera style variation in the same training dataset, thereby improving performance. Besides, when evaluating pseudo-labels, Wu et al. [8] use the features of the labeled data as the evaluation features. However, we adopt an average feature method to average the features of the labeled data and the features after the camera style transformation, as the final evaluation feature. In the same way, there will be a better improvement. Our method can obtain a more robust initialized CNN model, and in subsequent iterations, it will obtain better performance.

    Our contributions are summarized as follow:

    ● We propose a random style transfer strategy to transform the camera style of the training data, which eliminates the style variation between different cameras in the same datasets and obtains a better initialized CNN model.

    ● We adopt an average feature strategy to estimate the pseudo-labels of unlabeled data, and the estimation results are improved.

    ● Our method has achieved good results on commonly used datasets.

    In recent years, with the advent of the CNN model [9] and the emergence of large datasets, methods based on deep learning have been widely used in computer vision tasks [10,11,12], including person re-ID. More methods [13,14,15,16,17,18,19] are used for person re-ID tasks and achieve good performance results. Deng et al. [20] use triplets as input on the siamese model. Zheng et al. [21] use a pair of images as the input of the network, and indicates whether the two images are a person according to the similarity value of the output. Ahmed et al. [22] propose a joint learning framework that combines end-to-end re-ID learning and data generation to make better use of the generated data. Zheng et al. [18] use a conventional fine-tuning approach called the Identity Discriminative Embedding (IDE) on the Market-1501 dataset and finally obtain competitive results. Zheng et al. [21] combine the verification model and recognition model to learn more differentiated pedestrian descriptors with a pair of training images. Huynh-The et al. [23] learn the full gait information of an individual by comprehensively studying gait information from 3D human skeleton data with a deep learning-based identifier.

    Semi-supervised learning methods [24,25,26,27] use partially labeled data and partially unlabeled data to solve a given task. In recent years, some semi-supervised person re-ID [28] tasks use the Progressive Cross-camera Soft Label Learning (PCSL) framework. Kipf et al. [29] use a scalable method of semi-supervised learning on graph structure data. Yu et al. [30] propose an asymmetric metric clustering to discover potential labels in unlabeled target data. Liu et al. [31] use K-Nearest Neighbor (KNN) to update the classifier. Ye et al. [32] adopt a Dynamic Graph Matching (DGM) method to iteratively update graph matching and label estimation. Liu et al. [33] propose a newemantics-Guided Clustering with Deep Progressive Learning (SGC-DPL) to gradually enhance the labeled training data.

    In this paper, we follow the one-example setting person re-ID as in [8]. It assumes that only one image of each person in the training set is labeled, while the rest of the data in the training set is not labeled. In [8], a progressive sampling strategy is proposed to increase the number of the selected pseudo-labeled candidates step by step. However, it lacks consideration of the cameras variation when estimating the pseudo-label. This can be solved with style transfer.

    The unsupervised domain adaptation re-ID uses an auxiliary source dataset to label an unlabeled target dataset. Recently, some cross-domain learning methods [34,35,36,37,38] have emerged. Peng et al. [34] propose an asymmetric multi-task dictionary learning model to learn a discriminative represtation for target data. Fan et al. [35] propose a learning via translation framework to reduce the performance deviation after dataset conversion and the unsupervised self-similarity and domain-dissimilarity to ensure potential ID information. Zhong et al. [38] propose to learn camera invariance and domain connectedness simultaneously to improve the generalization ability of re-ID models on the target testing set. Xiang et al. [39] propose a new unsupervised Re-ID method through domain adaptation to use synthetic data to get rid of heavy data annotation and improve the performance of re-identification in a completely unsupervised way. Ge et al. [40] propose an unsupervised framework, namely Mutual Mean-Teaching (MMT), to reduce the inevitable label noise caused by the clustering procedure.

    Style transfer refers to the transfer of the style of image A to image B to obtain a new image. The new image contains both the content of image B and the style of image A. The conversion process is achieved through GAN. Since Goodfellow et al. propose GAN [41], many variants of GAN [42,43,44,45,46] have been proposed to handle different tasks, such as natural style conversion, super-resolution, image conversion etc. Zheng et al. [47] use GAN to generate new samples for data augmentation of person re-ID, which is an early work on personnel migration done by GAN for person re-ID. Isola et al. [48] propose a conditional adversarial network to learn the mapping function from input to output image. However, this method requires paired training data, which is difficult to obtain in many tasks. Therefore, for the task of unpaired image-to-image conversion, Zhu et al. [49] propose a loop consistency loss training unpaired data. Later, Person Transfer GAN (PTGAN) [50] proposed by Wei et al. is similar to Cycle-GAN [49], and it can also perform image-to-image conversion. The difference is that, in order to ensure that the transmitted images can be used for model training, additional constraints are imposed on the identity of the person. In Camstyle [6], CycleGAN is used to transfer the labeled training image style to each camera, and form an enhanced training set with the original training samples. Zhang et al. [51] propose a novel semi-supervised re-ID by Similarity-Embedded Cycle GANs (SECGAN), which can learn cross-view features with limited labeled data by using cycle GAN. Our work focuses on using style transfer on one-example setting to eliminate the cameras variation. Liu et al. [52] propose a UnityStyle adaption method to solve the problem of more image artifacts when the difference between the images taken by different cameras is large. Chong et al. [53] propose an unsupervised domain-adaptive person re-identification method based on style transfer (STReID) to solve the potential image distinctions between different domains.

    This section consists of five parts: We first describe the process of using CycleGAN [49] to generate camera conversion data in subsection 3.1. We then introduce the preliminary work in subsection 3.2. The random style transfer (RST) strategy and average feature estimation (Avg) strategy are described in subsection 3.3 and subsection 3.4 respectively. The last subsection shows the overall progressive iteration strategy. The overall framework is shown in Figure 2. In each iteration, (1) In the training phase, we use the style transfer dataset(Cam set) to perform random style transfer on labeled data, pseudo-labeled data and unlabeled data to train the CNN model. We train label data and pseudo-label data through the Cross-Entropy loss, and train unlabeled data through the Exclusive Loss. (2) In the label estimation stage, we first use style transfer data to average the features of the labeled data. Then, according to the distance in the feature space, some reliable pseudo-label candidates are selected from the unlabeled data U. Nodes with different colors in the feature space frame represent different identification samples.

    Figure 2.  An overview of the framework.

    Given two datasets {x}Mi=1 and {y}Nj=1 collected from domain A and domain B, {x}Mi=1 belongs to A and {y}Nj=1 belongs to B. The goal of GAN is to learn a generator G and a discriminator D, where G:AB is to realize the conversion of the image from domain A to domain B, that is, G(A)B, D is to identify whether the image is from another domain conversion. CycleGAN contains two mapping functions GA2B:AB and GB2A:AB. Two adversarial discriminators DA and DB. The overall loss function of CycleGAN is as follows:

    L(GA2B,GB2A,DA,DB)=LG(GA2B,DB,A,B)+LG(GB2A,DA,B,A)+λLcyc(GA2B,GB2A,A,B) (3.1)

    where LG(GA2B,DB,A,B) and LG(GB2A,DA,B,A) are the loss functions for the mapping functions GA2B and GB2A and for the discriminators DB and DA. Lcyc(GA2B,GB2A,A,B) is the cycle consistency loss, in which each image can be reconstructed after a cycle mapping. λ weighs the importance of LG and Lcyc. More details about CycleGAN can be accessed in [49].

    CycleGAN is employed to transform the style of one domain into the style of another domain to realize the generation of different camera data. Given a re-ID dataset collected from C cameras, the styles between different cameras are regarded as different domains and CycleGAN is used to learn the image-image conversion model of each camera pair to realize the generation of data of different camera styles.

    In this paper, we need to generate camera style transfer data. There are many deep learning models that can realize style transfer data generation. This paper only uses CycleGAN as an example.

    Using the learned CycleGAN model, for the training images collected from a specific camera, we generate C1 new training samples. Figure 3 shows pictures with other camera styles generated by the Market-1501 dataset with the help of CycleGAN. For example, the picture in cam1 uses CycleGAN to generate a picture of cam2 style. The cam1to2 retains the pedestrian of cam1 but changes the camera style of cam1 to the camera style of cam2. There are 6 cameras in the Market-1501 dataset. Under the action of CycleGAN, each image generates 5 other camera-style images. Compared with the original image, the generated image has the same identity labeled, only the camera style changes. In this work, we will generate each style conversion image to retain the content of the original image and have the same identity as the original image. This kind of data is called CameraStyle data.

    Figure 3.  Examples of style transfer on Market-1501 [18].

    We first introduce the necessary symbols for the one-example re-ID task. Let x and y represent the person image and the identity label, respectively. For the training of one-example re-ID task, we have a labeled dataset L={(x1,y1),,(xnl,ynl)}, an unlabeled dataset U={xnl+1,,xnl+nu}, and a CamStyle dataset Z. In the training phase, these data are used to train the re-ID model ϕ(θ,) in the form of identity classification. In the evaluation phase, the trained CNN model ϕ is used to embed query data and gallery data in the feature space. The query result is a ranking list of all gallery data based on the Euclidean distance between the query data and each gallery data, i.e., ϕ(θ;xq)ϕ(θ;xg), where xq and xg represent the query image and gallery image, respectively. And we denote St and Mt as the pseudo-labeled dataset and unlabeled dataset of the tth step, respectively.

    To utilize the abundant unlabeled data, we use the joint training method [8] in the training phase to perform joint training of labeled data, pseudo-labeled data, and unlabeled data. The objective function is as follows:

    minθ,ωλnli=1CE(f(ω;ϕ(θ;xi)),yi)+λnl+nui=nl+1st1iCE(f(ω;ϕ(θ;xi)),ˆyi)+(1λ)nl+nui=nl+1(1st1i)e(V;˜ϕ(θ;xi)) (3.2)

    where minθ,ωnli=1CE(f(ω;ϕ(θ;xi)),yi) represents the optimized part of the labeled dataset L, minθ,ωnl+nui=nl+1st1iCE(f(ω;ϕ(θ;xi)),ˆyi) represents the optimized part of the pseudo-labeled dataset St, and minθ,ωnl+nui=nl+1(1st1i)e(V;˜ϕ(θ;xi)) represents the optimized part of the unlabeled dataset Mt. f(w;) is an identity classifier, parameterized by w, to classify the embedded feature ϕ(θ;xi) into a k-dimension confidence estimation. si{0,1} is the selection indicator for the unlabeled sample xi. CE and e represent identity classification loss and exclusive loss, respectively, and λ is a hyper-parameter used to adjust their contribution. More details about joint training method can be accessed in [8].

    We first introduce the types of data used in training. In the tth iteration, we will use four kinds of data: label data L, pseudo label data St, unlabeled data Mt, and CameraStyle data Z. Before the training starts, we need to generate the CamStyle dataset. We use CycleGAN mentioned in subsection 3.1 to generate one corresponding picture of other camera styles for each picture in the training dataset, that is, we keep the person label, and only change the camera style to other camera styles.

    The random style transfer strategy is to perform a random conversion of the camera style with the help of CamStyle data for each piece of labeled data, pseudo-labeled data, and unlabeled data during the training process:

    xk˜xk,˜xk{xk}{x|xZk} (3.3)

    where xk represents a picture in the dataset, and Zk represents pictures of other cameras corresponding to A in the CamStyle dataset, so ˜xk represents a camera-style image of A randomly converted, including itself.

    After adopting a random style conversion strategy for training, (3.2) will be optimized as:

    minθ,ωλnli=1CE(f(ω;ϕ(θ;˜xi)),yi)+λnl+nui=nl+1st1iCE(f(ω;ϕ(θ;˜xi)),ˆyi)+(1λ)nl+nui=nl+1(1st1i)e(V;˜ϕ(θ;˜xi)) (3.4)

    where minθ,ωnli=1CE(f(ω;ϕ(θ;˜xi)),yi) represents the optimized part of the labeled dataset L after random style transfer, minθ,ωnl+nui=nl+1st1iCE(f(ω;ϕ(θ;˜xi)),ˆyi) represents the optimized part of the pseudo-labeled dataset St after random style transfer, and minθ,ωnl+nui=nl+1(1st1i)e(V;˜ϕ(θ;˜xi)) represents the optimized part of the unlabeled dataset Mt after random style transfer.

    Therefore, the training data will reduce the domain difference caused by different cameras, thereby making the initial training model more robust and making the subsequent training process more effective.

    Previous works use the distance between labeled data and unlabeled data in the feature space as a measure of the reliability of pseudo-labels. For label estimation of unlabeled data, the nearest neighbor (NN) classifier is used to assign pseudo labels to each unlabeled data through the nearest labeled neighbor in the feature space [8]. However, it is difficult to eliminate the domain difference caused by different cameras by just using a labeled picture feature to calculate the distance from the unlabeled data. To solve this problem, we propose the average feature estimation strategy. For label estimation of unlabeled data, we find other camera-style pictures corresponding to the labeled picture in CamStyle data, and then take the average feature of all the pictures. Finally, the distance between the average feature and the unlabeled data is used as a measure of the reliability of the pseudo-label.

    We evaluate pseudo-labels for each unlabeled data xiU by:

    x,y=argmin(xl,yl)ϕ(θ;xi)ϕ(θ;˜xl) (3.5)
    d(θ;xi)=ϕ(θ;xi)ϕ(θ;x) (3.6)
    ˆyi=y (3.7)

    where ˜xl is the average of label data xlL and other camera pictures in the corresponding CamStyle data. Where d(θ;xi) is the dissimilarity cost of label estimation. In order to select candidates, in iterative step t, we sample pseudo-labeled candidates into training by setting the selection indicators as follows:

    st=argminst0=mtnl+nui=nl+1sid(θ;xi) (3.8)

    where mt denotes the size of selected pseudo-labeled set. st is the vertical concatenation of all si. (3.8) selects the top mt nearest unlabeled data for all the labeled data at the iteration step t.

    We train the CNN model iteratively. In the initial iteration, we only use the labeled data with a random style transfer strategy to initialize the model. Then in each subsequent iteration, we first optimize the model through (3.4). Then we use (3.7) to estimate the pseudo label of the unlabeled data, and select some reliable pseudo-label data by applying the trained model on (3.8).

    When selecting pseudo-labels, we adopt a dynamic sampling strategy to ensure the reliability of the selected pseudo-labeled samples. It starts with a small amount of pseudo-labeled data in the initial stage, and then merges more samples in the following stages. We set the sampled pseudo-labeled data m0=0 and unlabeled data M0=U at the beginning. In subsequent iterations, we gradually increase the size of the selected pseudo-labeled candidate set St. In iterative step t, we expand the size of the sampled pseudo-labeled data by setting mt=mt1+pnu, where p(0,1) is the selection factor, which represents the speed of magnifying the candidate set during the iteration. As described in Algorithm 1,

    Algorithm 1 The proposed method
    Require: Labeled data L, unlabeled data U, selection factorr p(0,1) initialized CNN model θ0.
    Ensure: The best CNN model θ.
    1: Initialize the selected pseudo-labeled data S0, sampling size m1pnu, iteration step t0, best validation performance V0.
    2: while mt+1U do
    3:   tt+1
    4:   Update the model (θt,wt) on L, St and Mt after random style transfer via (3.4).
    5:   Estimate pseudo labels for U via (3.7)
    6:   Generate the selection indicators st via (3.8)
    7:   Update the sampling size: mt+1mt+pnu
    8: end while
    9: for i1 to T do
    10:   Evaluate θi on the validation set performance Vi
    11:   if Vi>V then
    12:     V, θVi, θi
    13:   end if
    14: end for

     | Show Table
    DownLoad: CSV

    We first introduce the datasets and settings show the results compared with advanced methods, and finally introduce ablation experiments and result analysis.

    Market-1501 [18] contains 32,668 pieces of labeled data with 1,501 identities taken under 6 cameras. Among them, 12,936 labeled images with 751 identities are used as the training set, and 19,732 labeled images with 750 identities are used as the test set.

    DukeMTMC-reID [21] is a re-ID dataset derived from the DukeMTMC dataset [54]. It contains 36,411 pieces of labeled data with 1,404 identities taken under 8 cameras. Among them, there are 16,522 training set images with 702 identities, 2,228 query images with 702 identities, and 17,661 gallery images.

    Camarket is a dataset of the other 5 camera styles of the Market-1501 dataset generated by CycleGAN [49], with a total of 64,680 images. It's just a collection of images in the training set under other 5 camera styles. The generation of Camarket dataset can refer to [6].

    Camduke is a dataset of the other 7 camera styles of the duke dataset generated CycleGAN [49], with a total of 115,654 images. It's just a collection of pictures in the training set under other 7 camera styles. The generation of Camduke dataset can refer to [6].

    We employ the Cumulative Matching Characteristic (CMC) curve and the mean average precision (mAP) for re-ID evaluation. The CMC scores reflect the accuracy of response retrieval and we use Rank-1, Rank-5, Rank-10 scores to represent the CMC curve. Rank-1 recognition rate means that after matching according to a certain similarity matching rule, the ratio of the number of tests with the correct label to the total number of test samples can be judged for the first time. Rank-5 means that there are five opportunities to judge. The mAP is the average of the Average Precision (AP) for each query, and the AP refers to the average of precision.

    For one-example experiments, we use the same protocol as [8]. In all datasets, we randomly select an image from camera 1 as the labeled data for each identity. If camera 1 does not record any data for an identity, we will randomly select a sample from the next camera to ensure that each identity has a sample as labeled data. Other data are used as unlabeled data. Before training, we first form a set, which contains each picture and its corresponding other camera style pictures on the camarket/camduke data. During training, labeled data and unlabeled data will be randomly converted into one picture in the corresponding set. In this paper, we only use an image (single-shot) as input, not a video (multi-shot).

    We use ResNet-50 (with the last classification layer removed) as the feature embedding model for all experiments. We initialize it with ImageNet [57] pre-trained model. We set the temperature scalar τ to 0.1. The setting of λ will be discussed in subsection 4.1. In each model update step, Stochastic Gradient Descent (SGD) with a momentum of 0.5 and a weight decay of 0.0005 is used to optimize the parameters of 70 epochs with a batch size of 16. The total learning rate is initialized to 0.1. In the last 15 epochs, in order to stabilize the model training, we change the learning rate to 0.01 and set λ=1.

    The recent work on one-example person re-ID did not eliminate the domain variation between different cameras. Our method solves this problem well. The re-ID performance of our method on the two large-scale re-ID datasets are summarized in Table 1. We use a selection factor of p=0.05. The baseline(λ=0.8) is a model training based on one-example label data. Ours(RST, 0.8) represents the result after using the random style transfer strategy when λ=0.8. Ours(RST+Avg, 0.8) represents the result of adopting random style transfer strategy and average feature estimation strategy when λ=0.8. Ours(RST+Avg, 0.9) represents the best result after adjusting the hyperparameters λ=0.9. Even if there is only one labeled example for each identity, our method achieves amazing performance on image-based re-ID tasks, i.e., we achieve 10.3% and 8.4% points of Rank-1 accuracy improvement over the Baseline (one-example) on Market-1501 and DukeMTMC-reID, respectively. The performance on two large-scale datasets proves the effectiveness of our method.

    Table 1.  Performance comparison.
    Methods Marekt-1501 DukeMTMC-reID
    Rank-1 Rank-5 Rank-10 mAP Rank-1 Rank-5 Rank-10 mAP
    LOMO [7] 27.2 41.6 49.1 8.0 12.3 21.3 26.6 4.8
    BoW [18] 35.8 52.4 60.3 14.8 17.1 28.8 34.9 8.3
    UDML [34] 34.5 - - 12.4 18.5 - - 7.3
    ISR [55] 40.3 62.2 - 14.3 - - - -
    CAMEL [30] 54.5 73.1 - 26.3 40.3 57.6 - 19.8
    PUL [35] 45.5 60.7 66.7 20.5 30.0 43.4 48.5 16.4
    TJ-AIDL [36] 58.2 74.8 81.1 26.5 44.3 59.6 65.0 23.0
    PTGAN [50] 38.6 57.3 - 15.7 27.4 43.6 - 13.5
    SPGAN [37] 51.5 70.1 76.8 22.8 41.1 56.6 63.0 22.3
    SPGAN+LMP [37] 57.7 75.8 82.4 26.7 46.4 62.3 68.0 26.2
    HHL [38] 62.2 78.8 84.0 31.4 46.9 61.0 66.7 27.2
    Rank [56] 26.0 41.4 49.2 9.0 16.4 27.9 32.8 6.8
    Baseline [8] 55.8 72.3 78.4 26.2 48.8 63.4 68.4 28.5
    Ours(RST, 0.8) 63.3 78.8 83.5 30.4 54.6 66.4 71.2 30.0
    Ours(RST+Avg, 0.8) 64.6 78.5 83.0 31.6 55.5 68.1 72.9 31.0
    Ours(RST+Avg, 0.9) 66.1 80.0 84.2 32.7 57.2 69.7 74.4 32.5

     | Show Table
    DownLoad: CSV

    We compare our method with the state-of-the-art image-based person re-ID approaches. Among them, there are four hand-crafted feature representation methods (LOMO [7], BoW [18], UDML [34], ISR [55]), and multiple deep-learning-based methods. The latter includes three recent pseudo-label-learning-based methods(CAMEL [30], PUL [35], TJ-AIDL [36]), three domain-adaptation-based methods(PTGAN [50], SPGAN [37], HHL [38]) and two recent one-example-based methods (Rank [56], Baseline [8]). The results show that our method is the best compared with all methods on the two datasets. The main reason for the gap with the hand-crafted feature method is that most of the early works are based on heuristic design, so they cannot learn the best distinguishing features. Our method is superior to the unsupervised re-ID method based on pseudo-label learning. The main reason is that our average feature strategy on KNN can better assign pseudo-labels to unlabeled data. On the contrary, the pseudo-label learning of [30,35] directly compares the visual features, ignoring potential label information and camera variation. Compared with domain-adaptation-based approaches, our approach achieves superior performance. A key reason is that we have adopted style transfer and information mining for unlabeled data. After adopting the random style transfer strategy (Ours(RST, 0.8)), on the market-1501 dataset, Rank-1 and mAP increase by 7.5% and 4.2% than Baseline [8], respectively; on the DukeMTMC-reID dataset, Rank-1 and mAP increase by 4.8% and 1.5% than Baseline [8], respectively. On this basis, after using the average feature estimation strategy (Ours(RST+Avg, 0.8)), the Rank-1 and mAP of the market-1501 dataset and DukeMTMC-reID dataset can be improved again. After adjusting the hyperparameter λ=0.9, the best performance can be obtained: on the market-1501 dataset, Rank-1 and mAP increase by 10.3% and 6.5% than Baseline [8], respectively; on the DukeMTMC-reID dataset, Rank-1 and mAP increase by 8.4% and 4.0% than Baseline [8], respectively. Compared with the state-of-the-art image-based person re-ID methods, it can be seen that our method performs better than existing methods, and will get higher accuracy in many practical fields. The specific process of adjusting hyperparameters is in the next section.

    p is a key parameter in our framework, and it controls the speed at which pseudo-marked data are selected in the iterative process. A smaller selection factor indicates a lower selection speed, therefore, more iteration steps and training time are required. The results of different selection factors can be found in Figure 4. Figure 4a is the result of Rank-1 with different selection factors on Marekt-1501. Figure 4b is the result of Rank-1 with different selection factors on DukeMTMC-reID. The x-axis represents the ratio of selected data from the entire unlabeled dataset. Here we set λ=0.8. In experiments, a smaller selection factor can produce better performance. An important reason is that each selection step is more cautious, so the label estimation is more accurate. We can also find that the gap between the five curves in the first few iterations is relatively small, and the gap gradually increases in the subsequent iterations, which indicates that the estimation error continues to accumulate during the iteration.

    Figure 4.  Rank-1 comparison of different selection factor p on Marekt-1501 and DukeMTMC-reID, respectively.

    The hyperparameter λ is the only parameter that adjusts the proportion of labeled data, pseudo-labeled data and unlabeled data in the optimization process. It represents the contribution degree of labeled data and pseudo-labeled data in training. We have selected three values, λ=0.7, λ=0.8 (λ=0.8 in the baseline), λ=0.9, The results with λ=0.7, λ=0.8, λ=0.9, are shown in Figure 5. Figure 5a is the Rank-1 result of different hyperparameter λ on Marekt-1501. Figure 5b is the Rank-1 result of different hyperparameters λ on DukeMTMC-reID. The x-axis represents the ratio of data selected from the entire unlabeled dataset. Here we set p=0.05. As seen, λ=0.9 can achieve the best performance under different selection factors. An important reason is that the random style transfer strategy eliminates camera variation and the average feature estimation strategy makes the label estimation more accurate, so the labeled data and pseudo-labeled data become more reliable.

    Figure 5.  Rank-1 comparison of different λ value on Marekt-1501 and DukeMTMC-reID respectively.

    The effectiveness of the average feature estimation strategy. The average feature estimation strategy uses the average feature of the labeled data and its corresponding style transfer data to evaluate the pseudo-label. The results of it under different selection factors are shown in Table 2. Here we only take values of 0.05, 0.10, 0.15 for p. Baseline(0.8) and Avg(0.8) represent the result of baseline when λ=0.8 and the result after we only adopt the average feature estimation strategy when λ=0.8, respectively. As seen, using the average feature estimation strategy in the experiment can produce better performance. An important reason is that the average feature is better than the original feature to reduce the impact of camera style variation, so the pseudo label estimation is more accurate.

    Table 2.  The results of the average feature estimation strategy under different selection factors.
    Selection factor Methods Marekt-1501 DukeMTMC-reID
    Rank-1 Rank-5 Rank-10 mAP Rank-1 Rank-5 Rank-10 mAP
    p=0.05 Baseline(0.8) 55.8 72.3 78.4 26.2 48.8 63.4 68.4 28.5
    Avg(0.8) 59.5 74.9 80.7 30.1 51.7 66.2 70.7 29.1
    p=0.10 Baseline(0.8) 51.5 66.8 73.6 23.2 40.5 53.9 60.2 21.8
    Avg(0.8) 60.1 75.0 81.0 27.4 53.1 64.6 69.2 28.3
    p=0.15 Baseline(0.8) 44.8 61.8 69.1 19.2 35.1 49.1 54.3 18.2
    Avg(0.8) 51.0 68.1 74.1 23.6 51.3 63.1 67.8 26.4

     | Show Table
    DownLoad: CSV

    The random style transfer strategy is to randomly change the camera style of labeled data and unlabeled data in the training phase. As shown in Table 3, here p is set to 0.05, Baseline-start(0.8) represents the model initialized in the first iteration when the random style conversion strategy is not used; Ours-start(RST, 0.8) represents the model initialized in the first iteration when the random style conversion strategy is used. As seen, using a random style transfer strategy can achieve better performance. This is because the random style transfer strategy eliminates the style variation between cameras, so the obtained initialization model will be better.

    Table 3.  The result of model initialization when we adopt the random style transfer strategy.
    Methods Marekt-1501 DukeMTMC-reID
    Rank-1 Rank-5 Rank-10 mAP Rank-1 Rank-5 Rank-10 mAP
    Baseline-start(0.8) 28.3 43.6 51.0 10.1 17.2 28.1 34.1 7.1
    Ours-start(RST, 0.8) 34.0 53.0 61.6 12.4 32.9 48.1 54.4 13.7

     | Show Table
    DownLoad: CSV

    In this paper, we propose a random style transfer strategy and the average feature estimation strategy. In the training process, we adopt a random style transfer strategy to randomly change the styles of labeled data and unlabeled data. In the pseudo-label evaluation process, we adopt an average feature estimation strategy to more accurately evaluate pseudo-labels for unlabeled data. These two strategies eliminate the camera style variation during the training process and the pseudo-label evaluation process. The obvious performance improvement proves the effectiveness of our method.

    In the future, we hope to use these two strategies to improve the accuracy of video-based person re-ID tasks.

    The work is partially supported by the National Natural Science Foundation of China (Nos. U1836216, 61772322, 62076153), the major fundamental research project of Shandong, China (No. ZR2019ZD03), and the Taishan Scholar Project of Shandong, China (No. ts20190924).

    The authors declare there is no conflict of interests.



    [1] M. E. Abd El Monsef, M. A. El-Gayar, R. M. Aqeel, A comparison of three types of rough fuzzy sets based on two universal sets, Int. J. Mach. Learn. Cybern., 8 (2017), 343–353. https://doi.org/10.1007/s13042-015-0327-8 doi: 10.1007/s13042-015-0327-8
    [2] M. E. Abd El Monsef, A. M. Kozae, M. K. El-Bably, On generalizing covering approximation space, J. Egypt Math. Soc., 23 (2015), 535–545.
    [3] M. E. Abd El Monsef, M. A. El-Gayar, R. M. Aqeel, On relationships between revised rough fuzzy approximation operators and fuzzy topological spaces, Int. J. Granul. Comput. Rough Sets Intell. Syst., 3 (2014), 257–271. https://doi.org/10.1504/IJGCRSIS.2014.068022 doi: 10.1504/IJGCRSIS.2014.068022
    [4] M. E. Abd El Monsef, O. A, Embaby, M. K. El-Bably, Comparison between rough set approximations based on different topologies, Int. J. Granul. Comput. Rough Sets Intell. Syst., 3 (2014), 292–305. https://doi.org/10.1504/IJGCRSIS.2014.068032 doi: 10.1504/IJGCRSIS.2014.068032
    [5] R. Abu-Gdairi, A. A. El Atik, M. K. El-Bably, Topological visualization and graph analysis of rough sets via neighborhoods: A medical application using human heart data, AIMS Math., 8 (2023), 26945–26967. https://doi.org/10.3934/math.20231379 doi: 10.3934/math.20231379
    [6] R. Abu-Gdairi, A. A. Nasef, M. A. El-Gayar, M. K. El-Bably, On fuzzy point applications of fuzzy topological spaces, Int. J. Fuzz. Logic Intell. Sys., 23 (2023), 162–172. https://doi.org/10.5391/IJFIS.2023.23.2.162 doi: 10.5391/IJFIS.2023.23.2.162
    [7] R. Abu-Gdairi, M. A. El-Gayar, M. K. El-Bably, K. K. Fleifel, Two different views for generalized rough sets with applications, Mathematics, 9 (2022), 2275. https://doi.org/10.3390/math9182275 doi: 10.3390/math9182275
    [8] R. Abu-Gdairi, M. A. El-Gayar, T. M. Al-shami, A. S. Nawar, M. K. El-Bably, Some topological approaches for generalized rough sets and their decision-making applications, Symmetry, 14 (2022). https://doi.org/10.3390/sym14010095
    [9] M. I. Ali, M. K. El-Bably, E. A. Abo-Tabl, Topological approach to generalized soft rough sets via near concepts, Soft Comput., 26 (2022), 499–509. https://doi.org/10.1007/s00500-021-06456-z doi: 10.1007/s00500-021-06456-z
    [10] A. A. Azzam, A topological tool to develop novel rough set, J. Math Comput., 32 (2024), 204–216.
    [11] A. A. Azzam, Comparison of two types of rough approximation via grill, Italian J. Pure Appl. Math., 47 (2022), 258–270.
    [12] N. Ba˘girmaz, I. Icen, A. F. ¨Ozcan, Topological rough groups, Topol. Algebra Appl., 4 (2016), 31–38.
    [13] B. Banaschewski, Extensions of topological spaces, Canad. Math. Bull., 7 (1964), 1–2.
    [14] G. Birkhoff, Lattice theory, Providence: American Mathematical Society Colloquium Publications, 1967.
    [15] C. J. R. Borges, On extensions of topologies, Canadian J. Math., 9 (1967), 474–487.
    [16] J. Chen, J. Li, Y. Lin, G. Lin, Z. Ma, Relations of reduction between covering generalized rough sets and concept lattices, Inform. Sci., 304 (2015), 16–27.
    [17] D. Chen, W. Zhang, D. Yeung, E. C. C. Tsang, Rough approximations on a complete completely distributive lattice with applications to generalized rough sets, Inform. Sci., 176 (2006), 1829–1848.
    [18] P. Crawley, R. P. Dilworth, Algebraic theory of lattices, Prent. Hall, 1973.
    [19] B. A. Davey, H. A. Priestely, Introduction to lattice and order, Cambr. Univ. Press, 1990.
    [20] B. A. Davey, H. A. Priestely, Introduction to lattice and order, Sec. Ed. Cambr. Univ. Press, 2002.
    [21] A. A. El Atik, A. S. Wahba, Topological approaches of graphs and their applications by neighborhood systems and rough sets, J. Intell. Fuzzy Syst., 39 (2020), 6979–6992. https://doi.org/10.3233/JIFS-200126 doi: 10.3233/JIFS-200126
    [22] A. A. El Atik, A. A. Nasef, Some topological structures of fractals and their related graphs, Filomat, 34 (2020), 153–165. https://doi.org/10.2298/FIL2001153A doi: 10.2298/FIL2001153A
    [23] M. K. El-Bably, R. Abu-Gdairi, M. A. El-Gayar, Medical diagnosis for the problem of Chikungunya disease using soft rough sets, AIMS Math., 8 (2023), 9082–9105. https://doi.org/10.3934/math.2023455 doi: 10.3934/math.2023455
    [24] M. K. El-Bably, A. A. El Atik, Soft β-rough sets and their application to determine COVID-19, Turk. J. Math., 45 (2021), 1133–1148. https://doi.org/10.3906/mat-2008-93 doi: 10.3906/mat-2008-93
    [25] M. A. El-Gayar, R. Abu-Gdairi, M. K. El-Bably, D. I. Taher, Economic decision-making using rough topological structures, J. Math., 2023 (2023). https://doi.org/10.1155/2023/4723233
    [26] M. A. El-Gayar, A. El Atik, Topological models of rough sets and decision making of COVID-19, Complexity, 2022 (2022). https://doi.org/10.1155/2022/2989236
    [27] A. A. Estaji, M. R. Hooshmandas, Z. B. Davva, Rough set theory applied to lattice theory, Inform. Sci., 200 (2012), 108–122.
    [28] R. B. Esmaeel, M. O. Mustafa, On nano topological spaces with grill-generalized open and closed sets, AIP Conf. Proc., 2414 (2023). https://doi.org/10.1063/5.0117062
    [29] R. B. Esmaeel, N. M. Shahadhuh, On grill-semi-P-separation axioms, AIP Conf. Proc., 2414 (2023). https://doi.org/10.1063/5.0117064
    [30] S. Fomin, Extensions of topological spaces, Ann. Math., Second Series, 44 (1943), 471–480.
    [31] G. Grätzer, General lattice theory, Basel: Birkhäuser Verlag, 1978.
    [32] C. Guido, On a characterization of simple extensions of topologies, Note De Mat., 1981.
    [33] J. Gutiˊerrez Garcia, I. Mardones-Pˊerez, M. A. de Prada Vicente, Insertion and extension theorems for lattice-valued functions on preordered topological spaces, Topology Appl., 158 (2011), 60–68.
    [34] G. Gr¨atzer, Lattice theory: Foundation, Springer Basel, 2011. https://doi.org/10.1007/978-3-0348-0018-1
    [35] W. He, D. Peng, M. Tkachenko, Z. Xiao, Gaps in lattices of (para) topological group topologies and cardinal functions, Topology Appl., 264 (2019), 89–104.
    [36] K. Hu, Y. Sui, Y. Lu, J. Wang, C. Shi, Concept approximation in concept lattice, knowledge discovery and data mining, Lecture Notes Comput. Sci., 2035 (2001), 167–173.
    [37] Z. Huang, J. Li, Covering based multi-granulation rough fuzzy sets with applications to feature selection, Expert Syst. Appl., 238 (2024), 121908. https://doi.org/10.1016/j.eswa.2023.121908 doi: 10.1016/j.eswa.2023.121908
    [38] Z. Huang, J. Li, C. Wang, Robust feature selection using multigranulation variable-precision distinguishing indicators for fuzzy covering decision systems, IEEE Trans. Syst. Man. Cyber. Syst., 2023. https://doi.org/10.1109/TSMC.2023.3321315
    [39] Z. Huang, J. Li, Multi-level granularity entropies for fuzzy coverings and feature subset selection, Artif. Intell. Rev., 56 (2023), 12171–12200. https://doi.org/10.1007/s10462-023-10479-3 doi: 10.1007/s10462-023-10479-3
    [40] Z. Huang, J. Li, Y. Qian, Noise-Tolerant Fuzzy-β-Covering-based multigranulation rough sets and feature subset selection, IEEE Trans. Fuzzy Syst., 30 (2022), 2721–2735. https://doi.org/10.1109/TFUZZ.2021.3093202 doi: 10.1109/TFUZZ.2021.3093202
    [41] Z. Huang, J. Li, Discernibility measures for fuzzy-β-covering and their application, IEEE Trans. Cyber., 52 (2022), 9722–9735. https://doi.org/10.1109/TCYB.2021.3054742 doi: 10.1109/TCYB.2021.3054742
    [42] J. Järvinen, Approximations and rough sets based on tolerances, Berlin Heidelberg: Springer-Verlag, 2001.
    [43] J. Järvinen, The ordered set of rough sets, Berlin Heidelberg: Springer-Verlag, 2004.
    [44] J. Järvinen, Lattice theory for rough sets, Berlin Heidelberg: Springer-Verlag, 2007.
    [45] N. Levine, Simple extensions of topologies, Amer. Math. Monthly, 71 (1964), 22–25.
    [46] M. Liu, M. Shao, W. Zhang, C. Wu, Reduction method for concept lattices based on rough set theory and its application, Comput. Math. Appl., 53 (2007), 1390–1410.
    [47] G. L. Liu, W. Zhu, The algebraic structures of generalized rough set theory, Inform. Sci., 178 (2008), 4105–4113.
    [48] R. E. Larson, S. J. Andima, The lattice of topologies, Rocky Mountain J. Math., 5 (1975), 177–198.
    [49] X. Li, H. Yi, S. Liu, Rough sets and matroids from a lattice-theoretic viewpoint, Inform. Sci., 342 (2016), 37–52.
    [50] H. Lu, A. M. Khalil, W.d. Alharbi, M. A. El-Gayar, A new type of generalized picture fuzzy soft set and its application in decision making, J. Intell. Fuzzy Syst., 40 (2021), 12459–12475.
    [51] M. N. Mukherjee, B. Roy, R. Sen, On extensions of topological spaces in terms of ideals, Topology Appl., 154 (2007), 3167–3172.
    [52] Z. M. Ma, B. Q. Hu, Topological and lattice structures of L-fuzzy rough sets determined by lower and upper sets, Inform. Sci., 218 (2013), 194–204.
    [53] S. Mapes, Finite atomic lattices and resolutions of monomial ideals, J. Algebra, 379 (2013), 259–276.
    [54] A. S. Nawar, M. A. El-Gayar, M. K. El-Bably, R. A. Hosny, θβ-ideal approximation spaces and their applications, AIMS Math., 7 (2022), 2479–2497. https://doi.org/10.3934/math.2022139 doi: 10.3934/math.2022139
    [55] Z. Pawlak, Rough sets, Int. J. Inform. Comput. Sci., 11 (1982), 341–356. https://doi.org/10.1007/BF01001956
    [56] Z. Pawlak, Rough sets. Algebric and topological approach, ICS PAS Reports, 482 (1982). Available from: https://api.semanticscholar.org/CorpusID: 111800999
    [57] Z. Pawlak, Rough sets, rough relation and rough function, Fund. Inform., 27 (1996), 103–108.
    [58] Z. Pei, D. Pei, L. Zheng, Topology vs generalized rough sets, Int. J. Approx. Reason., 52 (2011), 231–239.
    [59] G. L. Qi, W. R. Liu, Rough operations on Boolean algebras, Inform. Sci., 173 (2005), 49–63.
    [60] J. Reinhold, Finite intervals in the lattice of topologies, Appl. Cat. Struct., 8 (2000), 36–376.
    [61] S. K. Roy, S. Bera, Approxiamtion of rough, soft set and its application to lattice, Fuzzy Inf. Eng., 7 (2015), 379–387.
    [62] A. Tan, J. Li, G. Lin, Connections between covering-based rough sets and concept lattices, Int. J. Approx. Reason., 56 (2015), 43–58.
    [63] M. Vlach, Algebraic and topological aspects of rough set theory, Fourth Int. Workshop Comput. Intell. Appl., 10 (2008), 23–30.
    [64] P. Whitman, Lattices, Equivalence relations, and subgroups, Bull. Amer. Math. Soc., 52 (1946), 507–522.
    [65] L. Wei, J. J. Qi, Relation between concept lattice reduction and rough set reduction, Knowl. Based Syst., 23 (2010), 934–938.
    [66] Z. Wang, Q. Feng, H. Wang, The lattice and matroid representations of definable sets in generalized rough sets based on relations, Inform. Sci., 485 (2019), 505–520.
    [67] F. Wehrung, Cevian operations on distributive lattices, J. Pure Appl. Algebra, 224 (2020), 106–202.
    [68] Q. Xiao, Q. Li, L. Guo, Rough sets induced by ideals in lattices, Inform. Sci., 271 (2014), 82–92.
    [69] J. Yang, X. Zhang, Some weaker versions of topological residuated lattices, Fuzzy Sets Sys., 373 (2019), 62–77.
    [70] Y. Y. Yao, Two views of the theory of rough sets in finite unverses, Int. J. Approx. Reason., 15 (1996), 291–317.
    [71] Y. Y. Yao, Generalized rough set models, Rough Sets Knowl. Dis., 1 (1998). 286–318.
    [72] Y. Y. Yao, Two semantic issues in probabilistic rough set model, Fund. Inform., 108 (2011), 249–265.
    [73] Z. Yu, X. Bai, Z. Yun, A study of rough sets based on 1-neighborhood systems, Inform. Sci., 248 (2013), 103–113.
    [74] P. Zhang, T. Li, G. Wang, C. Luo, H. Chen, J. Zhang, et al., Multi-source information fusion based on rough set theory: A review, Inform. Fusion, 68 (2021), 85–117. https://doi.org/10.1016/j.inffus.2020.11.004 doi: 10.1016/j.inffus.2020.11.004
    [75] P. Zhang, T. Li, Z. Yuan, C. Luo, K. Liu, X. Yang, Heterogeneous feature selection based on neighborhood combination entropy, IEEE Trans. Neur. Net. Learn. Syst., 2022. https://doi.org/10.1109/TNNLS.2022.3193929
    [76] N. L. Zhou, B. Q. Hu, Rough sets based on complete completely distributive lattice, Inform. Sci., 269 (2014), 378–387.
    [77] D. Zou, Y. Xu, L. Li, Z. Ma, Novel variable precision fuzzy rough sets and three-way decision model with three strategies, Inform. Sci., 629 (2023), 222–248. https://doi.org/10.1016/j.ins.2023.01.141 doi: 10.1016/j.ins.2023.01.141
  • This article has been cited by:

    1. Junkai Deng, Zhanxiang Feng, Peijia Chen, Jianhuang Lai, 2021, Chapter 30, 978-3-030-88003-3, 366, 10.1007/978-3-030-88004-0_30
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1161) PDF downloads(40) Cited by(3)

Figures and Tables

Figures(5)  /  Tables(2)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog