
Biometric authentication prevents losses from identity misuse in the artificial intelligence (AI) era. The fusion method integrates palmprint and palm vein features, leveraging their stability and security and enhances counterfeiting prevention and overall system efficiency through multimodal correlations. However, most of the existing multi-modal palmprint and palm vein feature extraction methods extract only feature information independently from different modalities, ignoring the importance of the correlation between different modal samples in the class to the improvement of recognition performance. In this study, we addressed the aforementioned issues by proposing a feature-level joint learning fusion approach for palmprint and palm vein recognition based on modal correlations. The method employs a sparse unsupervised projection algorithm with a "purification matrix" constraint to enhance consistency in intra-modal features. This minimizes data reconstruction errors, eliminating noise and extracting compact, and discriminative representations. Subsequently, the partial least squares algorithm extracts high grayscale variance and category correlation subspaces from each modality. A weighted sum is then utilized to dynamically optimize the contribution of each modality for effective classification recognition. Experimental evaluations conducted for five multimodal databases, composed of six unimodal databases including the Chinese Academy of Sciences multispectral palmprint and palm vein databases, yielded equal error rates (EER) of 0.0173%, 0.0192%, 0.0059%, 0.0010%, and 0.0008%. Compared to some classical methods for palmprint and palm vein fusion recognition, the algorithm significantly improves recognition performance. The algorithm is suitable for identity recognition in scenarios with high security requirements and holds practical value.
Citation: Wei Wu, Yuan Zhang, Yunpeng Li, Chuanyang Li. Fusion recognition of palmprint and palm vein based on modal correlation[J]. Mathematical Biosciences and Engineering, 2024, 21(2): 3129-3145. doi: 10.3934/mbe.2024139
[1] | Zeynep Cındık, Ismail H. Armutlulu . A revision of Altman Z-Score model and a comparative analysis of Turkish companies' financial distress prediction. National Accounting Review, 2021, 3(2): 237-255. doi: 10.3934/NAR.2021012 |
[2] | Nikolaos T. Giannakopoulos, Damianos P. Sakas, Nikos Kanellos, Christos Christopoulos . Web analytics and supply chain transportation firms' financial performance. National Accounting Review, 2023, 5(4): 405-420. doi: 10.3934/NAR.2023023 |
[3] | Dachen Sheng, Opale Guyot . Corporate governance in Chinese manufacturing sector: Ownership structure, monitoring and firms' earning quality. National Accounting Review, 2023, 5(4): 421-443. doi: 10.3934/NAR.2023024 |
[4] | Juan De Lucio, Raúl Mínguez, Asier Minondo, Francisco Requena . Top exporters and regional export specialization. National Accounting Review, 2024, 6(4): 564-572. doi: 10.3934/NAR.2024026 |
[5] | Guillermo Peña . The inequality-credit linkage. National Accounting Review, 2025, 7(2): 143-176. doi: 10.3934/NAR.2025007 |
[6] | Alberto Bernardi, Daniela Bragoli, Davide Fedreghini, Tommaso Ganugi, Giovanni Marseguerra . COVID-19 and firms' financial health in Brescia: a simulation with Logistic regression and neural networks. National Accounting Review, 2021, 3(3): 293-309. doi: 10.3934/NAR.2021015 |
[7] | Serena Fatica, Wildmer Daniel Gregori . Income shifting by multinational enterprises as a source of mismeasurement in aggregate statistics. National Accounting Review, 2024, 6(4): 548-563. doi: 10.3934/NAR.2024025 |
[8] | Dina Krasnoselskaya, Venera Timiryanova . Exploring the impact of ecological dimension on municipal investment: empirical evidence from Russia. National Accounting Review, 2023, 5(3): 227-244. doi: 10.3934/NAR.2023014 |
[9] | Dev Gupta, Akanksha Parikh, Tapan Kumar Datta . A multi-criteria decision-making approach to rank the sectoral stock indices of national stock exchange of India based on their performances. National Accounting Review, 2021, 3(3): 272-292. doi: 10.3934/NAR.2021014 |
[10] | Jinhong Wang, Yanting Xu . Factors influencing the transition of China's economic growth momentum. National Accounting Review, 2024, 6(2): 220-244. doi: 10.3934/NAR.2024010 |
Biometric authentication prevents losses from identity misuse in the artificial intelligence (AI) era. The fusion method integrates palmprint and palm vein features, leveraging their stability and security and enhances counterfeiting prevention and overall system efficiency through multimodal correlations. However, most of the existing multi-modal palmprint and palm vein feature extraction methods extract only feature information independently from different modalities, ignoring the importance of the correlation between different modal samples in the class to the improvement of recognition performance. In this study, we addressed the aforementioned issues by proposing a feature-level joint learning fusion approach for palmprint and palm vein recognition based on modal correlations. The method employs a sparse unsupervised projection algorithm with a "purification matrix" constraint to enhance consistency in intra-modal features. This minimizes data reconstruction errors, eliminating noise and extracting compact, and discriminative representations. Subsequently, the partial least squares algorithm extracts high grayscale variance and category correlation subspaces from each modality. A weighted sum is then utilized to dynamically optimize the contribution of each modality for effective classification recognition. Experimental evaluations conducted for five multimodal databases, composed of six unimodal databases including the Chinese Academy of Sciences multispectral palmprint and palm vein databases, yielded equal error rates (EER) of 0.0173%, 0.0192%, 0.0059%, 0.0010%, and 0.0008%. Compared to some classical methods for palmprint and palm vein fusion recognition, the algorithm significantly improves recognition performance. The algorithm is suitable for identity recognition in scenarios with high security requirements and holds practical value.
Leukemia is always considered being a life-scaring disease [1]. It attacks the blood-forming capacity of bone marrow [2,3,4]. Bone marrow generates leukocytes (white blood cells) gradually through different stages involving the generation of the blast cells [5]. Due to the development of leukemia, these blasts cells in the human blood tend to increase [6]. A higher amount of blasts cells and other abnormal cells lead to severe leukemia [7]. The detection is the task of observing the morphologies of leukocytes and predicting the type and severity of leukemia [8].
It has four broad types: acute myeloid leukemia (AML), acute lymphoblastic leukemia (ALL), chronic myeloid leukemia (CML) and chronic lymphoblastic leukemia (CLL) [10]. In addition, there is the existence of the components like blasts cells (immature leukocytes), Auer rods (a very thin rod like structure) in the blood sample of leukemia patients.
Figure 1 shows the microscopic image of the blood cells having normal and leukemia cells [9]. Microscopic analysis is required to detect the presence of leukemia cells [11]. Figure 1 indicates that leukemia cells differ morphologically from normal cells. The diagnosis of leukemia involves the detection of these abnormal characteristics of blood cells [12]. The very first and the important step in the detection is segmentation [13]. In the microscopic image of a stained blood sample, there are many components, including leukocytes-white blood cells, erythrocytes-red blood cells, plasma, and parasites [14]. Staining is required for making the blood smear for microscopic analysis [15,16]. Segmentation in leukemia detection involves separating leukocytes from the given stained blood microscopic image to get the region of interest [17]. The popular segmentation techniques are thresholding [18], clustering [19], edge-based segmentation [20], region-based segmentation [21], artificial neural network-based segmentation [22], and partial differential equation-based segmentation [23]. However, a generalized segmentation methodology involves pre-processing, background separation, and the outcome will be the region of interest for further detection purposes [24].
In the proposed methodology, Otsu and Kapur thresholding are utilized for the segmentation of leukocytes. Furthermore, to enhance the thresholding performance, computational time utilization, and optimization, the learning enthusiasm-based teaching-learning-based optimization (LebTLBO) is used. It has been successfully used in various science and engineering applications in recent years. In the traditional teaching-learning-based optimization (TLBO)and most of its variants, all learners have an equal chance of learning from others. LebTLBO was inspired by the learner's varied chances of acquiring knowledge compared to others, and it incorporated a learning excitement mechanism into the basic TLBO [25].
This optimization involves the teaching-learning process in the class [26]. In this algorithm, there are three phases of working, namely teaching phase, learning phase, and poor student tutoring phase [27]. Broadly, in this case, the students with a good desire to take knowledge from their teacher will get good grades and alter their positions towards top. Students with less desire will fall to the bottom, and will not update their positions towards the top. In addition to this, the poor phase is also present, which analyzes the same thing for poor students. Here poor students are those who falls within the lower 10% of the whole class. The learning desire of the poor students is also analyzed and their positions are also considered to be updated. So this involves the position alteration of good students as well as poor students, giving more precise optimization [28]. The same is explained in more depth in Section 3.
The paper is divided into six different sections. Section 1 presents the introduction. Section 2 talks about the related work. Section 3 explores the proposed technique for segmentation with a detailed methodology. Different performance metrics are explained in Section 4. Finally, Section 5 contains the results and discussion, followed by the conclusion and future scope.
There are different segmentation methods employed for the white blood cell separation for the detection of leukemia by different researchers including mathematical morphology, fuzzy logic approach [29], watershed transform [30], k-means clustering [31], zack algorithm [32], marker controlled watershed transform [33], clustering based on stimulating discriminant measures (SDM) [34,35], hough transform [36], iterative thresholding with watershed transform [37], edge thresholding, triangular thresholding DOST algorithm [38], Otsu's method [39,40], a conventional neural network with laplacian of gaussian (LOG) and coupled edge profile active contours (C-EPAC) [41], sequential maximum angle convex cone (SMACC) clustering with iterative) algorithm and iterative self-organizing data analysis technique algorithm (ISODATA), and local binary pattern (LBP) [39], VGG16 segmentation [42].
The methods adopted by researchers are mainly Otsu thresholding which becomes a time-consuming technique when the number of threshold levels are increased [43] and needs an appropriate tuning for proper threshold selections [44,45]. The majority of methods adopting the thresholding concept [46,47,48] employed the binary classification of images, where the image is divided into background and foreground only. But the blood microscopic images comprise of 4 to 5 different parts. These include leukocyte, erythrocytes, platelets, plasma parasites (due to the cause of some infections). The development of leukemia in the human body, causes the generation of immature leukocytes (blasts cells), and other morphological component such as Auer rod in the blood. Therefore, the investigation comprises of the separation of these 4 to 5 components and detection of the morphological abnormalities in the blood cell images. Hence, it is very challenging to adopt binary thresholding methods for segmentation purposes.
Figure 2 shows the different parts in the microscopic image of blood-leukocyte, erythrocyte [49]. The blood image can contain plasma and some parasites as well. So, it is very difficult to segment the image with binary thresholding. As a result, more than one threshold is required for segmentation to separate these various components. Therefore, there is a need to adopt a multi-level thresholding technique for blood microscopic images. Artificial intelligence and deep learning-based methods will take a significant time for performing the segmentation due to the requirement of more iterations for better results. Most times, where deep learning may be adopted, there is a need to use tuned hyper-parameters to carry out the segmentation. These tuning always influences the performance of the model [50].
It also deals with the problems of under-fitting and over-fitting. So proper tuning of hypermeters is always an important issue in deep learning models. This tuning process is time-consuming and is also a resource-consuming task [52,53]. One more problem with deep learning algorithms is the requirement of very large datasets for better results [54,55]. As leukemia is cancer and its infected patients are comparatively in fewer numbers, it is difficult to get a large number of imaging samples for building a larger dataset. On the other hand, traditional image processing algorithms perform well with a small number of dataset images. Clustering is one more approach for segmentation that requires the user to initialize the cluster priory, which is the drawback of the same [56].
There are also different issues during image capturing, such as resolution, the contrast which are to be considered when these methods are to be used for segmentation purposes [57]. Otsu is the most popular thresholding method of segmentation [58]. Time expensiveness is the major drawback of this algorithm; it has very high time consumption [59]. According to Tuba [44] for 10 threshold levels, this algorithm requires a time of about 10,000 years with an I-7processor [44].
Therefore, the aforementioned issues have motivated the authors to come up with different optimization techniques to be adopted for image segmentation purposes. Figure 3 shows the different optimization techniques.
All evolutionary and swarm intelligence-based algorithms are probabilistic, and they all require the same set of regulating parameters, such as population size, generation number, elite size, and so on. Different algorithms require their algorithm-specific control parameters, in addition to the common control parameters as explained in Table 1.
Algorithm | Control parameters |
GA | Mutation probability, crossover probability and the selection operator. |
ABC | Number of spectator bees, employed bees, scout bees and limit. |
PSO | Inertia weight, social and cognitive parameters. |
Harmony search | Memory consideration rate, pitch adjustment rate and the number of improvisations are all factors in the HS algorithm. |
Algorithm-specific parameters are generally modified when using optimization algorithms [60]. The correct tuning of algorithm-specific parameters is a critical aspect that influences the performance of the algorithms discussed above. Incorrect tuning of algorithm-specific parameters either increases processing effort or produces the local best solution. Rao et al. [61] recognized this and developed the TLBO algorithm, which does not require any algorithm-specific parameters [62]. For the TLBO algorithm to work, it only needs a few basic governing parameters, like population size and generation number. Among optimization researchers, the TLBO algorithm has acquired widespread popularity [63].
From the discussion of the above methods, there is a need for a simple and less time-consuming technique for segmentation. Though there are a variety of optimization algorithms, the TLBO and its variant learning enthusiastic type prove to be simple and time inexpensive, as explained in the previous paragraph. For that reason, LebTLBO is adopted in this work.
Input microscopic images of the stained blood sample are taken from publicly available databases1. The segmentation methodology can be explained as below. Table 2 shows the dataset images of ALL-IDB dataset.
1 (Available at https://homes.di.unimi.it/scotti/all/)
Name | Image formats | Number of images | Color depth | Remark |
ALL-IDB-1 | JPEG | 109 (510 lymphoblast) |
24-bit, 2592 × 1944 |
Cancerous |
ALL-IDB-2 | JPEG | 260 (130 lymphoblast) |
24-bit 257 × 257 |
Cancerous |
1) Images loading: The images are taken from the publicly available databases. It consists of the microscopic images of stained blood smears of normal and leukemia infected patients. Here, ALL-IDB dataset is considered for the study and implementation of the algorithm [64,65,66]. This database has two parts ALL-IDB-1 consisting of 260 images of normal and leukemia patients, and ALL-IDB2 has 106 images of normal and leukemia patients.
2) Pre-processing: The images obtained from the microscope may have certain errors during the image capturing. Also, there is a requirement for images to convert in grayscale format for further processing. So the database images are gray-scaled for further processing.
3) Segmentation: The images are parsed in the further stage of segmentation. It involves the extraction of the region of interest. For detection of leukemia, white blood cell-leukocytes are the region of interest in the blood smear microscopic images, which are to be the outcome of segmentation. A broader classification of segmentation methods includes thresholding, edge-based methods, region-based methods, clustering methods, watershed-based methods, artificial neural network-based methods, etc. [67].
Figure 4 shows the proposed approach for segmentation for leukemia detection purposes.
Out of the different segmentation methods stated earlier in Section 2, thresholding is the most simple and easy step towards segmentation for separating foreground and background in an image. It is further sub-divided into local and global thresholding [68]. Global thresholding uses an image as a whole, while distinguished characteristics of local image areas may be used to choose thresholds of the image in local thresholding type. The methods adopted here are Otsu and Kapur multi-level thresholding for the problem formulation, optimizing each of them using LebTLBO [26]. Following are the different concepts used for segmentation.
The most important concept utilized here is multi-level thresholding rather than binary thresholding because binary threshold divides the image into two major parts-background and foreground. In the case of blood microscopic images, there are different components, including plasma, leukocyte, erythrocytes, and some other cells like parasites. So for segmentation, there is a need for more than one threshold to separate these different components. Hence, multi-level thresholding is preferred over binary type. In multi-level thresholding, more than two thresholds are generally used.
So instead of generating only two regions, as in the case of bi-level thresholding, multi-level thresholding generates several regions such as [r1,r2,r3,r4,…….rn] cconsidering the thresholding values, and this can lead to more accurate results of segmentation. In addition, different threshold values, such as th1,th2,th3,th4……..tc are considered in this case. Equation 1 could further explore this.
r1←d(m,n),if0<d(m,n)<th1r2←d(m,n),ifth1<d(m,n)<th2r3←d(m,n),ifth2<d(m,n)<th3ri←d(m,n),ifthi<d(m,n)<thi+1…rk←d(m,n),ifThne−1<d(m,n)<Thne | (1) |
where ne are the number of threshold levels.
For multi-level thresholding in the segmentation, Otsu and Kapur-based algorithms are the most popular [69]. Otsu comes under the global method of image thresholding [70]. Amongst the different thresholding methods, multi-level Otsu is more popular among the researchers [71,72]. Otsu maximizes the variances between the classes. The strategy here is to process the image histogram and segment the objects by minimizing the thresholds into the different classes. Generally, image histograms may contain two clear peaks representing different intensity ranges. The intra-class variance, defined as the weighted sum of the variances between the two classes, is minimized by finding a threshold, as shown in Eq (2).
σ2ω(t)=ω0(t)σ20(t)+ω1(t)σ21(t) | (2) |
Weights ω0 and ω1 are probabilities of different classes, t defines the threshold value, and σ0 and σ1 are the class variances. The histogram is used to obtain the class probability, ω0, 1 (t) is calculated from L bits of the histogram.
ω0(t)=∑t−1i=0p(i) | (3) |
ω1(t)=∑L−1i=tp(i) | (4) |
For two classes, minimization of the intra-class variance is the same as maximizing the inter-class variance.
σ2b(t)=σ2−σ2ω(t)=ω0(t)ω1(t)[μ0(t)−μ1(t)]2 | (5) |
The class means μ0(t), μ1(t) and μT are given by
μ0(t)=∑t−1t=0p(i)ω0(t) | (6) |
μ1(t)=∑L−1i=tp(i)ω1(t) | (7) |
μT(t)=∑L−1i=0ip(i) | (8) |
foTSU(th)=ϕ0=max(σ2B(th)),0≤th≤L−10<th<L−1,i=1,2,3….k | (9) |
When the histograms have a bimodal distribution and a deep and abrupt valley between the peaks, Otsu performs reasonably well. In the case of smaller object areas, the histogram may not possess bimodality. Also, if there is large additive noise, the histogram's sharp valley will get degraded. Then, there may be the segmentation error due to incorrect threshold values. As a result, tiny object size, a lower mean difference between background and foreground, significant variations between object and background pixels, a considerable quantity of noise, and other factors may limit object segmentation performance [73].
Kapur's method can prove a valuable metric for multi-level segmentation [74]. Kapur [75] developed the Kapur's entropy technique in 1985, which maximized the histogram entropy of segmented classes to find the best threshold values. This method proves to be effective and feasible for image segmentation [76]. Shannon's entropy function is conceptualized as an inverse proportion between the probabilities of occurrence of an event and the information [77]. Higher the probability of occurrence of the event exhibits fewer information contents. The entropy presented by Shannon is given by the equation,
H=−∑mn=1pklog(pk) | (10) |
where H- Entropy and Pk- the probability of the nth gray level, m-total number of levels in the gray image. The probability distribution considered here is between background and foreground in an image. The entropy of an image does not consider the spatial distribution of an image. Due to this, the same histogram having different images offers the same value of entropy and the threshold [77]. Assuming p1, p2, p3…p as the distribution of probability for a gray level image. Mostly, there are two probability distributions to be considered. Eqs (11) and (12) define them as,
A:P1PsP2Ps⋯⋅PsPs | (11) |
B:PS+11−PsPs+21−Ps⋯⋅Pn1−Ps | (12) |
H(A) and H(B) gives the entropies associated with A and B in Eqs (13) and (14), respectively, The value of ϕs, which gives the sum of H(A) and H(B) has to be maximum, So that maximum information will be obtained between background and foreground.
H(A)=−∑si=1PiPsInPiPs=−1Ps∑si=1PiInPi−PsInPs=InPs+HsPs | (13) |
H(B)=−∑n1+sPi1−PsInPi1−Ps=−11−Ps[∑ni=1+sPiInPi−1−PsIn(1−Ps)]=In(1−Ps)Hn−Hs1−Ps | (14) |
ϕs=HA+HB=InPs+HsPs+In(1−Ps)+Hn−Hs1−Ps | (15) |
ϕs=InPs(1−Ps)+(Hn−Hs)Ps+Hs(1−Ps)Ps(1−Ps) | (16) |
The threshold value is obtained from Eq (16) of ϕs where it is maximum, after considering the discrete values of s in the equation. To get the exact distinction between backgrounds and foreground, the maximum value of ϕs is considered. Most techniques consisting of the metaheuristic approach for optimization follow a similar technique [78], where the fitness function is the entropy to maximize output. Multilevel thresholding using the Kapur approach is always superior to the single level to get more precise and accurate results, as the image segmentation is considered, as follows by Eq (16) [79]. This entropy-based method considers the histogram of a gray level as the base. There will be scattered in the optimal threshold as the entropy reaches maximization:
H(0)=−∑t1−1i=0PiP0InPiP0,P0=∑t−1i=0Pi | (17) |
H(1)=−∑t2−1i=0PiP1InPiP1,P1=∑t2−1i=0Pi | (18) |
H(2)=−∑t3−1i=0PiP2InPiP2,P2=∑t3−1i=0Pi | (19) |
H(j)=−∑tj+1−1i=tjPiPjInPiPj,Pj=∑tj+1−1i=tjPi | (20) |
H(m)=−∑L−1i=tmPiPmInPiPm,Pm=∑L−1i=tmPm | (21) |
The objective function is to maximize when selecting the optimal threshold.
fkapur(th)=ϕk=argmax−∑mi=1Hi(th),0≤th≤L−1 | (22) |
Optimal threshold values for multiple thresholds are given as:
fkapur(TH)=fkapur(thi),i=1,2,3….,k | (23) |
th1,th2,…are the threshold values combined in a vector TH, I represent a specific class as provided in Eq (2).
Researchers have developed a variety of optimization algorithms [80]. There are certain popular categories of these algorithms. As the image segmentation is considered, some optimization algorithms such as genetic algorithm (GA) [81], particle swarm optimization (PSO) [82], ant bee colony (ABC) [83], etc. are presented. A newly developed algorithm by [61] is based on the classroom's teaching-learning process, TLBO. Some popular variants of these algorithms include basic, modified [84], differential learning approach [85], nonlinear inertia weighted approach [86], approach with learning experience from others [87], generalized oppositional approach [88], etc. In addition to these variants of TLBO, there is a recently developed variant known as LEBTLBO. Singh et al. [26] compared this optimization technique with GA, PSO and other popular traditional techniques for segmentation. They found that LebTLBO explores the segmentation with greater significance. For the segmentation problem of leukemia detection, the same optimization algorithm is adopted along with multi-level thresholding as an objective function here. This involved the classroom behavior of the teaching and learning process. Therefore, it has two phases- the teaching phase and the learning phase. Simple implementation and fast convergence are the main advantages of this algorithm. This basic algorithm is applied to many problems related to engineering, such as machining, drilling, ultrasonic machining, electrochemical machining, etc. [89], and also for image segmentation purposes [90]. There are certain modifications done are TLBO to enhance the search capability for finding the optimum solution. Although, as in traditional TLBO, every student is considered to have the same learning potential, LebTLBO works with the real-world problem related to enthusiasm for learning. More enthusiastic students in the class may also have good concentration and are keen to know the concepts taught to them by their teacher. On the contrary, less enthusiastic students may have less concentration towards teaching and are generally not open to getting information from teachers. LebTLBO is motivated by this kind of real-world teaching mechanism. In this approach, a third phase is introduced, called as "a poor phase of tutoring". Learning enthusiasm is calculated based on the grades of students. Students with good grades are considered to have higher learning enthusiasm and the students with less or lower grades are considered to have less learning enthusiasm.
Teaching learning is a unique process. The enthusiasm amongst the students towards learning defines the students as good students. This enthusiasm leads them to get good marks and also makes them interested to get the new knowledge from their teachers. When they gained knowledge from their teacher, they will be able to pursue good marks in their examinations. After getting good grades, their positions will be altered and they will achieve good ranks, updating their positions to the top. On the other hand, the poor students or the students with less enthusiasm will surely get less marks and will not move towards the top positions. This behavior is the basis of teaching phase of the LebTLBO algorithm. Figure 5 shows different steps of the teacher phase. In the mathematical manner, it is explained as below.
For this phase to construct, there are two presumptions considered.
1) The students having decent evaluations are more excited about learning and are more open to getting information from the instructor.
2) The students with fewer evaluations have less energy for learning and are less open to acquiring knowledge from their instructor. In this phase, learners' grades are taken as the base for sorting them into the different categories from best to worst. For a maximization problem, an assumption is made that,
f(x1)≥f(x2)≥…..≥f(XK) | (24) |
After this, learning enthusiasm values are defined as,
LEi=LEmin+(LEmax−LEmin)K−iK | (25) |
where, i = 1, 2, ...., K, where LEmax and LEmin are the maximum and minimum value of enthusiasm of learners towards learning from their teacher. Depending upon the value of LEi (learning enthusiasm value), students are classified into two categories as-gaining from the instructor or not gaining from the instructor. An irregular number is created for the student (xi) as ri ∈ [0, 1]. If ri ≤ LEi, the student will gain knowledge from his instructor. In the other case, the student will not be open to learning from the educator. When xi gains knowledge from his teacher, his position gets updated, and he gets upgraded his knowledge. Depending upon the learner's desire, whether to learn or not learn from the educator, the learner position will be updated with the following Eq (26).
xdi,new=f(x)={xdi,old+rand2⋅(xdteacher−TF⋅xdmean),ifrand1<0.5xdr1+F.(xdr2+xdr3),otherwise | (26) |
where, r1,r2,r3 are randomly selected integers from {1,2,3,….NP}, d∈{1,2,3,..D}, rand1andrand2 are two random variables distributed uniformly between [0,1], F is a scale factor in the range of [0,1].
The basis of this phase is the learners will learn from other learners depending on their desire towards learning. Their desire is lesser than the learning enthusiasm, then they will learn from other learners. This means, the learners take the knowledge from other students also instead of the teacher. However, the position of the student is updated, only when it gains the knowledge from the teacher. Figure 6 shows different steps of the learner phase.
The main assumption here is that learners with strong grades have a high level of learning excitement and that these learners have a high likelihood of gaining knowledge from their teachers. The students are rated from best to worst based on their grades for learning openness. This is based on Eq (25). A random number is created for the learner xi using the range of ri ∈ [0, 1]. If ri ≤ LEi, the learner xi will learn from other learners; otherwise, the present learner xi will ignore the learner's knowledge of the learner. However, if current learner xi can obtain knowledge from the teacher, its position will be revised, taking into account diversity-enhanced teaching. This is shown in Eq (27),
xdi,new={xdi,old+rand2⋅(xi−xj)iff(xi)≥f(xj)xdi,old+rand2⋅(xj−xi)iff(xi)≤f(xj) | (27) |
where, rand denotes a uniformly distributed random vector in the range [0, 1], f(xi) denotes the objective function, and xi, old is the ith learner's old position. If xi, new is superior to xi, old, xi, new will be accepted. Otherwise, there will be no change in xi, old.
Poor students are sorted first and then amongst the poor students, the desire of gaining knowledge is again considered for categorizing them. The learning desire is considered for updating their grades in this case as well. This phase will work same as the learning phase, with a change that, the students considered here are below 10% level amongst all the students. This can be explained in the mathematical manner and is described in the following paragraph.
This phase differentiates the LebTLBO algorithm from the basic TLBO. Here, improvement in the grades of the poor students is the prime motive. The Poor learners are considered if they fall below the 10% level of the whole class. A learner named xT is picked randomly from among the impoverished student's xi, with a rank in the top half of the class.
The following Equation shows the learning,
xdi,new=xdi,old+rand(xdT−xdi,old) | (28) |
where, d = 1, 2, ...., D. In this phase also, xi, new will be accepted depending upon whether it is better than xi, old. Otherwise, it earlier remains unchanged. The flow chart of the same is shown in Figure 7. Those who receive good grades are more likely to alter their positions, whereas students who receive poor grades are less likely to do so. This strategy is based on the real-time teaching-learning process. Therefore, it proves more feasible in its working regarding the position updating of learners.
In this approach, Otsu and Kapur methods are given by Eqs (9) and (22), fkapur (TH) and fOTSU (TH) are considered as the objective functions for the optimization. Multiple optimal threshold values are to be found with the maximization of the objective function. Let thi = (thi1, thi2, ...., thik) denotes the population vector of ith position of different threshold values-k, where thi(j) ∈ 0,255. Equation (29) gives the initial population of students,
thi(j)=thmin+rand()∗(thmax−thmin) | (29) |
where thmin and thmax are the image's minimum and maximum intensity values, j is the number of threshold levels, and a rand is a random number, respectively. Otsu and Kapur's entropy is used to calculate each's fitness value. The student with the maximum knowledge gives the best objective function for the algorithms.
For segmentation, the personal laptop is used with 8 GB of RAM and Intel-core I-5 processor. In addition, the Python software with version 3.8 is utilized for the same.
Segmentation is the method for extraction of the region of interest for the given application. The first thing to evaluate segmentation performance is to compare the segmented image with the ground truth image. In this respect, different measures are identified for the evaluation of the system's performance. These parameters include volume overlapping error [91], dice index [92], relative volume difference [92], jaccard index [93], mean-square-error and peak-signal-to-noise-ratio [94].
For this to evaluate, segmented images and the ground truth masks are considered in their volumes. It is given by,
VOE=AandBAorB | (30) |
where, A = ground truth image, B = segmented image
The volumes of both images are compared and considered to be overlapped on each other. The amount of overlap indicates the correct segmentation. The amount of volume which does not overlap indicates the error in segmentation. The amount of non-overlapping between two images gives the overlapping error. Its value should be lesser to indicate effective segmentation.
It evaluates the difference between the segmented image and the mask image which is the ground truth. It is given by,
RVD=A−BB | (31) |
where, A = ground truth image, B = segmented image.
In this performance metric, the ground truth image is subtracted from the segmented image. So its value will be lesser when the segmentation is effective. Less or zero value of this parameter indicates that, the segmented image and the mask image are exactly similar, showing the correct segmentation.
This accuracy metric will compare the ground truth mask (a mask prepared manually by a medical professional, such as a radiologist or a pathologist) against the mask developed using segmentation.
JaccardIndex=AandBAorB=TPTP+FN+FP | (32) |
where, TP = true positive, FN = false negative, FP = false positive.
It is also given in terms of dice score as,
JaccardIndex=DiceScore2−DiceScore | (33) |
Jaccard index and the dice coefficient are quite similar. The intersection is counted twice with the dice coefficient (TP). It is given by,
Dicescore=2×(AandB)(AandB)+(AorB) | (34) |
Jaccard index and dice coefficients indicates the number of truly segmented images out of the total images. As the expectation of effective segmentation is that, all the images are to be segmented. Hence, the value of Jaccard index and dice coefficient should be as large as possible. Its maximum value is 1 [95].
It calculates the squared error between the estimator and the estimated average. This parameter can analyze the segmentation performance.
It is the ratio of the signal with noise that signifies the loss during the segmentation process. It is expressed in dB with the following equation,
PSNR=10log10peakvalue2MSE | (35) |
MSE and PSNR are both used for getting the segmentation performance. MSE gives the error between mask image and the segmented image, while PSNR gives the signal-to-noise ratio. MSE value should be as low as possible to effective segmentation. MSE should always be lesser to indicate error-free segmentation. According to Eq (35), with lesser value of MSE, PSNR will lead to a greater value. According to the theory, the higher the PSNR, the better the degraded image has been reconstructed to match the original image, and the reconstructive algorithm is better.
Intersection over union (IoU) [96] is a type of image segmentation evaluation matrix that measures how much the goal ground truth mask overlaps with the prediction output. It is calculated by dividing the intersection of target and prediction pixels by the total number of pixels in both masks. According to Eq (36), IoU value will approach to high for error-free segmentation.
IoU=Target∩PredictionTarget∪Prediction | (36) |
The perceptual difference between two comparable images is measured using SSIM. The formula for calculating the similarity index (SI) is as follows [97]:
SI=mab2xy2mambmambx2+y2m2a+m2b | (37) |
where, x and y denote the mean values of images original image image1-K (i, j) and enhanced image image2-I (i, j) and ma2, mb2 and mab denotes the variance of image1-K (i, j), image2-I (i, j) and covariance of image1-K (i, j) and image2-I (i, j). This index is used to get the structural similarity between the ground truth image and the segmented image. Hence, its value should be higher to indicate good segmentation. Its maximum value will be 1.
It is defined as the difference between the original and improved image [98] and is expressed as below,
MAE=1n∑ni=0|fi−yi|=1n∑ni=1|ei| | (38) |
where, fi is segmented images and Yi is the mask or ground truth image. This parameter finds the difference between the ground truth image and the segmented image. Hence, to show true segmentation, its value approach to zero.
The results are visualized as shown in Figures 8–10. Figure 8a shows the original image of the dataset and the orange ring indicates the leukocyte to be segmented. Figure 8b is the mask, also called as ground truth image. It is obtained by manually segmenting the original image and the leukocyte is annotated by an expert pathologist. The segmentation task involves gray scaling of this original image, followed by the multi-level thresholding with Otsu and Kapur methods and then optimizing the performance with a popular LebTLBO optimization approach. Figures 9a, b are the segmented leukocytes by Kapur approach and optimized Kapur approach. In these Figures, the orange rings and the arrows indicate segmentation errors. Figure 10a, b are the outcomes of segmentation with Otsu and optimized Otsu approaches. These Figures indicate that optimized Otsu gives better segmentation as compared with other approaches in our proposed methodology.
In addition to the visualized results, different statistical parameters are also considered for the analysis of the performance of segmentation with our approach. Figure 11 shows the average results obtained as an outcome of segmentation on ALL-IDB dataset images.
Jaccard index and dice coefficient are the indicators of the true positives with respect to the total images. These two parameters indicate the amount of segmentation. These parameter approach to the maximum value to show good segmentation. In our results, as indicated in Figure 11a), d) dice score and jaccard index are maximum when optimization is applied with basic Otsu and Kapur methods. This justifies the improvement in segmentation with optimized thresholding. Other parameter is VOE that gives the error in overlapping of volumes of ground truth image and the segmented image. This parameter is expected to have lower value ensuring the good segmentation. As shown in the Figure 11b), this error has less value in optimized Kapur approach. In addition to this, other popular error metrics are also evaluated for analyzing the segmentation performance. These include RVD, MSE, and MAE. As these are the errors in segmentation showing volume difference and mean errors, their value should also be lower to indicate good segmentation performance in our result visualization as shown in Figure 11c), e), i). These error parameters also show less values with optimized approach. Another parameter, SSIM is also evaluated as performance metric of segmentation. It shows similarities between the ground truth image and the segmented image. The value of this parameter should be maximum to indicate good segmentation. As indicated in Figure 11h), SSIM has a maximum value with an optimized approach. The next metric for evaluation of segmentation performance is PSNR. Its value should be higher to ensure good segmentation. Also, in this case, optimized approach looks better than traditional thresholding approaches, as shown in Figure 11f). IoU is also considered as the performance metric, which gives higher value with optimized approach. This is indicated in Figure 11g).
In addition to these parameters, execution time analysis is performed. The time taken for segmentation for the traditional thresholding approach and with optimized approach is calculated. The time considered for single image segmentation with and without optimization. The visualization is shown in Figure 12. It indicates that optimized approaches require less time for segmentation.
The execution time analysis is also done with and without optimization which is shown in Figure 12. The segmentation is carried out on Core-I3 system with 4 GB RAM, and on Pycharm IDE platform. The segmentation time is considered for a single image with two thresholds.
The proposed approach is compared with the approach stated by Dhal et al. [99]. In their approach, the segmentation of leukocytes is presented by using some popular bio-inspired optimization algorithms and a clustering approach using K-means and mean shift. The bio-inspired optimization approaches used are stochastic fractal search (SFS), ant bee colony (ABC), particle swarm optimization (PSO) and differential evolution (DE). PSNR and similarity index are considered for comparison of these methods. This comparison is shown in Figure 13a)–c).
Two thresholding methods, Otsu and Kapur, are used for segmentation. The search capabilities of these methods are improved by the LebTLBO optimization method. Otsu deals with the within-class and between-class variance, while Kapur has entropy as the thresholding parameter. Parameters considered for measuring the performance of the segmentation are Dice Score, which gives the relation of union and intersection between the segmented image and the ground truth image. As shown in Figure 11a, the dice score of the proposed methodology with Kapur and Optimized Kapur varies from 0.565731 to 0.60281, and the same for Otsu and Optimized Otsu varies from 0.628412 to 0.632528. This comparison indicates that Otsu thresholding with an optimization approach with LebTLBO gives a better dice score during segmentation evaluation. Jaccard index also varies from 0.466313 to 0.472316 for all four approaches. It also indicates the optimized Otsu offers good performance as the Jaccard index is considered.
The other parameter considered for evaluation is the relative volume difference, which is the difference in the volumes of mask-ground truth image and the segmented image. Its value is lesser in all four cases. Although, optimized Otsu indicates lesser RVD compared to the other three approaches with the value -0.0964. In addition to this, volume overlapping error is also considered for analysis. It gives AND and OR operations of ground truth image and segmented image. Its value lies between 0.44 to 0.52, while the lesser error is obtained in the optimized Kapur method. MSE and PSNR are other parameters considered for the evaluation of segmentation performance. MSE and MAE are also lesser in the Optimized Otsu approach with corresponding values of 0.0981 and 0.0761, respectively, indicating the lowest error with the optimized Otsu method. Regarding MSE, another parameter PSNR is considered, which should be larger for better segmentation. PSNR shows the signal-to-noise ratio, which also proves a parameter showing the segmented image quality compared to ground truth after segmentation. In the proposed methodology, it varies from 56.55 to 58.75, with different approaches. It also clearly states that Optimized Otsu proves to be better than other approaches.
IoU varies from 0.471 to 0.5573, indicating that, Optimized Otsu provides better IoU compared to other approaches in the methodology. The similarity index is also one more measure considered for analysis. Its value ranges from 0.9872 to 0.9906. Also, in this case, optimized Otsu proves to be a better segmentation approach.
Our proposed approach is found to offer a good comparative yield in terms of PSNR with the highest value of 58.80 with optimized Otsu, in comparison with approaches stated by [99]. In their methodology, PSNR varied from 25.43 to 32.39. The similarity index is another measure considered for comparison. It is found to be 0.99 in the case of our methodology, which is quite superior to the methodology proposed by [99], with a higher value of 0.8893 in the SFS approach.
Our methodology is also compared with the one proposed by Shahzad [42]. Here, the measure considered is IoU. Our approach gives IoU varying from 0.44 to 0.56, which is better than the measure obtained by [42] with a value of 0.3199.
Figures 9 and 10, shows the images as a result of segmentation with different methods. Kapur method shows an under segmentation (with more segmented regions than expected). When the optimization is applied, the segmentation is improved. Otsu has an over-segmentation (with fewer segmented regions than expected), and it is also improved with optimization. The main reason for the differential behavior of Otsu and Kapur is the basis of these two algorithms. Otsu works on maximization of between-class variance, whereas Kapur maximizes the entropy for measuring the homogeneity of the classes. Hence, the segmentation results differ with Otsu and Kapur. Although, optimization improves the segmentation results of both the algorithms.
In biological assays, there is a requirement of counting and segmentation of cells. This was considered to be a prospective research area of microscopic imaging. In practice, the manual counting of cell colony forming units is used to calculate the surviving fraction (SF) in clonogenic assays. This method is operator specific and may be prone to error. Furthermore, because of the fast development rate that causes nearby colonies to merge, determining the actual colony number is frequently impossible. The colony size, which is often connected with the given radiation dose or the administered cytotoxic agent, is not considered in traditional assessment. Researchers adopted different methodologies for these problems. These methods include Circular Hough Transform and adaptive thresholding [100], multi-level thresholding with feedback based on watershed algorithm [101], bilateral filtering followed by watershed transform and morphological filtering [102], binary classifier inspired by quantum machine learning theory with homogeneity considered to be the relevant feature [103]. The stated approach in our paper can be explored with the problem of clonological assay in biomedical field. Our proposed approach will provide novel insights related to the segmentation in biological assays.
In this research, an improved and optimized method for image segmentation is proposed and evaluated for the detection of leukemia via microscopic imaging. Images are taken from a popular and publicly available dataset ALL-IDB. Segmentation involved separation between background and foreground from microscopic images. Multi-level thresholding is applied for the segmentation purpose using two methods, Otsu and Kapur. Otsu is based on intensity, while Kapur is an entropy-based algorithm. The histogram is considered for the threshold selection in these cases. There is a problem of threshold selection with Otsu and Kapur if there are fewer histogram changes that lead to improper threshold values. So the performance of the system is boosted by optimizing the same with the LebTLBO optimization algorithm. This has increased the search capacity of the algorithm for getting proper threshold values for segmentation. Different parameters used for the analysis of segmentation, performance indicates the assurance of the optimization technique to improve the segmentation performance. A comparative analysis of the proposed approach is done against the segmentation with bio-inspired optimization techniques such as SFS, ABC, PSO and clustering algorithm such as K-means and, other algorithms including DE and mean shift for segmentation purpose of leukocyte with the same dataset ALL-IDB. Performance measures such as PSNR and Similarity index are considered and the proposed approach was found to be better in terms of both parameters. The comparison is performed with a convolutional encoder- decoder framework with VGG-16 model, where IoU is the measure considered for the comparison. This system proved to be better over bio-inspired and some clustering algorithms. Therefore, for leukocyte detection, LebTLBO optimization will prove better than bio-inspired and clustering systems. These segmented images could be used further for the classification and diagnosis of leukemia. This work comes under traditional image processing, hence is always self-explainable and could always be more trusted than the unexplainable natures of newly developed AI techniques.
In the proposed methodology, an ALL-IDB database is utilized. This methodology is to be tested with other available databases of leukemia and with real patients' microscopic images, in order to prove the robustness of the proposed algorithm. Generative Adversarial Networks (GAN) could be employed further for improvement in segmentation performance. GAN could be combined with a transfer learning approach (TLA) for further enhancement in system performance.
This work was supported in part by the Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and IT, University of Technology Sydney (UTS), in part by the Researchers Supporting Project, King Saud University, Riyadh, Saudi Arabia, under Grant RSP-2021/14.
The authors declare not competing interests.
[1] |
S. P. Zhao, L. K. Fei, J. Wen, Multiview-learning-based generic palmprint recognition: A literature review, Mathematics, 11 (2023), 1261–1261. https://doi.org/10.3390/math11051261 doi: 10.3390/math11051261
![]() |
[2] |
S. Y. Li, B. Zhang, Joint discriminative sparse coding for robust hand-based multimodal recognition, IEEE Trans. Inf. Forensics and Secur., 16 (2021), 3186–3198. https://doi.org/10.1109/TIFS.2021.3074315 doi: 10.1109/TIFS.2021.3074315
![]() |
[3] | K. Zhang, H. Wang, C. Yu, M. Du, L. Tao, Class constraint-based discriminative features learning algorithm for palm print and palm vein fusion recognition, in 2022 7th International Conference on Signal and Image Processing (ICSIP), (2022), 275–280. https://doi.org/10.1109/ICSIP55141.2022.9886189 |
[4] | H. S. Kala, S. Kumar, R. B. Reddy, N. Shastry, R. Thakur, Contactless authentication device using palm vein and palm print fusion biometric technology for post covid world, in 2021 International Conference on Design Innovations for 3Cs Compute Communicate Control (ICDI3C), (2021), 281–285. https://doi.org/10.1109/ICDI3C53598.2021.00063 |
[5] | T. R. Yashavanth, M. Suresh, Performance analysis of multimodal biometric system using LBP and PCA, in 2023 International Conference on Recent Trends in Electronics and Communication (ICRTEC), (2023), 1–5. https://doi.org/10.1109/ICRTEC56977.2023.10111925 |
[6] | S. Li, B. Zhang, L. Wu, R. Ma, X. Ning, Robust and sparse least square regression for finger vein and finger knuckle print recognition, IEEE Trans. Inf. Forensics Secur., (2024), 2709–2719. https://doi.org/10.1109/TIFS.2024.3352389 |
[7] |
T. Wu, L. Leng, M. K. Khan, F. A. Khan, Palmprint-palmvein fusion recognition based on deep hashing network, IEEE Access, 9 (2021), 135816–135827. https://doi.org/10.1109/ACCESS.2021.3112513 doi: 10.1109/ACCESS.2021.3112513
![]() |
[8] | L. G. Oldal, A. Kovács, Hand geometry and palmprint-based authentication using image processing, in 2020 IEEE 18th International Symposium on Intelligent Systems and Informatics (SISY), (2020), 125–130. https://doi.org/10.1109/SISY50555.2020.9217068 |
[9] | C. Ramachandran, D. Sankar, Score level based fusion method for multimodal biometric recognition using palmprint and Iris, in 2020 Advanced Computing and Communication Technologies for High Performance Applications (ACCTHPA), (2020), 281–286. https://doi.org/10.1109/ACCTHPA49271.2020.9213216 |
[10] |
N. Bala, R. Gupta, A. Kumar, Multimodal biometric system based on fusion techniques: A review, Inf. Secur. J., 31(2022), 289–337. https://doi.org/10.1080/19393555.2021.1974130 doi: 10.1080/19393555.2021.1974130
![]() |
[11] |
S. C. Cho, B. S. Oh, K. A. Toh, Z. P. Lin, Extraction and cross-matching of palm-vein and palmprint from the RGB and the NIR spectrums for identity verification, IEEE Access, 8(2020), 4005–4021. https://doi.org/10.1109/ACCESS.2019.2963078 doi: 10.1109/ACCESS.2019.2963078
![]() |
[12] | H. S. Mohammad, A. D. Mohammad, S. Hamed, A hybrid approach for multimodal biometric recognition based on feature level fusion in reproducing Kernel Hilbert Space, in 2022 30th International Conference on Electrical Engineering (ICEE), (2022), 796–802. https://doi.org/10.1109/ICEE55646.2022.9827154 |
[13] | D. V. R. Devi, K. N. Rao, Decision level fusion schemes for a multimodal biometric system using local and global wavelet features, in 2020 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT), (2020), 1–6. http://dx.doi.org/10.1109/conecct50063.2020.9198547 |
[14] | W. L. Yang, J. D. Huang, D. Luo, W. X. Kang, Efficient disentangled representation learning for multi-modal finger biometrics, Pattern Recognit., 145 (2024). https://doi.org/10.1016/j.patcog.2023.109944 |
[15] | H. S. Kala, K. Shreya, B. R. Rakshanda, S. Nishanth, T. Rounak, Contactless authentication device using palm vein and palmprint fusion biometric technology for post covid world, in 2021 International Conference on Design Innovations for 3Cs Compute Communicate Control (ICDI3C), (2021), 281–285. https://doi.org/10.1109/ICDI3C53598.2021.00063 |
[16] |
T. F. Wu, L. Leng, M. K. Khan, F. A. Khan, Palmprint-palmvein fusion recognition based on deep hashing network, IEEE Access, 9 (2021), 135816–135827. https://doi.org/10.1109/ACCESS.2021.3112513 doi: 10.1109/ACCESS.2021.3112513
![]() |
[17] |
C. Y. Zhou, J. Huang, F. Yang, Y. Q. Liu, A hybrid fusion model of iris, palm vein and finger vein for multi-biometric recognition system, Multimedia Tools Appl., 79 (2020), 1–22. https://doi.org/10.1007/s11042-020-08914-6 doi: 10.1007/s11042-020-08914-6
![]() |
[18] | S. A. Bala, Z. B. Abdullahi, C. Ponlawat, A. Muhammad, Sequence-wise multimodal biometric fingerprint and finger-vein recognition network (STMFPFV-Net), Intell. Syst. Appl., 19 (2023). https://doi.org/10.2139/ssrn.4432356 |
[19] |
W. Wu, Q. Wang, S. Q. Yu, Q. Luo, S. Lin, Z. Han, Y. D. Tang, Outside box and contactless palm vein recognition based on a wavelet denoising ResNet, IEEE Access, 9 (2021), 82471–82484. https://doi.org/10.1109/ACCESS.2021.3086811 doi: 10.1109/ACCESS.2021.3086811
![]() |
[20] | L. Zhang, L. Li, A. Yang, Y. Shen, M. Yang, Towards contactless palmprint recognition: A novel device, a new benchmark, and a collaborative representation based identification approach, Pattern Recognit., 69 (2017), 199–212. https://doi.org/10.1016/j.patcog.2017.04.016 |
[21] |
Y. Zhou, A. Kumar, Human identification using palm-vein images, IEEE Trans. Inf. Forensics Secur., 6 (2011), 1259–1274. https://doi.org/10.1109/TIFS.2011.2158423 doi: 10.1109/TIFS.2011.2158423
![]() |
[22] |
D. Zhang, Z. Guo, G. M. Lu, L. Zhang, W. M. Zuo, An online system of multispectral palmprint verification, IEEE Trans. Instrum. Measurement, 59 (2010), 480–490. https://doi.org/10.1109/TIM.2009.2028772 doi: 10.1109/TIM.2009.2028772
![]() |
[23] | W. Q. Yuan, W. Wu, S. Lin, H. Song, H. T. Zhang, Non-contact biometric recognition of palm vein based on block and partial least squares, Chinese J. Sci. Instrum., 34 (2013), 31–39. |
[24] |
B. Jay, K. Ajay, On estimating performance indices for biometric identification, Pattern Recognit., 42 (2009), 1803–1815. https://doi.org/10.1016/j.patcog.2008.10.004 doi: 10.1016/j.patcog.2008.10.004
![]() |
[25] |
J. Y. Wang, L. Wang, F. P. Nie, X. L. Li, Joint feature selection and extraction with sparse unsupervised projection, IEEE Trans. Neural Networks Learn. Syst., 34 (2021), 3071–3081. https://doi.org/10.1109/TNNLS.2021.3111714 doi: 10.1109/TNNLS.2021.3111714
![]() |
[26] |
E. Alhatami, U. A. Bhatti, M. X. Huang, S. L. Feng, Review and enhancement of discrete cosine transform (DCT) for medical image fusion, Springer Nat. Singapore, 614 (2023), 89–97. http://dx.doi.org/10.1007/978-981-19-9331-2_8 doi: 10.1007/978-981-19-9331-2_8
![]() |
[27] |
J. Almaghtuf, F. Khelifi, A. Bouridane, Fast and efficient difference of block means code for palmprint recognition, Mach. Vision Appl., 31 (2020), 478–487. https://doi.org/10.1007/s00138-020-01103-3 doi: 10.1007/s00138-020-01103-3
![]() |
[28] |
G. Manas, D. Aniruddha, Fractional-weighted entropy-based fuzzy G-2DLDA algorithm: A new facial feature extraction method, Multimedia Tools Appl., 82 (2022), 2689–2707. https://doi.org/10.1007/s11042-022-13328-7 doi: 10.1007/s11042-022-13328-7
![]() |
[29] |
A. Zainab, W. L. Jong, A. K. Muhammad, S. Muhammad, A. K. Sajid, R. Naveed, Optical character recognition (OCR) using partial least square (PLS) based feature reduction: an application to artificial intelligence for biometric identification, J. Enterp. Inf. Manage., 36 (2020), 767–789. https://doi.org/10.1108/JEIM-02-2020-0076 doi: 10.1108/JEIM-02-2020-0076
![]() |
[30] |
X. W. Zhao, J. Guo, F. P. Nie, L. Chen, Z. H. Li, H. X. Zhang, Joint principal component and discriminant analysis for dimensionality reduction, IEEE Trans. Neural Networks Learn. Syst., 31 (2020), 433–444. https://doi.org/10.1109/TNNLS.2019.2904701 doi: 10.1109/TNNLS.2019.2904701
![]() |
1. | Stephen Sarfo Adu-Yeboah, Yuanchun Jiang, Michelle Frempomaa Frempong, Md Altab Hossin, Richard Amoako, Corporate sustainability and firm performance in small and medium enterprises in Ghana: Mediating role of green innovation, 2022, 32, 1433-0237, 311, 10.1080/14330237.2022.2066341 | |
2. | Michael Appiah, Mingxing Li, Saba Sehrish, Emad Eddin Abaji, Investigating the connections between innovation, natural resource extraction, and environmental pollution in OECD nations; examining the role of capital formation, 2023, 81, 03014207, 103312, 10.1016/j.resourpol.2023.103312 | |
3. | Muhammad Abubakr Naeem, Michael Appiah, Sitara Karim, Larisa Yarovaya, What abates environmental efficiency in African economies? Exploring the influence of infrastructure, industrialization, and innovation, 2023, 186, 00401625, 122172, 10.1016/j.techfore.2022.122172 | |
4. | Michael Appiah, Mingxing Li, John Taden, Sania Ashraf, Aviral Kumar Tiwari, Prosper Basommi Laari, Enhancing natural resource rents through industrialization, technological innovation, and foreign capital in the OECD countries: Does financial development matter?, 2024, 89, 03014207, 104520, 10.1016/j.resourpol.2023.104520 | |
5. | Ying Fu, Jiaying Li, Ying Wu, Islam Abdeljawad, International operations and corporate R&D investment: Evidence from China, 2024, 19, 1932-6203, e0308290, 10.1371/journal.pone.0308290 | |
6. | Lloyd George Banda, Emmanuel George Yusufu, Greenson Chimjinthi Nyirenda, Exploring the Modification Effect of Corporate Governance on the Relationship Between Good Governance and Logistics Performance, 2024, 06, 74, 10.12944/JBSFM.06.01.06 | |
7. | Natalia Sokrovolska, Alina Korbutiak, Artur Oleksyn, Oleh Boichenko, Natalia Danik, STATE REGULATOR’S ROLE IN THE COUNTRY’S BANKING SYSTEM DURING WARTIME, 2023, 2, 2310-8770, 43, 10.55643/fcaptp.2.49.2023.3985 | |
8. | Prince Dorian Rivel Bambi, Marly Loria Diabakanga Batatana, Michael Appiah, Derrick Tetteh, Governance, institutions, and climate change resilience in Sub-Saharan Africa: assessing the threshold effects, 2024, 12, 2296-665X, 10.3389/fenvs.2024.1352344 | |
9. | Bahadır Karakoç, A Nonlinear Relationship Between Corporate Size and Profitability, 2025, 18, 1308-3740, 151, 10.29067/muvu.1569770 |
Algorithm | Control parameters |
GA | Mutation probability, crossover probability and the selection operator. |
ABC | Number of spectator bees, employed bees, scout bees and limit. |
PSO | Inertia weight, social and cognitive parameters. |
Harmony search | Memory consideration rate, pitch adjustment rate and the number of improvisations are all factors in the HS algorithm. |
Name | Image formats | Number of images | Color depth | Remark |
ALL-IDB-1 | JPEG | 109 (510 lymphoblast) |
24-bit, 2592 × 1944 |
Cancerous |
ALL-IDB-2 | JPEG | 260 (130 lymphoblast) |
24-bit 257 × 257 |
Cancerous |
Algorithm | Control parameters |
GA | Mutation probability, crossover probability and the selection operator. |
ABC | Number of spectator bees, employed bees, scout bees and limit. |
PSO | Inertia weight, social and cognitive parameters. |
Harmony search | Memory consideration rate, pitch adjustment rate and the number of improvisations are all factors in the HS algorithm. |
Name | Image formats | Number of images | Color depth | Remark |
ALL-IDB-1 | JPEG | 109 (510 lymphoblast) |
24-bit, 2592 × 1944 |
Cancerous |
ALL-IDB-2 | JPEG | 260 (130 lymphoblast) |
24-bit 257 × 257 |
Cancerous |