Loading [MathJax]/jax/output/SVG/jax.js
Research article

DSCA-Net: A depthwise separable convolutional neural network with attention mechanism for medical image segmentation


  • Received: 27 April 2022 Revised: 18 September 2022 Accepted: 28 September 2022 Published: 08 October 2022
  • Accurate segmentation is a basic and crucial step for medical image processing and analysis. In the last few years, U-Net, and its variants, have become widely adopted models in medical image segmentation tasks. However, the multiple training parameters of these models determines high computation complexity, which is impractical for further applications. In this paper, by introducing depthwise separable convolution and attention mechanism into U-shaped architecture, we propose a novel lightweight neural network (DSCA-Net) for medical image segmentation. Three attention modules are created to improve its segmentation performance. Firstly, Pooling Attention (PA) module is utilized to reduce the loss of consecutive down-sampling operations. Secondly, for capturing critical context information, based on attention mechanism and convolution operation, we propose Context Attention (CA) module instead of concatenation operations. Finally, Multiscale Edge Attention (MEA) module is used to emphasize multi-level representative scale edge features for final prediction. The number of parameters in our network is 2.2 M, which is 71.6% less than U-Net. Experiment results across four public datasets show the potential and the dice coefficients are improved by 5.49% for ISIC 2018, 4.28% for thyroid, 1.61% for lung and 9.31% for nuclei compared with U-Net.

    Citation: Tong Shan, Jiayong Yan, Xiaoyao Cui, Lijian Xie. DSCA-Net: A depthwise separable convolutional neural network with attention mechanism for medical image segmentation[J]. Mathematical Biosciences and Engineering, 2023, 20(1): 365-382. doi: 10.3934/mbe.2023017

    Related Papers:

    [1] B. El-Sobky, G. Ashry . An interior-point trust-region algorithm to solve a nonlinear bilevel programming problem. AIMS Mathematics, 2022, 7(4): 5534-5562. doi: 10.3934/math.2022307
    [2] M Junaid Basha, S Nandhini . A fuzzy based solution to multiple objective LPP. AIMS Mathematics, 2023, 8(4): 7714-7730. doi: 10.3934/math.2023387
    [3] B. El-Sobky, G. Ashry, Y. Abo-Elnaga . An active-set with barrier method and trust-region mechanism to solve a nonlinear Bilevel programming problem. AIMS Mathematics, 2022, 7(9): 16112-16146. doi: 10.3934/math.2022882
    [4] Jiaxin Chen, Yuxuan Wu, Shuai Huang, Pei Wang . Multi-objective optimization for AGV energy efficient scheduling problem with customer satisfaction. AIMS Mathematics, 2023, 8(9): 20097-20124. doi: 10.3934/math.20231024
    [5] B. El-Sobky, Y. Abo-Elnaga, G. Ashry, M. Zidan . A nonmonton active interior point trust region algorithm based on CHKS smoothing function for solving nonlinear bilevel programming problems. AIMS Mathematics, 2024, 9(3): 6528-6554. doi: 10.3934/math.2024318
    [6] Muhammad Akram, Syed Muhammad Umer Shah, Mohammed M. Ali Al-Shamiri, S. A. Edalatpanah . Extended DEA method for solving multi-objective transportation problem with Fermatean fuzzy sets. AIMS Mathematics, 2023, 8(1): 924-961. doi: 10.3934/math.2023045
    [7] Sema Akin Bas, Beyza Ahlatcioglu Ozkok . An iterative approach for the solution of fully fuzzy linear fractional programming problems via fuzzy multi-objective optimization. AIMS Mathematics, 2024, 9(6): 15361-15384. doi: 10.3934/math.2024746
    [8] Shima Soleimani Manesh, Mansour Saraj, Mahmood Alizadeh, Maryam Momeni . On robust weakly ε-efficient solutions for multi-objective fractional programming problems under data uncertainty. AIMS Mathematics, 2022, 7(2): 2331-2347. doi: 10.3934/math.2022132
    [9] Habibe Sadeghi, Maryam Esmaeili . A modification of the trilevel Kth-Best algorithm. AIMS Mathematics, 2018, 3(4): 524-538. doi: 10.3934/Math.2018.4.524
    [10] Shima Soleimani Manesh, Mansour Saraj, Mahmood Alizadeh, Maryam Momeni . Correction: On robust weakly ε-efficient solutions for multi-objective fractional programming problems under data uncertainty. AIMS Mathematics, 2022, 7(5): 7462-7463. doi: 10.3934/math.2022417
  • Accurate segmentation is a basic and crucial step for medical image processing and analysis. In the last few years, U-Net, and its variants, have become widely adopted models in medical image segmentation tasks. However, the multiple training parameters of these models determines high computation complexity, which is impractical for further applications. In this paper, by introducing depthwise separable convolution and attention mechanism into U-shaped architecture, we propose a novel lightweight neural network (DSCA-Net) for medical image segmentation. Three attention modules are created to improve its segmentation performance. Firstly, Pooling Attention (PA) module is utilized to reduce the loss of consecutive down-sampling operations. Secondly, for capturing critical context information, based on attention mechanism and convolution operation, we propose Context Attention (CA) module instead of concatenation operations. Finally, Multiscale Edge Attention (MEA) module is used to emphasize multi-level representative scale edge features for final prediction. The number of parameters in our network is 2.2 M, which is 71.6% less than U-Net. Experiment results across four public datasets show the potential and the dice coefficients are improved by 5.49% for ISIC 2018, 4.28% for thyroid, 1.61% for lung and 9.31% for nuclei compared with U-Net.



    Multilevel programming (MLP) is developed to model the decentralized decision-making situations wherein decision makers are arranged within a hierarchical structure. In a MLP problem, when only two levels exist, the problem is referred to as the bilevel programming (BLP) problem. In actual, BLP problems involve two nested optimization problems where the constraint region of first-level problem contains the second-level optimization problem. In this model, a decision maker at the first-level is termed as a leader, and at the second-level, a follower. BLP problems have many applications in the real world, such as those in supply chain management [14], planning [16], and logistics [19]. Most of the research concerning MLP have focused on bilevel programming (BLP) [1,5]. In this paper, we consider linear bilevel problems with one leader and multiple followers. This model is called linear bilevel multi-follower programming (LBLMFP) problem. Shi et al. extended the Kth-Best algorithm to solve LBLMFP problems [21]. Several researchers developed heuristic algorithms to the BLP problems, for instance, see [10]. Lu et al, applied the extended Kuhn-Tucker approach to solve LBLMFP problem [13]. Also, one approach is developed to solve the LBLMFP problems based on multi-parametric programming [7]. In addition, some researchers have exploited multiple objective programming techniques to solve BLP problems [9,18]. These kinds of solution approaches are developed based on some relationships between BLP and multiple objective programming that were first presented by F¨ul¨op [8]. For example, Glackin et al. presented an algorithm to solve linear bilevel programming (LBLP) problems based on multiple objective linear programming (MOLP) techniques [9].

    In this paper, Motivated by the relationship between LBLP and MOLP, a relationship between a class of LBLMFP problems and MOLP problems is introduced for the first time. Moreover, we present an algorithm for solving LBLMFP problems based on the proposed relation.

    The paper is organized as follows: In the next section, we present the basic definitions, and some notions about LBLMFP problems and MOLP problems. In Section 3, we introduce two MOLP problems such that each feasible solution for the LBLMFP problem is efficient for both of the two MOLP problems. Next, we obtain results, based on which the LBLMFP problem can be reduced to optimize the first-level objective function over a certain efficient set. Based on this result, a new algorithm is developed in Section 4 for solving the LBLMFP problem. Furthermore, we obtain some results for special cases. Section 5 presents a number of numerical examples to illustrate the proposed algorithm. Finally, in Section 6, we present the conclusions.

    In this section, we present the formulation of LBLMFP problem which we shall investigate, accompanied by basic definitions. In addition, we recall some notions and theoretical properties about MOLP problems.

    In this article, we consider a linear bilevel programming (LBLP) problem with two followers. There are individual decision variables in separate objective functions and constraints between the followers. Each follower takes other followers' decision variables as a reference. This is called a reference-uncooperative LBLMFP problem [23]. This model can be formulated as follows:

    minx1X1F(x1,x2,x3)=(c11)Tx1+(c12)Tx2+(c13)Tx3,minx2X2f1(x2,x3)=(c22)Tx2+(c23)Tx3,s.t.A1x1+A2x2+A3x3b1,minx3X3f2(x2,x3)=(c32)Tx2+(c33)Tx3,s.t.B1x1+B2x2+B3x3b2,

    where xiXiRni+,i=1,2,3, F:X1×X2×X3R, and fk:X2×X3R,k=1,2, cij, for i,j{1,2,3}, are the vectors of conformal dimension, and for each i{1,2,3}, Ai is a m×ni matrix, Bi is a q×ni matrix, b1Rm, and b2Rq. Also, although xi0, for i=1,2,3, do not explicitly appear in this problem, we assume that they exist in the set of constraints.

    Notice that, the followers' objective functions are linear in x2 and x3 and for each follower, the value for the variable x1 is given. Thus, a problem equivalent to the LBLMFP problem is obtained if the followers' objective functions are replaced by 3j=1(cij)Txj, for i = 2, 3.

    We need to introduce the following definitions that can be found in [23].

    (1) Constraint region of the LBLMFP problem:

    S:={(x1,x2,x3)X1×X2×X3:A1x1+A2x2+A3x3b1,B1x1+B2x2+B3x3b2}.

    We suppose that S is non-empty and compact.

    (2) Projection of S onto the leader's decision space:

    S(X1):={x1X1:(x2,x3)X2×X3,(x1,x2,x3)S}.

    (3) Feasible sets for the first and second followers, respectively:

    S1(x1,x3):={x2X2:(x1,x2,x3)S},
    S2(x1,x2):={x3X3:(x1,x2,x3)S}.

    The feasible region of each follower is affected by the leader's choice of x1, and the other followers' decisions.

    (4) The first and second followers' rational reaction sets, respectively:

    P1(x1,x3):={x2:x2argmin[f1(ˆx2,x3):ˆx2S1(x1,x3)]},

    where

    argmin[f1(ˆx2,x3):ˆx2S1(x1,x3)]={x2S1(x1,x3):f1(x2,x3)f1(ˆx2,x3),ˆx2S1(x1,x3)},

    and

    P2(x1,x2):={x3:x3argmin[f2(x2,ˆx3):ˆx3S2(x1,x2)]},

    where

    argmin[f2(x2,ˆx3):ˆx3S2(x1,x2)]={x3S2(x1,x2):f2(x2,x3)f2(x2,ˆx3),ˆx3S2(x1,x2)}.

    (5) Inducible region (IR):

    IR:={(x1,x2,x3)S:x2P1(x1,x3),x3P2(x1,x2)}.

    In actual, the inducible region is the set of feasible solutions of the LBLMFP problem. Therefore, determining the solution for the LBLMFP problem is equivalent to solving the following problem:

    min{F(x1,x2,x3):(x1,x2,x3)IR}. (2.1)

    In order to assure that (2.1) has an optimal solution [23], we supposed that the following assumptions hold:

    (1) IR is non-empty.

    (2) P1(x1,x3) and P2(x1,x2) are point-to-point mappings with respect to (x1,x3) and (x1,x2), respectively. In other words, for each leader's choice of x1, there will be a unique solution to each follower.

    Assume that p2 is an integer and ciRn,i=1,2,,p are row vectors. Let C be a p×n matrix whose i-th row is given by ci,i=1,2,,p, and U is a non-empty, compact and convex polyhedral set in Rn. A multiple objective linear programming (MOLP) problem is formulated in general as follows:

    min{Cx:xU},(MOLP)

    where U is called feasible region.

    Definition 1. [22] A feasible point, ˉxU, is called an efficient solution if there exists no xU such that CxCˉx and CxCˉx.

    An efficient solution is also called a Pareto-optimal solution.

    Let E denote the set of all efficient solutions of the MOLP problem. Note that E [[6], Theorem 2.19].

    Let dRn. Consider the following mathematical programming problem:

    min{dTx:xE}. (2.2)

    This problem is a non-convex linear optimization problem over the efficient set E of the MOLP problem. Let E denote the set of optimal solutions to the problem (2.2) and Uex denote the set of extreme points of the polyhedron U. Because U is a non-empty, compact and convex polyhedron in Rn, Uex is non-empty [[3], Theorem 2.6.5]. So, the following result holds:

    Theorem 1. Let U be a non-empty, compact, and convex polyhedral set, and let E be non-empty. Then, at least one element of E is an extreme point of U [[4], Theorem 4.5].

    By considering the relation EU, we immediately get the following corollary to Theorem 1.

    Corollary 1. Let E. Then, there is an extreme point of E, which is an optimal solution to the problem (2.2).

    From Theorem 3.40 in [6], because U is a non-empty, compact and convex polyhedron in Rn, and the objective functions cix,i=1,2,,p are convex, we can conclude that the efficient set E is connected.

    Moreover, it is well-known that the efficient set E, is comprised of a finite union of faces of U [6]. Therefore, we obtain the following corollary:

    Corollary 2. The efficient set E is comprised of a finite union of connected faces of the polyhedron U.

    The Corollary 2 obviously implies that the efficient set E is closed.

    Also, the following definition is used in sequel:

    Definition 2. [22] Let URn, and x1,x2,,xrU. The notation γ(x1,x2,,xr) denotes the set of all convex combinations of x1,x2,,xr.

    In this section, we will introduce two MOLP problems in such a way that any extreme efficient solution for both problems simultaneously, is an extreme feasible solution (in other words, an extreme point of IR) to LBLMFP problem.

    Note that ˉx=(ˉx1,ˉx2,ˉx3)IR, i.e., ˉx2P1(ˉx1,ˉx3) and ˉx3P2(ˉx1,ˉx2). Therefore, ˉx2 is an optimal solution for the following linear program (LP):

    minx2X2f1(x2,ˉx3)=(c22)Tx2+(c23)Tˉx3, (3.1)
    s.t.A2x2b1A1ˉx1A3ˉx3, (3.2)

    and ˉx3 is an optimal solution for the following LP:

    minx3X3f2(ˉx2,x3)=(c32)Tˉx2+(c33)Tx3, (3.3)
    s.t.B3x3b2B1ˉx1B2ˉx2. (3.4)

    Now, set n:=3i=1ni. We construct augmented matrix Aa1=[A1A3]. Let rank(Aa1)=r1. Without loss of generality, we assume that augmented matrix Aa1 can be rewritten as [ˉA1ˉA3ˆA1ˆA3], where [ˉA1ˉA3] is r1×(n1+n3) matrix and rank([ˉA1ˉA3])=rank(Aa1)=r1. [Notice that the augmented matrix Aa1 contains an (n1+n3)×(n1+n3) identity submatrix corresponding to non-negativity constraints. So, r1=n1+n3]. Similarly, we construct augmented matrix Ba1=[B1B2]. Let rank(Ba1)=r2 and augmented matrix Ba1 can be rewritten as [ˉB1ˉB2ˆB1ˆB2], where [ˉB1ˉB2] is r2×(n1+n2) matrix and rank([ˉB1ˉB2])=rank(Ba1)=r2. Let ki=ri+2,i=1,2 and the ki×n, criterion matrices Ci,i=1,2 be defined as follows:

    C1:=[ˉA1OˉA3eTˉA10eTˉA30(c22)T0],C2:=[ˉB1ˉB2OeTˉB1eTˉB2000(c33)T], (3.5)

    where O and 0, are zero matrix and zero vectors of conformal dimension, respectively. Also, e is a vector having each entry equal to 1.

    Next, consider the following two MOLP problems:

    min{Cix:xS},i=1,2.(MOLPi)

    Let Ei be the set of efficient solutions of the MOLPi, i=1,2. We have the following result:

    Proposition 1. Let IR and E1 be defined as above. Then, IRE1.

    Proof. Assume that ˉx=(ˉx1,ˉx2,ˉx3)IR. It suffices to show that ˉx is an efficient solution for the MOLP1. Let us suppose the contrary, i.e., there exists x=(x1,x2,x3)S such that C1xC1ˉx and C1xC1ˉx. Because C1xC1ˉx, using the structure of matrix C1, one obtains:

    ˉA1x1+ˉA3x3ˉA1ˉx1+ˉA3ˉx3, (3.6)
    eTˉA1x1eTˉA3x3eTˉA1ˉx1eTˉA3ˉx3. (3.7)

    From inequality (3.6), one obtains:

    eTˉA1x1eTˉA3x3eTˉA1ˉx1eTˉA3ˉx3. (3.8)

    The last two inequalities imply that:

    eTˉA1x1+eTˉA3x3=eTˉA1ˉx1+eTˉA3ˉx3. (3.9)

    From equality (3.9) and inequality (3.6), it is easy to see that ˉA1x1+ˉA3x3=ˉA1ˉx1+ˉA3ˉx3. So, we have:

    A1x1+A3x3=A1ˉx1+A3ˉx3. (3.10)

    Obviously, x2 is a feasible solution for the problem (3.1)- (3.2). Moreover, due to C1xC1ˉx and in view of equalities (3.9) and (3.10), we deduce that (c22)Tx2<(c22)Tˉx2 which contradicts ˉx2 is an optimal solution for the problem (3.1)– (3.2). This completes the proof.

    The proof of the following proposition is similar to that of Proposition 1, and we omit it.

    Proposition 2. Let IR and E2 be defined as above. Then, IRE2.

    Now, if we set E:=E1E2, from Propositions 1 and 2, we can get:

    Remark 1. Let IR,E1, and E2 be defined as above. Then, IRE1E2=E, and regarding to Assumption (1), E is a non-empty set.

    Therefore, this allows us to prove the main result of this section.

    Theorem 2. The extreme points of IR and E are identical.

    Proof. Let x=(x1,x2,x3)IRex be arbitrary. Then, x is an extreme point of S [[23], Corollary 4.9]. Because IRE, we get xE. Therefore, x is an extreme point of E (Note that ES). Since xIRex was chosen arbitrarily, we obtain IRexEex. Now, we will show that EexIRex. Let x=(x1,x2,x3)Eex be chosen arbitrarly. There are two cases to be considered here.

    Case 1: xIR. In this case, we conclude from the definition of IR that at least one of the following two subcases should occur:

    Subcase 1: x2P1(x1,x3). It follows from Assumption (2) that there exists ˆx2P1(x1,x3), so that f1(ˆx2,x3)<f1(x2,x3). As a consequence, we get:

    (c22)Tˆx2<(c22)Tx2. (3.11)

    Set ˆx=(x1,ˆx2,x3). By definition of P1(x1,x3), ˆxS. Due to the structure of matrix C1 and inequality (3.11), one has C1ˆxC1x,C1ˆxC1x that contradicts xEE1.

    Subcase 2: x3P2(x1,x2). By Assumption (2), there exists ˆx3P2(x1,x2) so that f2(x2,ˆx3)<f2(x2,x3). As a consequence, we have:

    (c33)Tˆx3<(c33)Tx3. (3.12)

    Set ˆx=(x1,x2,ˆx3). By definition of P2(x1,x2), ˆxS. Due to the structure of matrix C2 and inequality (3.12), one has C2ˆxC2x,C2ˆxC2x, which contradicts xEE2.

    Case 2: xIR. Since xEex and IRE (see Remark 1), we conclude that xIRex. This completes the proof.

    Now, let us consider the following problem:

    min{F(x1,x2,x3)=(c11)Tx1+(c12)Tx2+(c13)Tx3:(x1,x2,x3)E}. (3.13)

    Remark 2. Note that Corollary 2 obviously implies that the efficient set E is closed. Also, since E is a closed subset of the compact polyhedral set of S, itself is a compact set. Therefore, problem (3.13) involves the optimization of a linear function over a compact set. Hence, there exists an optimal solution to the problem (3.13), i.e., E. Then, according to the Corollary 1, there is an extreme point of E, which is an optimal solution for the problem (3.13).

    The relation between LBLMFP problem and problem (3.13) is stated in the following theorem.

    Theorem 3. A point ˉx=(ˉx1,ˉx2,ˉx3)IRex is an optimal solution to LBLMFP problem if, and only if, it is an optimal solution to the problem (3.13).

    Proof. It is well-known that solving LBLMFP problem is equivalent to solving problem (2.1). Furthermore, there exists an extreme point of IR, which is an optimal solution for problem (2.1) [23]. Because the optimal solution of problem (3.13) occurs at an extreme point of E and the extreme points of E and IR are the same, we conclude that if ˉx is an optimal solution for problem (3.13), it is an optimal solution for LBLMFP problem, and vice versa. This completes the proof.

    In the next section, based on Theorem 3, we propose a new algorithm to solve LBLMFP problem.

    Owing to the preceding discussion, we will propose a new algorithm for solving LBLMFP problems. First, one can solve two MOLPi, i=1,2 in order to obtain the sets Ei,i=1,2. Let E be the set of efficient points that have been discovered to be efficient solutions to both problems. Then, one can minimize the leader objective function over the set S. If the obtained extreme optimal solution lies in the set E, it solves the LBLMFP problem. Otherwise, one can solve problem (3.13). Then, by Theorem 3, the obtained extreme optimal solution solves the LBLMFP problem. The algorithm can be described as follows:

    The Algorithm:

    Step 1. Construct the MOLPi, i=1,2.

    Step 2. Find efficient sets Ei with respect to the MOLPi, i=1,2.

    For instance, approaches presented in [6,17,20,22] can be used in Step 2.

    Step 3. Set E=E1E2.

    Step 4. Solve the following LP:

    min{F(x1,x2,x3)=(c11)Tx1+(c12)Tx2+(c13)Tx3:(x1,x2,x3)S}. (4.1)

    Let x be an optimal solution (One can use the common known solution methods for LPs in [2]). If xE, stop. Then, x is an optimal solution to LBLMFP problem. Otherwise, go to Step 5.

    Step 5. Find an optimal solution to problem (3.13) using, for instance, approaches developed in [4,11,12,15]. Let x be an optimal solution for problem (3.13). Then, it is an optimal solution to LBLMFP problem, too.

    Notice that, by Remark 2, there exists an extreme point of E which is an optimal solution for problem (3.13). Hence, in Step 5, for a few number of variables and extreme points, we can find an optimal solution by picking the minimum objective function value among all extreme points of E.

    Note that because the efficient set E is not convex in general, the problem (3.13) is a non-convex optimization problem. Thus, in Step 5, we face a non-convex optimization problem that is generally difficult to solve. Moreover, the algorithm stops by obtaining an efficient optimal solution. The Step 4 solves the LP (4.1) and checks whether xE. If xE, the algorithm stops at Step 4, and does not enter Step 5. Therefore, Step 4 is necessary in this algorithm to reduce the computations in some cases.

    Prior to applying the proposed algorithm for solving numerical examples, we will state the following lemma:

    Lemma 1. Let E be defined as above. Then EconvIR, where convIR be the smallest convex set containing IR.

    Proof. Let ˆxE be chosen arbitrarly. Suppose the contrary, that is ˆxconvIR. Because IR is closed [[23], Theorem 4.9], convIR is a convex and closed set. Therefore, there exists a non-zero vector P [[3], Theorem 2.4.4], such that PTˆx<PTy, for all yconvIR. Since S is a compact set and convIR is a closed subset of S, convIR is a compact set as well. Furthermore, because yconvIR, according to the Representation Theorem [3], we have y=ni=1λixi,ni=1λi=1,0λi1,xi(convIR)ex. On the other hand, (convIR)ex=IRex=Eex, thus,

    PTˆx<PT(ni=1λixi),ni=1λi=1,0λi1,xiEex. (4.2)

    Now, consider the problem min{PTx:xE}. By Remark 2, there exists xjEex, which is an optimal solution for this problem. Then, PTxjPTx, for all xE. Because ˆxE, we have:

    PTxjPTˆx. (4.3)

    Since xjEex, it follows from (4.2) that PTˆx<PTxj, which contradicts (4.3). Hence, ˆxconvIR.

    Note that the efficient set E and the inducible region IR are not convex sets, generally [1,5,6,22]. Up to now, we have proved that IREconvIR. In the following, we show that these inclusions convert to equality in some special cases, as convexity.

    Corollary 3. If the efficient set E is a convex set, then E=convIR.

    Proof. By the convexity of E and from Remark 1, we obtain convIRE. Also, according to Lemma 1, we have EconvIR. Then, E=convIR.

    Corollary 4. If the inducible region IR is a convex set, then E=IR.

    Proof. According to Remark 1, IRE. Because IR is a convex set, convIR=IR. It follows from Lemma 1 that EIR. Then, E=IR.

    In this section, we apply the proposed algorithm for solving some numerical examples.

    Example 1. Consider the following LBLMFP problem:

    minx1F(x1,x2,x3)=3x1+x2+x3,minx2f1(x2,x3)=x2+x3,s.t.x1+x28,x1+4x28,x1+2x213,7x12x20,x10,x20,x30,minx3f2(x2,x3)=2x3,s.t.x10,x20,0x32.

    In this problem, S={(x1,x2,x3):x1+x28,x1+4x28,x1+2x213,7x12x20,x10,x20,0x32}, and the extreme points of S are as follows:

    a1=(8,0,0),a2=(8,0,2),a3=(3,5,0),a4=(3,5,2),a5=(138,9116,0),a6=(138,9116,2),a7=(815,2815,0),a8=(815,2815,2).

    By [7], one can obtain:

    IR={(x1,x2,x3):815x1138, x2=72x1, x3=0}{(x1,x2,x3): 138x13, x2=13212x1, x3=0}{(x1,x2,x3):3x18, x2=8x1, x3=0}.

    The Figure 1 displays S and IR. The inducible region IR is denoted by the hatched lines.

    Figure 1.  Constraint region and IR of Example 1.

    Note that in this example, S and IR are non-empty and S is compact. Also, one can obtain:

    P1(x1,x3)={(x1,x2,x3)S: x2=72x1,815x1138}{(x1,x2,x3)S: x2=13212x1,138x13}{(x1,x2,x3)S: x2=8x1,3x18},P2(x1,x2)={(x1,x2,x3)S: x3=0}.

    Hence, P1(x1,x3), P2(x1,x2), and P1(x1,x3) and P2(x1,x2) are point-to-point maps with respect to (x1,x3) and (x1,x2), respectively. Therefore, the Assumptions (1) and (2) hold, and we can solve this problem by the proposed algorithm. Using the proposed algorithm, the process is as follows:

    Step 1. The MOLPi problems, for i=1,2 are constructed as follows, respectively:

    min{(x1,x3,x1x3,x2):(x1,x2,x3)S},min{(x1,x2,x1x2,2x3):(x1,x2,x3)S}.

    Step 2. In order to find the sets Ei,i=1,2, using the approach described in [20], one can obtain:

    E1=γ(a1,a3,a4,a2)γ(a3,a5,a6,a4)γ(a5,a7,a8,a6),E2=γ(a1,a3,a5,a7).

    The sets Ei,i=1,2 are shown in Figure 2 (a) and (b), by the hatched regions and gray area, respectively.

    Figure 2.  Constraint region, E1 and E2 of Example 1.

    Step 3. Set E=E1E2. We obtain E=γ(a1,a3)γ(a3,a5)γ(a5,a7)=IR which is shown in Figure 1 by the hatched lines.

    Step 4. Solve the following LP:

    min{3x1+x2+x3:(x1,x2,x3)S}.

    The optimal solution is x=(815,2815,0). Because xE, stop. We deduce that x is an optimal solution for this problem. Also, we solve this example using the multi-parametric approach [7]. We obtain x=(815,2815,0) which is equal with the obtained optimal solution using the proposed algorithm.

    Example 2. Consider the following LBLMFP problem:

    minx1F(x1,x2,x3)=3x1+x2x3,minx2f1(x2,x3)=2x23x3,s.t.x1+x21,x10,x20,x30,minx3f2(x2,x3)=4x2+x3,s.t.2x1+x2+x35,x10,x20,x30.

    In this problem, S={(x1,x2,x3): 2x1+x2+x35,x1+x21,x10,x20,x30}, and the extreme points of S are as follows:

    a1=(2.5,0,0),a2=(1,0,0),a3=(0,1,0),a4=(0,5,0),a5=(0,1,4),a6=(1,0,3).

    By [7], one can obtain:

    IR={(x1,x2,x3): 0x11, x2=1x1, x3=0}{(x1,x2,x3): 1x152, x2=0, x3=0}.

    The sets S and IR are drawn in Figure 3. The inducible region IR is denoted by the hatched lines.

    Figure 3.  Constraint region and IR of Example 2.

    Note that in this example, S and IR are non-empty and S is compact. Also, one can obtain:

    P1(x1,x3)={(x1,x2,x3)S: x2=1x1,0x11}{(x1,x2,x3)S: x2=0,1x15212x3},P2(x1,x2)={(x1,x2,x3)S: x3=0}.

    Hence, P1(x1,x3), P2(x1,x2), and P1(x1,x3) and P2(x1,x2) are point-to-point maps with respect to (x1,x3) and (x1,x2), respectively. Therefore, the Assumptions (1) and (2) hold, and we can solve this problem by the proposed algorithm. Applying the proposed algorithm to this example, we have:

    Step 1. The MOLPi problems, for i=1,2 are constructed as follows, respectively:

    min{(x1,x3,x1x3,2x2):(x1,x2,x3)S},min{(x1,x2,x1x2,x3):(x1,x2,x3)S}.

    Step 2. One can obtain the sets Ei,i=1,2, using the approach described in [20]:

    E1=γ(a1,a2,a6)γ(a2,a3,a5,a6),E2=γ(a1,a2,a3,a4).

    The sets Ei,i=1,2 are drawn in Figure 4 (a) and (b) by the hatched regions and gray area, respectively.

    Figure 4.  Constraint region, E1 and E2 of Example 2.

    Step 3. Set E=E1E2. So, we obtain E=γ(a1,a2)γ(a2,a3). Here, E coincides with IR which is shown in Figure 3 by the hatched lines.

    Step 4. Solve the following LP:

    min{3x1+x2x3:(x1,x2,x3)S}.

    The optimal solution is x=(0,1,4). Because xE, go to step 5.

    Step 5. Solve the following problem:

    min{3x1+x2x3:(x1,x2,x3)E}.

    This problem is a non-convex optimization problem. According to Remark 2, we just consider the extreme points of E. Then, we conclude that the point x=(0,1,0) is an optimal solution. Also, one can solve this example using the multi-parametric approach and obtain x=(0,1,0). The optimal solution is equal with the obtained optimal solution using the proposed algorithm.

    Example 3. Consider the following LBLMFP problem:

    minx1F(x1,x2,x3)=x14x2,minx2f1(x2,x3)=3x22x3,s.t.x1+x22,x10,x20,minx3f2(x2,x3)=x2+4x3,s.t.x10,x20,2x34.

    In this problem, S={(x1,x2,x3): x1+x22,x10,x20,2x34}, and the extreme points of S are as follows:

    a1=(2,0,2),a2=(0,0,2),a3=(0,2,2),a4=(0,2,4),a5=(0,0,4),a6=(2,0,4).

    By [7], one can obtain:

    IR={(x1,x2,x3): 0x12, x2=0, x3=2}.

    The Figure 5 displays constraint region S and inducible region IR. The inducible region is denoted by the hatched line. Note that in this example, S and IR are non-empty and S is compact. Also, one can obtain:

    Figure 5.  Constraint region and IR of Example 3.
    P1(x1,x3)={(x1,x2,x3)S: x2=0},P2(x1,x2)={(x1,x2,x3)S: x3=2}.

    Hence, P1(x1,x3), P2(x1,x2), and P1(x1,x3) and P2(x1,x2) are point-to-point maps with respect to (x1,x3) and (x1,x2), respectively. Therefore, the Assumptions (1) and (2) hold, and we can solve this problem by the proposed algorithm. Applying the proposed algorithm to this example, we have:

    Step 1. The MOLPi problems, for i=1,2 are constructed as follows, respectively:

    min{(x1,x3,x1x3,3x2):(x1,x2,x3)S},min{(x1,x2,x1x2,4x3):(x1,x2,x3)S}.

    Step 2. One can obtain the sets Ei,i=1,2, using the approach described in [20]:

    E1=γ(a1,a2,a5,a6),E2=γ(a1,a2,a3).

    The sets Ei,i=1,2 are drawn in Figure 6 (a) and (b) by the gray areas, respectively.

    Figure 6.  Constraint region, E1 and E2 of Example 3.

    Step 3. Set E=E1E2. So, we obtain E=γ(a1,a2)=IR. It is shown in Figure 5 by the hatched line.

    Step 4. Solve the following LP:

    min{x14x2:(x1,x2,x3)S}.

    The extreme optimal solutions are x=(0,2,2) and y=(0,2,4). Because x,yE, go to Step 5.

    Step 5. Solve the following problem:

    min{x14x2:(x1,x2,x3)E}.

    Since E is convex, this problem is the linear programming problem. An optimal solution is x=(2,0,2). Also, one can solve this example using the multi-parametric approach and obtain x=(2,0,2). The obtained optimal solution is equal with the obtained optimal solution using the proposed algorithm.

    In this paper, we have presented a relation between a class of LBLMFP problems and multi-criteria optimization for the first time. We have shown how to construct two MOLP problems so that the extreme points of the set of efficient solutions for both problems are the same as those of the set of feasible solutions for the LBLMFP problem. It is proved that solving the given LBLMFP problem is equivalent to optimizing the leader objective function over a certain efficient set. Based on this result, we proposed an algorithm for solving the LBLMFP problem, and we also showed that it can be simplified in some special cases. Further studies are being conducted to improve the performance of the proposed algorithm and to extend to other classes of LBLMFP problems.

    All authors declare no conflicts of interest in this paper.



    [1] J. C. Caicedo, J. Roth, A. Goodman, T. Becker, K. W. Karhohs, M. Broisin, et al., Evaluation of deep learning strategies for nucleus segmentation in fluorescence images, Cytometry, Part A, J. Quant. Cell Sci., 95 (2019), 952–965. https://doi.org/10.1002/cyto.a.23863 doi: 10.1002/cyto.a.23863
    [2] Y. Fu, Y. Lei, T. Wang, W. J. Curran, T. Liu, X. Yang, A review of deep learning based methods for medical image multi-organ segmentation, Physica Med., 85 (2021), 107–122. https://doi.org/10.1016/j.ejmp.2021.05.003 doi: 10.1016/j.ejmp.2021.05.003
    [3] R. Merjulah, J. Chandra, Segmentation technique for medical image processing: a survey, in 2017 International Conference on Inventive Computing and Informatics (ICICI), (2017), 1055–1061. https://doi.org/10.1109/ICICI.2017.8365301
    [4] Z. Gu, J. Cheng, H. Fu, K. Zhou, H. Hao, Y. Zhao, et al., CE-Net: context encoder network for 2D medical image segmentation, IEEE Trans. Med. Imaging, 38 (2019), 2281–2292. https://doi.org/10.1109/TMI.2019.2903562 doi: 10.1109/TMI.2019.2903562
    [5] N. C. F. Codella, D. Gutman, M. E. Celebi, B. Helba, M. A. Marchetti, S. W. Dusza, et al., Skin lesion analysis toward melanoma detection: a challenge at the 2017 International symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (ISIC), in 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), (2018), 168–172. https://doi.org/10.1109/ISBI.2018.8363547
    [6] H. Yu, L. T. Yang, Q. Zhang, D. Armstrong, M. J. Deen, Convolutional neural networks for medical image analysis: state-of-the-art, comparisons, improvement and perspectives, Neurocomputing, 444 (2021), 92–110. https://doi.org/10.1016/j.neucom.2020.04.157 doi: 10.1016/j.neucom.2020.04.157
    [7] N. Tajbakhsh, L. Jeyaseelan, Q. Li, J. N. Chiang, Z. Wu, X. Ding, Embracing imperfect datasets: a review of deep learning solutions for medical image segmentation, Med. Image Anal., 63 (2020), 101693. https://doi.org/10.1016/j.media.2020.101693 doi: 10.1016/j.media.2020.101693
    [8] V. Badrinarayanan, A. Kendall, R. Cipolla, SegNet: a deep convolutional encoder-decoder architecture for image segmentation, IEEE Trans. Pattern Anal. Mach. Intell., 39 (2017), 2481–2495. https://doi.org/10.1109/TPAMI.2016.2644615 doi: 10.1109/TPAMI.2016.2644615
    [9] J. Long, E. Shelhamer, T. Darrell, Fully convolutional networks for semantic segmentation, in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2015), 3431–3440. https://doi.org/10.1109/CVPR.2015.7298965
    [10] O. Ronneberger, P. Fischer, T. Brox, U-Net: convolutional networks for biomedical image segmentation, in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, (2015), 234–241. https://doi.org/10.1007/978-3-319-24574-4_28
    [11] R. Gu, G. Wang, T. Song, R. Huang, M. Aertsen, J. Deprest, et al., CA-Net: comprehensive attention convolutional neural networks for explainable medical image segmentation, IEEE Trans. Med. Imaging, 40 (2021), 699–711. https://doi.org/10.1109/TMI.2020.3035253 doi: 10.1109/TMI.2020.3035253
    [12] N. Ibtehaz, M. S. Rahman, MultiResUNet: rethinking the U-Net architecture for multimodal biomedical image segmentation, Neural Networks, 121 (2020), 74–87. https://doi.org/10.1016/j.neunet.2019.08.025 doi: 10.1016/j.neunet.2019.08.025
    [13] C. Y. Chang, P. C. Chung, Y. C. Hong, C. H. Tseng, A neural network for thyroid segmentation and volume estimation in CT images, IEEE Comput. Intell. Mag., 6 (2011), 43–55. https://doi.org/10.1109/MCI.2011.942756 doi: 10.1109/MCI.2011.942756
    [14] S. Nandamuri, D. China, P. Mitra, D. Sheet, Sumnet: fully convolutional model for fast segmentation of anatomical structures in ultrasound volumes, in 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), (2019), 1729–1732. https://doi.org/10.1109/ISBI.2019.8759210
    [15] M. Z. Alom, C. Yakopcic, T. M. Taha, V. K. Asari, Recurrent residual convolutional neural network based on U-Net (R2U-Net) for medical image segmentation, in NAECON 2018 - IEEE National Aerospace and Electronics Conference, (2018), 228–233. https://doi.org/10.1109/NAECON.2018.8556686
    [16] Y. Wen, L. Chen, Y. Deng, J. Ning, C. Zhou, Towards better semantic consistency of 2D medical image segmentation, J. Visual Commun. Image Represent., 80 (2021), 103311. https://doi.org/10.1016/j.jvcir.2021.103311 doi: 10.1016/j.jvcir.2021.103311
    [17] P. Naylor, M. Lae, F. Reyal, T. Walter, Nuclei segmentation in histopathology images using deep neural networks, in 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), (2017), 933–936. https://doi.org/10.1109/ISBI.2017.7950669
    [18] H. Noh, S. Hong, B. Han, Learning deconvolution network for semantic segmentation, in 2015 IEEE International Conference on Computer Vision (ICCV), (2015), 1520–1528. https://doi.org/10.1109/ICCV.2015.178
    [19] Q. Kang, Q. Lao, T. Fevens, Nuclei segmentation in histopathological images using two-stage learning, in Medical Image Computing and Computer Assisted Intervention – MICCAI 2019, (2019), 703–711. https://doi.org/10.1007/978-3-030-32239-7_78
    [20] Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, J. Liang, UNet++: a nested U-Net architecture for medical image segmentation, in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, 2018 (2018), 3–11. https://doi.org/10.1007/978-3-030-00889-5_1
    [21] L. Kaiser, A. N. Gomez, F. Chollet, Depthwise separable convolutions for neural machine translation, preprint, arXiv: 1706.03059.
    [22] J. Hu, L. Shen, G. Sun, Squeeze-and-excitation networks, in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2018), 7132–7141. https://doi.org/10.1109/CVPR.2018.00745
    [23] Q. Wang, B. Wu, P. Zhu, P. Li, W. Zuo, Q. Hu, ECA-Net: efficient channel attention for deep convolutional neural networks, in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2020), 11531–11539. https://doi.org/10.1109/CVPR42600.2020.01155
    [24] V. Mnih, N. Heess, A. Graves, K. Kavukcuoglu, Recurrent models of visual attention, in Proceedings of the 27th International Conference on Neural Information Processing Systems, 2 (2014), 2204–2212. https://dl.acm.org/doi/abs/10.5555/2969033.2969073
    [25] D. Bahdanau, K. Cho, Y. Bengio, Neural machine translation by jointly learning to align and translate, preprint, arXiv: 1409.0473.
    [26] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, et al., Attention is all you need, in Advances in Neural Information Processing Systems, 30 (2017), 5998–6008. Available from: https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf.
    [27] O. Oktay, J. Schlemper, L. L. Folgoc, M. Lee, M. Heinrich, K. Misawa, et al., Attention U-Net: learning where to look for the pancreas, preprint, arXiv: 1804.03999.
    [28] F. Chollet, Xception: deep learning with depthwise separable convolutions, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2017), 1800–1807. https://doi.org/10.1109/CVPR.2017.195
    [29] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, et al., Mobilenets: Efficient convolutional neural networks for mobile vision applications, preprint, arXiv: 1704.04861.
    [30] X. Zhang, X. Zhou, M. Lin, J. Sun, Shufflenet: an extremely efficient convolutional neural network for mobile devices, in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2018), 6848–6856. https://doi.org/10.1109/CVPR.2018.00716
    [31] L. C. Chen, Y. Zhu, G. Papandreou, F. Schroff, H. Adam, Encoder-decoder with atrous separable convolution for semantic image segmentation, in Computer Vision – ECCV 2018, (2018), 833–851. https://doi.org/10.1007/978-3-030-01234-2_49
    [32] K. Qi, H. Yang, C. Li, Z. Liu, M. Wang, Q. Liu, et al., X-Net: brain stroke lesion segmentation based on depthwise separable convolution and long-range dependencies, in Medical Image Computing and Computer Assisted Intervention – MICCAI 2019, (2019), 247–255.https://doi.org/10.1007/978-3-030-32248-9_28
    [33] A. Wibowo, S. R. Purnama, P. W. Wirawan, H. Rasyidi, Lightweight encoder-decoder model for automatic skin lesion segmentation, Inf. Med. Unlocked, 25 (2021), 100640. https://doi.org/10.1016/j.imu.2021.100640 doi: 10.1016/j.imu.2021.100640
    [34] C. Meng, K. Sun, S. Guan, Q. Wang, R. Zong, L. Liu, Multiscale dense convolutional neural network for DSA cerebrovascular segmentation, Neurocomputing, 373 (2020), 123–134. https://doi.org/10.1016/j.neucom.2019.10.035 doi: 10.1016/j.neucom.2019.10.035
    [35] G. Huang, Z. Liu, L. V. D. Maaten, K. Q. Weinberger, Densely connected convolutional networks, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2017), 2261–2269. https://doi.org/10.1109/CVPR.2017.243
    [36] Y. Wu, K. He, Group normalization, Int. J. Comput. Vision, 128 (2020), 742–755. https://doi.org/10.1007/s11263-019-01198-w doi: 10.1007/s11263-019-01198-w
    [37] X. Zhang, Y. Zou, W. Shi, Dilated convolution neural network with LeakyReLU for environmental sound classification, in 2017 22nd International Conference on Digital Signal Processing (DSP), (2017), 1–5. https://doi.org/10.1109/ICDSP.2017.8096153
    [38] P. Tschandl, C. Rosendahl, H. Kittler, The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions, Sci. Data, 5 (2018), 180161. https://doi.org/10.1038/sdata.2018.161 doi: 10.1038/sdata.2018.161
    [39] T. Wunderling, B. Golla, P. Poudel, C. Arens, M. Friebe, C. Hansen, Comparison of thyroid segmentation techniques for 3D ultrasound, Med. Imaging 2017: Image Process., 10133 (2017), 1013317. https://doi.org/10.1117/12.2254234 doi: 10.1117/12.2254234
    [40] D. P. Kingma, J. Ba, Adam: a method for stochastic optimization, preprint, arXiv: 1412.6980.
    [41] G. Lin, A. Milan, C. Shen, I. Reid, RefineNet: Multi-path refinement networks for high-resolution semantic segmentation, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2017), 5168–5177. https://doi.org/10.1109/CVPR.2017.549
    [42] R. Ma, S. Zhang, C. Gan, H. Zhao, EOCNet: Improving edge omni-scale convolution networks for skin lesion segmentation, in 2020 3rd International Conference on Digital Medicine and Image Processing, (2020), 45–50. https://doi.org/10.1145/3441369.3441377
    [43] L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, A. L. Yuille, DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, IEEE Trans. Pattern Anal. Mach. Intell., 40 (2018), 834–848. https://doi.org/10.1109/TPAMI.2017.2699184 doi: 10.1109/TPAMI.2017.2699184
    [44] S. Chen, Y. Zou, P. X. Liu, IBA-U-Net: Attentive BConvLSTM U-Net with redesigned inception for medical image segmentation, Comput. Biol. Med., 135 (2021), 104551. https://doi.org/10.1016/j.compbiomed.2021.104551 doi: 10.1016/j.compbiomed.2021.104551
  • This article has been cited by:

    1. Fatemeh Moslemi, Habibeh Sadeghi, On solving uncooperative linear bilevel multi-follower programming problems, 2022, 2220-5810, 1, 10.18187/pjsor.v18i1.3261
    2. Ben Tian, Zhaogang Xie, Wei Peng, Jun Liu, Application Analysis of Multi-Intelligence Optimization Decision-Making Method in College Students’ Ideological and Political Education System, 2022, 2022, 1939-0122, 1, 10.1155/2022/8999757
    3. Francisco Salas-Molina, David Pla-Santamaria, Maria Luisa Vercher-Ferrandiz, Ana Garcia-Bernabeu, New decision rules under strict uncertainty and a general distance-based approach, 2023, 8, 2473-6988, 13257, 10.3934/math.2023670
    4. Olumide F. Odeyinka, Oluwafemi Ipinnimo, Folorunso Ogunwolu, Parametric multi-level modeling of Lean Six Sigma, 2024, 71, 1110-1903, 10.1186/s44147-024-00528-1
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3845) PDF downloads(296) Cited by(12)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog