Under the background that Covid-19 is spreading across the world, the lifestyle of people has to confront a series of changes and challenges. This also presents new problems and requirements to automation facilities. For example, nowadays masks have almost become necessities for people in public places. However, most access control systems (ACS) cannot recognize people wearing masks and authenticate their identities to deal with increasingly serious epidemic pressure. Consequently, many public entries have turned to an attendant mode that brings low efficiency, infection potential, and high possibility of negligence. In this paper, a new security classification framework based on face recognition is proposed. This framework uses mask detection algorithm and face authentication algorithm with anti-spoofing function. In order to evaluate the performance of the framework, this paper employs the Chinese Academy of Science Institute of Automation-Face Anti-spoofing Datasets (CASIA-FASD) and Reply-Attack datasets as benchmarks. Performance evaluation indicates that the Half Total Error Rate (HTER) is 9.7%, the Equal Error Rate (EER) is 5.5%. The average process time of a single frame is 0.12 seconds. The results demonstrate that this framework has a high anti-spoofing capability and can be employed on the embedded system to complete the mask detection and face authentication task in real-time.
Citation: Dongzhihan Wang, Guijin Ma, Xiaorui Liu. An intelligent recognition framework of access control system with anti-spoofing function[J]. AIMS Mathematics, 2022, 7(6): 10495-10512. doi: 10.3934/math.2022585
[1] | Lizhen Huang, Qunying Wu . Precise asymptotics for complete integral convergence in the law of the logarithm under the sub-linear expectations. AIMS Mathematics, 2023, 8(4): 8964-8984. doi: 10.3934/math.2023449 |
[2] | Mingzhou Xu, Xuhang Kong . Note on complete convergence and complete moment convergence for negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(4): 8504-8521. doi: 10.3934/math.2023428 |
[3] | Haiye Liang, Feng Sun . Exponential inequalities and a strong law of large numbers for END random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(7): 15585-15599. doi: 10.3934/math.2023795 |
[4] | He Dong, Xili Tan, Yong Zhang . Complete convergence and complete integration convergence for weighted sums of arrays of rowwise m-END under sub-linear expectations space. AIMS Mathematics, 2023, 8(3): 6705-6724. doi: 10.3934/math.2023340 |
[5] | Mingzhou Xu . Complete convergence and complete moment convergence for maximal weighted sums of extended negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(8): 19442-19460. doi: 10.3934/math.2023992 |
[6] | Yi Wu, Duoduo Zhao . Law of the single logarithm for randomly weighted sums of dependent sequences and an application. AIMS Mathematics, 2024, 9(4): 10141-10156. doi: 10.3934/math.2024496 |
[7] | Mingzhou Xu . Complete convergence of moving average processes produced by negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(7): 17067-17080. doi: 10.3934/math.2023871 |
[8] | Li Wang, Qunying Wu . Almost sure convergence theorems for arrays under sub-linear expectations. AIMS Mathematics, 2022, 7(10): 17767-17784. doi: 10.3934/math.2022978 |
[9] | Lunyi Liu, Qunying Wu . Complete integral convergence for weighted sums of negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(9): 22319-22337. doi: 10.3934/math.20231138 |
[10] | Shuyan Li, Qunying Wu . Complete integration convergence for arrays of rowwise extended negatively dependent random variables under the sub-linear expectations. AIMS Mathematics, 2021, 6(11): 12166-12181. doi: 10.3934/math.2021706 |
Under the background that Covid-19 is spreading across the world, the lifestyle of people has to confront a series of changes and challenges. This also presents new problems and requirements to automation facilities. For example, nowadays masks have almost become necessities for people in public places. However, most access control systems (ACS) cannot recognize people wearing masks and authenticate their identities to deal with increasingly serious epidemic pressure. Consequently, many public entries have turned to an attendant mode that brings low efficiency, infection potential, and high possibility of negligence. In this paper, a new security classification framework based on face recognition is proposed. This framework uses mask detection algorithm and face authentication algorithm with anti-spoofing function. In order to evaluate the performance of the framework, this paper employs the Chinese Academy of Science Institute of Automation-Face Anti-spoofing Datasets (CASIA-FASD) and Reply-Attack datasets as benchmarks. Performance evaluation indicates that the Half Total Error Rate (HTER) is 9.7%, the Equal Error Rate (EER) is 5.5%. The average process time of a single frame is 0.12 seconds. The results demonstrate that this framework has a high anti-spoofing capability and can be employed on the embedded system to complete the mask detection and face authentication task in real-time.
Let {X,Xn;n≥1} be a sequence of independent and identically distributed (i.i.d.) random variables. Complete convergence first established by Hsu and Robbins [1] (for the sufficiency) and Erdős [2,3] (for the necessity) proceeds as follows:
∞∑n=1P(|Sn|≥ϵn)<∞,foranyϵ>0, |
if and only if EX=0 and EX2<∞. Baum and Katz [4] extended the above result and obtained the following theorem:
∞∑n=1nr/p−2P(|Sn|≥ϵn1/p)<∞,for0<p<2,r≥p,anyϵ>0, | (1.1) |
if and only if E|X|r<∞, and when r≥1, EX=0.
There are several extensions of the research on complete convergence. One of them is the study of the convergence rate of complete convergence. The first work was the convergence rate, achieved by Heyde [5]. He got the result of limϵ→0ϵ2∑∞n=1P(|Sn|≥ϵn)=EX2 under the conditions EX=0 and EX2<∞. For more results on the convergence rate, see Chen [6], Sp˘ataru [7], Gut and Sp˘ataru [8], Sp˘atarut and Gut [9], Gut and Steinebach [10], He and Xie [11], Kong and Dai [12], etc.
But (1.1) does not hold for p=2. However, by replacing n1/p by √nlnn and √nlnlnn, Gut and Sp˘ataru [8] and Sp˘atarut and Gut [9] established the following results called the convergence rate of the law of the (iterated) logarithm. Supposing that {X,Xn;n≥1} is a sequence of i.i.d. random variables with EX=0 and EX2=σ2<∞, Gut and Sp˘ataru [8] and Sp˘atarut and Gut [9] obtained the following results respectively:
limϵ→0ϵ2+2δ∞∑n=1lnδnnP(|Sn|≥ϵ√nlnn)=E|N|2+2δσ2+2δδ+1,0≤δ≤1, | (1.2) |
where N is the standard normal distribution, and
limϵ→0ϵ2∞∑n=31nlnnP(|Sn|≥ϵ√nlnlnn)=σ2. | (1.3) |
Motivated by the above results, the purpose of this paper is to extend (1.2) and (1.3) to sub-linear expectation space (to be introduced in Section 2), which was introduced by Peng [13,14], and to study the necessary conditions of (1.2).
Under the theoretical framework of the traditional probability space, in order to infer the model, all statistical models must assume that the error (and thus the response variable) is subject to a certain uniquely determined probability distribution, that is, the distribution of the model is deterministic. Classical statistical modeling and statistical inference are based on such distribution certainty or model certainty. "Distribution certainty" modeling has yielded a set of mature theories and methods. However, the real complex data in economic, financial and other fields often have essential and non negligible probability and distribution uncertainty. The probability distribution of the response variable to be studied is uncertain and does not meet the assumptions of classical statistical modeling. Therefore, classical probability statistical modeling methods cannot be used for this type of data modeling. Driven by uncertainty issues, Peng [14,15] established a theoretical framework for sub-linear expectation spaces from the perspective of expectations. Sub-linear expectation has a wide range of application backgrounds and prospects. In recent years, a series of research achievements on limit theory in sub-linear expectation spaces has been established. See Peng [14,15], Zhang [16,17,18], Hu [19], Wu and Jiang [20,21], Wu et al. [22], Wu and Lu [23], etc. Wu [24], Liu and Zhang [25], Ding [26] and Liu and Zhang [27] obtained the convergence rate for complete moment convergence. However, the convergence rate results for the (iterative) logarithmic law have not been reported yet. The main difficulty in studying it is that the sub-linear expectation and capacity are not additive, which makes many traditional probability space tools and methods no longer effective; thus, it is much more complex and difficult to study it.
In Section 2, we will provide the relevant definitions of sub-linear expectation space, the basic properties and the lemmas that need to be used in this paper.
Let (Ω,F) be a measurable space and let H be a linear space of random variables on (Ω,F) such that if X1,…,Xn∈H then φ(X1,…,Xn)∈H for each φ∈Cl,Lip(Rn), where Cl,Lip(Rn) denotes the set of local Lipschitz functions on Rn. In this case, for X∈H, X is called a random variable.
Definition 2.1. A sub-linear expectation ˆE on H is a function: H→R satisfying the following for all X,Y∈H:
(a) Monotonicity: If X≥Y then ˆEX≥ˆEY;
(b) Constant preservation: ˆEc=c;
(c) Sub-additivity: ˆE(X+Y)≤ˆEX+ˆEY;
(d) Positive homogeneity: ˆE(λX)=λˆEX,λ≥0.
The triple (Ω,H,ˆE) is called a sub-linear expectation space. The conjugate expectation ˆε of ˆE is defined by
ˆεX:=−ˆE(−X),∀X∈H. |
Let G⊂F. A function V:G→[0,1] is called a capacity if
V(∅)=0,V(Ω)=1andV(A)≤V(B)for∀A⊆B,A,B∈G. |
The upper and lower capacities (V,ν) corresponding to (Ω,H,ˆE) are respectively defined as
V(A):=inf{ˆEξ;I(A)≤ξ,ξ∈H},ν(A):=1−V(Ac),∀A∈F,Ac:=Ω−A. |
The Choquet integrals is defined by
CV(X):=∫∞0V(X>x)dx+∫0−∞(V(X>x)−1)dx. |
From all of the definitions above, it is easy to obtain the following Proposition 2.1.
Proposition 2.1. Let X,Y∈H and A,B∈F.
(i) ˆεX≤ˆEX,ˆE(X+a)=ˆEX+a,∀a∈R;
(ii) |ˆE(X−Y)|≤ˆE|X−Y|,ˆE(X−Y)≥ˆEX−ˆEY;
(iii) ν(A)≤V(A),V(A∪B)≤V(A)+V(B),ν(A∪B)≤ν(A)+V(B);
(iv) If f≤I(A)≤g,f,g∈H, then
ˆEf≤V(A)≤ˆEg,ˆεf≤ν(A)≤ˆεg. | (2.1) |
(v)(Lemma 4.5 (iii) in Zhang [16]) For any c>0,
ˆE(|X|∧c)≤∫c0V(|X|>x)dx≤CV(|X|), | (2.2) |
where, here and hereafter, a∧b:=min(a,b), and a∨b:=max(a,b) for any a,b∈R.
(vi)Markov inequality: V(|X|≥x)≤ˆE(|X|p)/xp,∀x>0,p>0;
Jensen inequality: (ˆE(|X|r))1/r≤(ˆE(|X|s))1/sfor0<r≤s.
Definition 2.2. (Peng [14,15])
(ⅰ) (Identical distribution) Let X1 and X2 be two random variables on (Ω,H,ˆE). They are called identically distributed, denoted by X1d=X2, if
ˆE(φ(X1))=ˆE(φ(X2)),forallφ∈Cl,Lip(Rn). |
A sequence {Xn;n≥1} of random variables is said to be identically distributed if for each i≥1, Xid=X1.
(ⅱ) (Independence) In a sub-linear expectation space (Ω,H,ˆE), a random vector Y=(Y1,…,Yn), Yi∈H is said to be independent of another random vector X=(X1,…,Xm),Xi∈H under ˆE if for each φ∈Cl,Lip(Rm×Rn), there is ˆE(φ(X,Y))=ˆE[ˆE(φ(x,Y))|x=X].
(ⅲ) (Independent and identically distributed) A sequence {Xn;n≥1} of random variables is said to be i.i.d., if Xi+1 is independent of (X1,…,Xi) and Xid=X1 for each i≥1.
From Definition 2.2 (ⅱ), it can be verified that if Y is independent of X, and X≥0,ˆEY≥0, then ˆE(XY)=ˆE(X)ˆE(Y). Further, if Y is independent of X and X,Y≥0, then
ˆE(XY)=ˆE(X)ˆE(Y),ˆε(XY)=ˆε(X)ˆε(Y). | (2.3) |
For convenience, in all subsequent parts of this article, let {X,Xn;n≥1} be a sequence of random variables in (Ω,H,ˆE), and Sn=∑ni=1Xi. For any X∈H and c>0, set X(c):=(−c)∨X∧c. The symbol c represents a positive constant that does not depend on n. Let ax∼bx denote limx→∞ax/bx=1, ax≪bx denote that there exists a constant c>0 such that ax≤cbx for sufficiently large x, [x] denote the largest integer not exceeding x, and I(⋅) denote an indicator function.
To prove the main results of this article, the following three lemmas are required.
Lemma 2.1. (Theorem 3.1 (a) and Corollary 3.2 (b) in Zhang [16]) Let {Xk;k≥1} be a sequence of independent random variables in (Ω,H,ˆE).
(i) If ˆEXk≤0, then for any x,y>0,
V(Sn≥x)≤V(max1≤k≤nXk>y)+exp(−x22(xy+Bn){1+23ln(1+xyBn)}); |
(ii) If ˆεXk≤0, then there exists a constant c>0 such that for any x>0,
ν(Sn≥x)≤cBnx2, |
where Bn=∑nk=1ˆEX2k.
Here we give the notations of a G-normal distribution which was introduced by Peng [14].
Definition 2.3. (G-normal random variable) For 0≤σ_2≤ˉσ2<∞, a random variable ξ in (Ω,H,ˆE) is called a G-normal N(0,[σ_2,ˉσ2]) distributed random variable (write ξ∼N(0,[σ_2,ˉσ2]) under ˆE), if for any φ∈Cl,Lip(R), the function u(x,t)=ˆE(φ(x+√tξ)) (x∈R,t≥0) is the unique viscosity solution of the following heat equation:
∂tu−G(∂2xxu)=0,u(0,x)=φ(x), |
where G(α)=(ˉσ2α+−σ_2α−)/2.
From Peng [14], if ξ∼N(0,[σ_2,ˉσ2]) under ˆE, then for each convex function φ,
ˆE(φ(ξ))=1√2π∫∞−∞φ(ˉσx)e−x2/2dx. | (2.4) |
If σ=ˉσ=σ_, then N(0,[σ_2,ˉσ2])=N(0,σ2) which is a classical normal distribution.
In particular, notice that φ(x)=|x|p,p≥1 is a convex function, taking φ(x)=|x|p,p≥1 in (2.4), we get
ˆE(|ξ|p)=2ˉσp√2π∫∞0xpe−x2/2dx<∞. | (2.5) |
Equation (2.5) implies that
CV(|ξ|p)=∫∞0V(|ξ|p>x)dx≤1+∫∞1ˆE(|ξ|2p)x2dx<∞,foranyp≥1/2. |
Lemma 2.2. (Theorem 4.2 in Zhang [17], Corollary 2.1 in Zhang [18]) Let {X,Xn;n≥1} be a sequence of i.i.d. random variables in (Ω,H,ˆE). Suppose that
(i)limc→∞ˆE(X2∧c) is finite;
(ii)x2V(|X|≥x)→0 as x→∞;
(iii)limc→∞ˆE(X(c))=limc→∞ˆE((−X)(c))=0.
Then for any bounded continuous function φ,
limn→∞ˆE(φ(Sn√n))=ˆE(φ(ξ)), |
and if F(x):=V(|ξ|≥x), then
limn→∞V(|Sn|>x√n)=F(x),ifxisacontinuouspointofF, | (2.6) |
where ξ∼N(0,[σ_2,ˉσ2]) under ˆE, ˉσ2=limc→∞ˆE(X2∧c) and σ_2=limc→∞ˆε(X2∧c).
Lemma 2.3. (Lemma 2.1 in Zhang [17]) Let {Xn;n≥1} be a sequence of independent random variables in (Ω,H,ˆE), and 0<α<1 be a real number. If there exist real constants βn,k such that
V(|Sn−Sk|≥βn,k+ϵ)≤α,forallϵ>0k≤n, |
then
(1−α)V(maxk≤n(|Sk|−βn,k)>x+ϵ)≤V(|Sn|>x),forallx>0,ϵ>0. |
The results of this article are as follows.
Theorem 3.1. Let {X,Xn;n≥1} be a sequence of i.i.d. random variables in (Ω,H,ˆE). Suppose that
CV(X2)<∞,limc→∞ˆE(X(c))=limc→∞ˆE((−X)(c))=0. | (3.1) |
Then for 0≤δ≤1,
limϵ→0ϵ2+2δ∞∑n=2lnδnnV(|Sn|≥ϵ√nlnn)=CV(|ξ|2δ+2)δ+1, | (3.2) |
where, here and hereafter, ξ∼N(0,[σ_2,ˉσ2]) under ˆE, ˉσ2=limc→∞ˆE(X2∧c) and σ_2=limc→∞ˆε(X2∧c).
Conversely, if (3.2) holds for δ=1, then (3.1) holds.
Theorem 3.2. Under the conditions of Theorem 3.1,
limϵ→0ϵ2∞∑n=31nlnnV(|Sn|≥ϵ√nlnlnn)=CV(ξ2). | (3.3) |
Remark 3.1. Theorems 3.1 and 3.2 not only extend Theorem 3 in [8] and Theorem 2 in [9], respectively, from the probability space to sub-linear expectation space, but they also study and obtain necessary conditions for Theorem 3.1.
Remark 3.2. Under the condition limc→∞ˆE(|X|−c)+=0 (limc→∞ˆE(X2−c)+=0⇒limc→∞ˆE(|X|−c)+=0), it is easy to verify that ˆE(±X)=limc→∞ˆE((±X)(c)). So, Corollary 3.9 in Ding [26] has two more conditions than Theorem 3.2: ˆE is continuous and limc→∞ˆE(X2−c)+=0. Therefore, Corollary 3.9 in Ding [26] and Theorem 3.2 cannot be inferred from each other.
Proof of the direct part of Theorem 3.1.. Note that
ϵ2+2δ∞∑n=2lnδnnV(|Sn|≥ϵ√nlnn)=ϵ2+2δ∞∑n=2lnδnnV(|ξ|≥ϵ√lnn)+ϵ2+2δ∞∑n=2lnδnn(V(|Sn|≥ϵ√nlnn)−V(|ξ|≥ϵ√lnn)):=I1(ϵ)+I2(ϵ). |
Hence, in order to establish (3.2), it suffices to prove that
limϵ→0I1(ϵ)=CV(|ξ|2δ+2)δ+1 | (3.4) |
and
limϵ→0I2(ϵ)=0. | (3.5) |
Given that lnδnn and V(|ξ|≥ε√lnn) is monotonically decreasing with respect to n, it holds that
I1(ϵ)=ϵ2+2δ∞∑n=2lnδnnV(|ξ|≥ϵ√lnn)=ϵ2+2δlnδ22V(|ξ|≥ϵ√ln2)+ϵ2+2δ∞∑n=3∫nn−1lnδnnV(|ξ|≥ϵ√lnn)dx≤ϵ2+2δlnδ22+ϵ2+2δ∞∑n=3∫nn−1lnδxxV(|ξ|≥ϵ√lnx)dx=ϵ2+2δlnδ22+ϵ2+2δ∫∞2lnδxxV(|ξ|≥ϵ√lnx)dx, |
and
I1(ϵ)=ϵ2+2δ∞∑n=2lnδnnV(|ξ|≥ϵ√lnn)=ϵ2+2δ∞∑n=2∫n+1nlnδnnV(|ξ|≥ϵ√lnn)dx≥ϵ2+2δ∞∑n=2∫n+1nlnδxxV(|ξ|≥ϵ√lnx)dx=ϵ2+2δ∫∞2lnδxxV(|ξ|≥ϵ√lnx)dx. |
Therefore, (3.4) follows from
limϵ→0I1(ϵ)=limϵ→0ϵ2+2δ∫∞2lnδxxV(|ξ|≥ϵ√lnx)dx=limϵ→0∫∞ε√ln22y2δ+1V(|ξ|≥y)dy(lety=ε√lnx)=∫∞02y2δ+1V(|ξ|≥y)dy=CV(|ξ|2+2δ)δ+1. |
Let M≥40; write AM,ϵ:=exp(Mϵ−2).
|I2(ϵ)|≤ϵ2+2δ∑2≤n≤[AM,ϵ]lnδnn|V(|Sn|√n≥ϵ√lnn)−V(|ξ|≥ϵ√lnn)|+ϵ2+2δ∑n>[AM,ϵ]lnδnnV(|Sn|≥ϵ√nlnn)+ϵ2+2δ∑n>[AM,ϵ]lnδnnV(|ξ|≥ϵ√lnn):=I21(ϵ)+I22(ϵ)+I23(ϵ). | (3.6) |
Let us first estimate I21(ϵ). For any β>ϵ2,
I21(ϵ)∼ϵ2+2δ∫AM,ϵ2lnδxx|V(|S[x]|√[x]≥ϵ√lnx)−V(|ξ|≥ϵ√lnx)|dx≤ϵ2+2δ∫Aβ,ϵ22lnδxxdx+ϵ2+2δ∫AM,ϵAβ,ϵlnδxxsupn≥Aβ,ϵ|V(|Sn|√n≥ϵ√lnx)−V(|ξ|≥ϵ√lnx)|dx≤2β1+δ+∫√M02y1+2δsupn≥Aβ,ϵ|V(|Sn|√n≥y)−F(y)|dy. | (3.7) |
By (2.2), ˆE(X2∧c)≤∫c0V(X2≥x)dx; also, notice that V(X2≥x) is a decreasing function of x. So, CV(X2)=∫∞0V(X2≥x)dx<∞ implies that limc→∞ˆE(X2∧c) is finite and limx→∞x2V(|X|≥x)=limx→∞xV(X2≥x)=0. Therefore, (3.1) implies the conditions of Lemma 2.2. From (2.6),
limϵ→0supn≥Aβ,ϵ|V(|Sn|√n≥y)−F(y)|=0,ifyisacontinuouspointofF. | (3.8) |
Note that F(y) is a monotonically decreasing function, so its discontinuous points are countable. Hence (3.8) holds for each y, except on a set with the null Lebesgue measure. Combining y2δ+1supn≥Aβ,ϵ|V(|Sn|√n≥y)−F(y)|≤2Mδ+1/2 for any 0≤y≤√M, by the Lebesgue bounded convergence theorem, (3.8) leads to the following:
limϵ→0∫√M0y2δ+1supn≥Aβ,ϵ|V(|Sn|√n≥y)−F(y)|dy=0. | (3.9) |
Let ϵ→0 first, then let β→0; from (3.7) and (3.9), we get
limϵ→0I21(ϵ)=0. | (3.10) |
Next, we estimate that I22(ϵ). For 0<μ<1, let φμ(x)∈Cl,Lip(R) be an even function such that 0≤φμ(x)≤1 for all x and φμ(x)=0 if |x|≤μ and φμ(x)=1 if |x|>1. Then
I(|x|≥1)≤φμ(x)≤I(|x|≥μ). | (3.11) |
Given (2.1) and (3.11), and that X,Xi are identically distributed, for any x>0 and 0<μ<1, we get
V(|Xi|≥x)≤ˆE[φμ(Xix)]=ˆE[φμ(Xx)]≤V(|X|≥μx). | (3.12) |
Without loss of generality, we assume that ˉσ=1. For n≥exp(Mϵ−2)≥exp(40ϵ−2), set bn:=ϵ√nlnn/20; from Proposition 2.1 (ii) and the condition that limc→∞ˆE(X(c))=0,
n∑i=1|ˆEX(bn)i|=n|limc→∞ˆE(X(c))−ˆEX(bn)|≤nlimc→∞ˆE|X(c)−X(bn)|=nlimc→∞ˆE(|X|∧c−bn)+≤nlimc→∞ˆE(|X|∧c)2bn=nˉσ2bn=20√nϵ√lnn≤ϵ2√nlnn,forM≥40,n≥exp(Mϵ−2). |
Using Lemma 2.1 for {X(bn)i−ˆEX(bn)i;1≤i≤n}, and taking x=ϵ√nlnn/2 and y=2bn=ϵ√nlnn/10 in Lemma 2.1 (ⅰ), by Proposition 2.1 (ⅰ), ˆE(X(bn)i−ˆEX(bn)i)=0, and noting that |X(bn)i−ˆEX(bn)i|≤y, Bn=n∑i=1ˆE(X(bn)i−ˆEX(bn)i)2≤4nˆE(X(bn)i)2≤4n; combining this with (3.12) we get
V(Sn≥ϵ√nlnn)≤V(n∑i=1(X(bn)i−ˆEX(bn)i)≥ϵ√nlnn/2)+n∑i=1V(|Xi|≥bn)≤exp(−ϵ2nlnn4(ϵ2nlnn/20+4n){1+23lnϵ2nlnn80n})+nV(|X|≥μbn)≤c(ϵ2lnn)−3+nV(|X|≥μϵ√nlnn/20) |
from ϵ2nlnn4(ϵ2nlnn/20+4n){1+23ln(1+ϵ2lnn80)}≥3ln(ϵ2lnn80).
Since {−X,−Xi} also satisfies the (3.1), we can replace the {X,Xi} with {−X,−Xi} in the upper form
V(−Sn≥ϵ√nlnn)≤c(ϵ2lnn)−3+nV(|X|≥μϵ√nlnn/20). |
Therefore
V(|Sn|≥ϵ√nlnn)≪(ϵ2lnn)−3+nV(|X|≥cϵ√nlnn). |
This implies the following from Markov's inequality and (2.5),
I22(ϵ)+I23(ϵ)≪ϵ2+2δ∑n≥AM,ϵlnδnn(nV(|X|≥cϵ√nlnn)+1ϵ6ln3n+ˆE|ξ|6ϵ6ln3n)∼ϵ2+2δ∫∞AM,ϵlnδxV(|X|≥cϵ√xlnx)dx+cϵ−4+2δ∫∞AM,ϵdxxln3−δx≤ϵ2+2δ∫∞Mϵ−12δyln1−δyV(|X|≥cϵy)dy+cM−2+δ≪ϵ2+2δ∫∞Mϵ−1yV(|X|≥ϵy)dy+M−2+δ≤ϵ2δ∫∞0zV(|X|≥z)dz+M−2+δ=ϵ2δCV(X2)/2+M−2+δ. |
Let ϵ→0 first, then let M→∞; we get
limϵ→0(I22(ϵ)+I23(ϵ))=0. |
Combining this with (3.10) and (3.6), (3.5) is established.
Proof of the converse part of Theorem 3.1. If (3.2) holds for δ=1, then
∞∑n=2lnnnV(|Sn|≥ϵ√nlnn)<∞foranyϵ>0. | (3.13) |
Take ξ as defined by Lemma 2.2 (ˆE|ξ|<∞ from (2.5)) and the bounded continuous function ψ such that I(x>qˆE|ξ|+1)≤ψ(x)≤I(x>qˆE|ξ|) for any fixed q>0. Therefore, for any ϵ>0,q>0 and n≥exp(qˆE|ξ|+1ϵ)2, according to (2.1), Lemma 2.2 and the Markov inequality, one has
V(|Sn|≥ϵ√nlnn)≤V(|Sn|≥(qˆE|ξ|+1)√n)≤ˆE(ψ(|Sn|√n))→ˆE(ψ(|ξ|))≤V(|ξ|>qˆE|ξ|)≤ˆE|ξ|qˆE|ξ|=1q. |
From the arbitrariness of q, letting q→∞, we get the following for any ϵ>0,
V(|Sn|≥ϵ√nlnn)→0,n→∞. | (3.14) |
So, there is an n0 such that V(|Sn|≥ϵ√nlnn)<1/4 for n≥n0. Now for n≥2n0, if k≤n/2, then n−k≥n/2≥n0, and, combining this with (2.1), (3.11) and (3.12), we get that,
V(|Sn−Sk|≥2ϵ√nlnn)≤ˆE(φ1/2(|Sn−Sk|2ϵ√nlnn))=ˆE(φ1/2(|Sn−k|2ϵ√nlnn))≤V(|Sn−k|≥ϵ√(n−k)ln(n−k))<1/2. |
Also, if n/2<k≤n, then n,k≥n/2≥n0; thus,
V(|Sn−Sk|≥2ϵ√nlnn)≤V(|Sn|≥ϵ√nlnn)+V(|Sk|≥ϵ√nlnn≥ϵ√klnk)<1/2. |
Taking α=1/2,βn,k=0 in Lemma 2.3, for n≥2n0,
V(maxk≤n|Sk|≥4ϵ√nlnn)≤V(|Sn|≥2ϵ√nlnn). |
Since maxk≤n|Xk|≤2maxk≤n|Sk|, it follows that for n≥2n0
V(maxk≤n|Xk|≥8ϵ√nlnn)≤V(|Sn|≥2ϵ√nlnn). | (3.15) |
Let Yk=φ8/9(Xk9ϵ√nlnn). Then,
I(maxk≤n|Xk|≥8ϵ√nlnn)=1−I(maxk≤n|Xk|<8ϵ√nlnn)=1−n∏k=1I(|Xk|<8ϵ√nlnn)≥1−n∏k=1(1−Yk). |
Since {Xk;k≥1} is a sequence of i.i.d. random variables, {1−Yk;k≥1} is also a sequence of i.i.d. random variables, and 1−Yk≥0; given (2.1), (2.3) and ˆE(−X)=−ˆε(X), it can be concluded that,
V(maxk≤n|Xk|≥8ϵ√nlnn)≥ˆE(1−n∏k=1(1−Yk))=1−ˆε(n∏k=1(1−Yk))=1−n∏k=1ˆε(1−Yk)=1−n∏k=1(1−ˆEYk)≥1−n∏k=1e−ˆEYk=1−e−nˆEY≥1−e−nV(|X|≥9ϵ√nlnn)∼nV(|X|≥9ϵ√nlnn). |
Hence, by (3.15) and (3.13)
∞∑n=2lnnV(|X|≥√nlnn)<∞. |
On the other hand,
∞∑n=2lnnV(|X|≥√nlnn)∼∫∞2lnxV(|X|≥√xlnx)dx∼∫∞√2ln22yV(|X|≥y)dy∼CV(X2). |
Hence,
CV(X2)<∞. | (3.16) |
Next, we prove that limc→∞ˆE(X(c))=limc→∞ˆE((−X)(c))=0. For c1>c2>0, by (2.2) and (3.16),
|ˆE(±X)(c1)−ˆE(±X)(c2)|≤ˆE|(±X)(c1)−(±X)(c2)|=ˆE(|X|∧c1−c2)+≤ˆE(|X|∧c1)2c2≤CV(X2)c2≪1c2. |
This implies that
limc1>c2→∞|ˆE(±X)(c1)−ˆE(±X)(c2)|=0. |
By the Cauchy criterion, limc→∞ˆE(X(c)) and limc→∞ˆE((−X)(c)) exist and are finite. It follows that limc→∞ˆE(X(c))=limn→∞ˆE(X(n)):=a. So, for any ϵ>0, when n is large enough, |ˆE(X(n))−a|<ϵ; by Proposition 2.1 (iii), Lemma 2.1 (ii), ˆE(−X(n)k+ˆEX(n)k)2≤4ˆE(X(n)k)2≤4CV(X2) and (3.16),
ν(Snn<a−2ϵ)≤ν((Snn<a−2ϵ,∀1≤k≤n,|Xk|≤n)∪(∃1≤k≤n,|Xk|>n))≤ν(n∑k=1X(n)k<(a−2ϵ)n)+n∑k=1V(|Xk|>n)=ν(n∑k=1(−X(n)k+ˆEX(n)k)>(2ϵ−a)n+nEX(n))+n∑k=1V(|Xk|>n)≤ν(n∑k=1(−X(n)k+ˆEX(n)k)>ϵn)+n∑k=1V(|Xk|>n)≪n∑k=1ˆE(−X(n)k+ˆEX(n)k)2n2+n∑k=1ˆE(|Xk|∧n)2n2≪1n→0,n→∞. |
It is concluded that,
limn→∞V(Snn≥a−2ϵ)=1foranyϵ>0. |
If a>0, taking ϵ<a/2, then ϵ1:=a−2ϵ>0, and
limn→∞V(|Sn|n≥ϵ1)≥limn→∞V(Snn≥ϵ1)=1. | (3.17) |
On the other hand, by (3.14),
limn→∞V(|Sn|n≥ϵ1)≤limn→∞V(|Sn|≥ϵ1√nlnn)=0, |
which contradicts (3.17). It follows that a≤0. Similarly, we can prove that b:=limc→∞ˆE((−X)(c))≤0. From (−X)(c)=−X(c) and
0≥a+b=limc→∞(ˆE(X(c))+ˆE(−X(c)))≥limc→∞ˆE(X(c)−X(c))=0, |
we conclude that a=b=0, i.e., limc→∞ˆE(X(c))=limc→∞ˆE((−X)(c))=0. This completes the proof of Theorem 3.1.
Proof of Theorem 3.2. Note that
ϵ2∞∑n=31nlnnV(|Sn|≥ϵ√nlnlnn)=ϵ2∞∑n=31nlnnV(|ξ|≥ϵ√lnlnn)+ϵ2∞∑n=31nlnn(V(|Sn|≥ϵ√nlnlnn)−V(|ξ|≥ϵ√lnlnn)):=J1(ϵ)+J2(ϵ). |
Hence, in order to establish (3.3), it suffices to prove that
limϵ→0J1(ϵ)=CV(ξ2) | (3.18) |
and
limϵ→0J2(ϵ)=0. | (3.19) |
Obviously, (3.18) follows from
limϵ→0J1(ϵ)=limϵ→0ϵ2∫∞31xlnxV(|ξ|≥ϵ√lnlnx)dx=limϵ→0∫∞ε√lnln32yV(|ξ|≥y)dy(lety=ε√lnlnx)=∫∞02yV(|ξ|≥y)dy=CV(ξ2). |
Let M≥32; write BM,ϵ:=exp(exp(Mϵ−2)).
|J2(ϵ)|≤ϵ2∑3≤n≤[BM,ϵ]1nlnn|V(|Sn|√n≥ϵ√lnlnn)−V(|ξ|≥ϵ√lnlnn)|+ϵ2∑n>[BM,ϵ]1nlnnV(|Sn|≥ϵ√nlnlnn)+ϵ2∑n>[BM,ϵ]1nlnnV(|ξ|≥ϵ√lnlnn):=J21(ϵ)+J22(ϵ)+J23(ϵ). | (3.20) |
Let us first estimate J21(ϵ). For any β>ϵ2,
I21(ϵ)∼ϵ2∫BM,ϵ31xlnx|V(|S[x]|√[x]≥ϵ√lnlnx)−V(|ξ|≥ϵ√lnlnx)|dx≤ϵ2∫Bβ,ϵ32xlnxdx+ϵ2∫BM,ϵBβ,ϵ1xlnxsupn≥Bβ,ϵ|V(|Sn|√n≥ϵ√lnlnx)−V(|ξ|≥ϵ√lnlnx)|dx≤2β+∫√M02ysupn≥Bβ,ϵ|V(|Sn|√n≥y)−F(y)|dy. |
Similar to (3.9) we have
limϵ→0∫√M0ysupn≥Bβ,ϵ|V(|Sn|√n≥y)−F(y)|dy=0. |
Therefore, let ϵ→0 first, then let β→0; we get
limϵ→0J21(ϵ)=0. | (3.21) |
Next, we estimate that J22(ϵ). Without loss of generality, we still assume that ˉσ=1. For n≥exp(exp(Mϵ−2))≥exp(exp(32ϵ−2)), set an:=ϵ√nlnlnn/16; from Proposition 2.1 (ii) and the condition that limc→∞ˆE(X(c))=0,
n∑i=1|ˆEX(an)i|=n|limc→∞ˆE(X(c))−ˆEX(an)|≤nlimc→∞ˆE|X(c)−X(an)|=nlimc→∞ˆE(|X|∧c−an)+≤nlimc→∞ˆE(|X|∧c)2an=nˉσ2an=16√nϵ√lnlnn≤ϵ2√nlnlnn. |
Using Lemma 2.1 for {X(an)i−ˆEX(an)i;1≤i≤n}, and taking x=ϵ√nlnlnn/2 and y=2an=ϵ√nlnlnn/8 in Lemma 2.1 (i), if we note that |X(an)i−ˆEX(an)i|≤y, and Bn≤4n, combined with (3.12) we get
V(Sn≥ϵ√nlnlnn)≤V(n∑i=1(X(an)i−ˆEX(an)i)≥ϵ√nlnlnn/2)+n∑i=1V(|Xi|≥an)≤exp(−ϵ2nlnlnn4(ϵ2nlnlnn/16+4n){1+23lnϵ2nlnlnn64n})+nV(|X|≥μan)≤c(ϵ2lnlnn)−2+nV(|X|≥μϵ√nlnlnn/16) |
from ϵ2nlnlnn4(ϵ2nlnlnn/16+4n){1+23ln(1+ϵ2lnlnn64)}≥2ln(ϵ2lnlnn64).
Since {−X,−Xi} also satisfies (3.1), we can replace the {X,Xi} with {−X,−Xi} in the upper form
V(−Sn≥ϵ√nlnlnn)≤c(ϵ2lnlnn)−2+nV(|X|≥μϵ√nlnlnn/16). |
Therefore
V(|Sn|≥ϵ√nlnlnn)≪(ϵ2lnlnn)−2+nV(|X|≥cϵ√nlnlnn). |
This implies the following from Markov's inequality and (2.5):
J22(ϵ)+J23(ϵ)≪ϵ2∑n≥BM,ϵ1nlnn(nV(|X|≥cϵ√nlnlnn)+1ϵ4(lnlnn)2+ˆE|ξ|4ϵ4(lnlnn)2)∼ϵ2∫∞BM,ϵV(|X|≥cϵ√xlnlnx)lnxdx+cϵ−2∫∞BM,ϵdxxlnx(lnlnx)2≤ϵ2∫∞√Mϵ−1ylnylnlnyV(|X|≥cϵy)dy+cM−1≤∫∞√MzV(|X|≥z)dz+cM−1→0,M→∞. |
Hence
limϵ→0(J22(ϵ)+J23(ϵ))=0. |
Combining this with (3.20) and (3.21), (3.19) is established.
Statistical modeling is one of the key and basic topics in statistical theory research and practical application research. Under the theoretical framework of traditional probability space, in order to infer the model, all statistical models must assume that the error (and therefore the response variable) follows a unique and deterministic probability distribution, that is, the distribution of the model is deterministic. However, complex data in the fields of economics, finance, and other fields often have inherent and non negligible probability and distribution uncertainties. The probability distribution of the response variables that need to be studied is uncertain and does not meet the assumptions of classical statistical modeling. Therefore, classical probability statistical modeling methods cannot be used to model these types of data. How to analyze and model uncertain random data has been an unresolved and challenging issue that has long plagued statisticians. Driven by uncertainty issues, Peng [13] established a theoretical framework for the sub-linear expectation space from the perspective of expectations, providing a powerful tool for analyzing uncertainty problems. The sub-linear expectation has a wide range of potential applications. In recent years, the limit theory for sub-linear expectation spaces has attracted much attention from statisticians, and a series of research results have been achieved. This article overcomes the problem of many traditional probability space tools and methods no longer being effective due to the non additivity of sub-linear expectations and capacity; it also demonstrates the development of sufficient and necessary conditions for the rate convergence of logarithmic laws in sub-linear expectation spaces.
The authors declare that they have not used artificial intelligence tools in the creation of this article.
This paper was supported by the National Natural Science Foundation of China (12061028) and Guangxi Colleges and Universities Key Laboratory of Applied Statistics.
Regarding this article, the author claims no conflict of interest.
[1] |
B. Qin, D. Li, Identifying facemask-wearing condition using image super-resolution with classification network to prevent COVID-19, Sensors, 10 (2020), 5236. https://doi.org/10.3390/s20185236 doi: 10.3390/s20185236
![]() |
[2] |
M. S. Ejaz, M. R. Islam, M. Sifatullah, A. Sarker, Implementation of principal component analysis on masked and non-masked face recognition, 2019 1st Int. Conf. Adv. Sci., Eng. Rob. Technol. (ICASERT), 2019, 1–5. https://doi.org/10.1109/ICASERT.2019.8934543 doi: 10.1109/ICASERT.2019.8934543
![]() |
[3] | M. Jiang, X. Fan, H. Yan, Retinamask: A face mask detector, arXiv, unpublished work. |
[4] |
J. Hosang, R. Benenson, B. Schiele, Learning non-maximum suppression, 2017 IEEE Conf. Comput. Vision Pattern Recognit. (CVPR), 2017, 4507–4515. https://doi.org/10.1109/CVPR.2017.685 doi: 10.1109/CVPR.2017.685
![]() |
[5] | S. Woo, J. Park, J. Lee, I. Kweon, Cbam: Convolutional block attention module, Proc. Eur. Conf. Comput. Vision (ECCV), 2018, 3–19. |
[6] | Y. Taigman, M. Yang, M. A. Ranzato, L. Wolf, Deepface: Closing the gap to human-level performance in face verification, Proc. IEEE Conf. Comput. Vision Pattern Recognit. (CVPR), 2014, 1701–1708. |
[7] | Y. Sun, X. Wang, X. Tang, Deep learning face representation from predicting 10,000 classes, Proc. IEEE Conf. Comput. Vision Pattern Recognit. (CVPR), 2014, 1891–1898. |
[8] |
D. Nguyen, K. Nguyen, S. Sridharan, D. Dean, C. Fookes, Deep spatio-temporal feature fusion with compact bilinear pooling for multimodal emotion recognition, Comput. Vis. Image Und., 174 (2018), 33–42. https://doi.org/10.1016/j.cviu.2018.06.005 doi: 10.1016/j.cviu.2018.06.005
![]() |
[9] | J. Deng, J. Guo, N. Xue, S. Zafeiriou, Arcface: Additive angular margin loss for deep face recognition, Proc. IEEE/CVF Conf. Comput. Vision Pattern Recognit. (CVPR), 2019, 4690–4699. |
[10] | H. Liu, X. Zhu, Z. Lei, S. Z. Li, Adaptiveface: Adaptive margin and sampling for face recognition, Proc. IEEE/CVF Conf. Comput. Vision Pattern Recognit. (CVPR), 2019, 11947–11956. |
[11] |
Y. Jiang, W. Li, M. S. Hossain, M. Chen, A. Alelaiwi, M. Al-Hammadi, A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition, Inform. Fusion, 53 (2019), 145–156. https://doi.org/10.1016/j.inffus.2019.06.019 doi: 10.1016/j.inffus.2019.06.019
![]() |
[12] | Y. Huang, Y. Wang, Y. Tai, X. Liu, P. Shen, S. Li, et al., Curricularface: Adaptive curriculum learning loss for deep face recognition, Proc. IEEE/CVF Conf. Comput. Vision Pattern Recognit. (CVPR), 2020, 5901–5910. |
[13] |
Z. Boulkenafet, J. Komulainen, A. Hadid, Face anti-spoofing based on color texture analysis, 2015 IEEE Int. Conf. Image Proc. (ICIP), 2015, 2636–2640. https://doi.org/10.1109/ICIP.2015.7351280 doi: 10.1109/ICIP.2015.7351280
![]() |
[14] |
Z. Boulkenafet, J. Komulainen, A. Hadid, Face spoofing detection using colour texture analysis, IEEE T. Inf. Forensics Secur., 11 (2016), 1818–1830. https://doi.org/10.1109/TIFS.2016.2555286 doi: 10.1109/TIFS.2016.2555286
![]() |
[15] |
X. Li, J. Komulainen, G. Zhao, P. C. Yuen, M. Pietikäinen, Generalized face anti-spoofing by detecting pulse from face videos, 2016 23rd Int. Conf. Pattern Recognit. (ICPR), 2016, 4244–4249. https://doi.org/10.1109/ICPR.2016.7900300 doi: 10.1109/ICPR.2016.7900300
![]() |
[16] | I. Chingovska, N. Erdogmus, A. Anjos, S. Marcel, Face recognition systems under spoofing attacks, In: T. Bourlai, Face recognition across the imaging spectrum, Springer, 2016, 165–194. https://doi.org/10.1007/978-3-319-28501-6_8 |
[17] | S. Q. Liu, X. Lan, P. C. Yuen, Remote photoplethysmography correspondence feature for 3D mask face presentation attack detection, Proc. Eur. Conf. Comput. Vision (ECCV), 2018,558–573. |
[18] |
I. Manjani, S. Tariyal, M. Vatsa, R. Singh, A. Majumdar, Detecting silicone mask-based presentation attack via deep dictionary learning, IEEE T. Inf. Forensics Secur., 2017, 1713–1723. https://doi.org/10.1109/TIFS.2017.2676720 doi: 10.1109/TIFS.2017.2676720
![]() |
[19] |
R. Shao, X. Lan, P. C. Yuen, Joint discriminative learning of deep dynamic textures for 3d mask face anti-spoofing, IEEE T. Inf. Forensics Secur., 14 (2018), 923–938. https://doi.org/10.1109/TIFS.2018.2868230 doi: 10.1109/TIFS.2018.2868230
![]() |
[20] |
J. Määttä, A. Hadid, M. Pietikäinen, Face spoofing detection from single images using micro-texture analysis, 2011 Int. Joint Conf. Biometrics (IJCB), 2011, 1–7. https://doi.org/10.1109/IJCB.2011.6117510 doi: 10.1109/IJCB.2011.6117510
![]() |
[21] |
J. Määttä, A. Hadid, M. Pietikäinen, Face spoofing detection from single images using texture and local shape analysis, IET Biom., 1 (2012), 3–10. https://doi.org/10.1049/iet-bmt.2011.0009 doi: 10.1049/iet-bmt.2011.0009
![]() |
[22] |
Y. Atoum, Y. Liu, A. Jourabloo, X. Liu, Face anti-spoofing using patch and depth-based CNNs, 2017 IEEE International Joint Conference on Biom. (IJCB), 2017,319–328. https://doi.org/10.1109/BTAS.2017.8272713 doi: 10.1109/BTAS.2017.8272713
![]() |
[23] | Y. Liu, A. Jourabloo, X. Liu, Learning deep models for face anti-spoofing: Binary or auxiliary supervision, Proc. IEEE Conf. Comput. Vision Pattern Recognit. (CVPR), 2018,389–398. |
[24] |
G. Pan, L. Sun, Z. Wu, S. Lao, Eyeblink-based anti-spoofing in face recognition from a generic web camera, 2007 IEEE 11th Int. Conf. Comput. Vision, 2007, 1–8. https://doi.org/10.1109/ICCV.2007.4409068 doi: 10.1109/ICCV.2007.4409068
![]() |
[25] | A. Zadeh, P. P. Liang, N. Mazumder, S. Poria, E. Cambria, L. P. Morency, Memory fusion network for multi-view sequential learning, Thirty-Second AAAI Conf. Artif. Intell., 32 (2018), 5642–5649. |
[26] |
T. Baltruaitis, C. Ahuja, L. P. Morency, Multimodal machine learning: A survey and taxonomy, IEEE T. Pattern Anal. Mach. Intell., 41 (2019), 154–163. https://doi.org/10.1109/TPAMI.2018.2798607 doi: 10.1109/TPAMI.2018.2798607
![]() |
[27] | T. Li, Q. Yang, S. Rong, L. Chen, B. He, Distorted underwater image reconstruction for an autonomous underwater vehicle based on a self-attention generative adversarial network, Appl. Opt., 59 (2020), 10049–10060. |
[28] |
T. Li, S. Rong, X. Cao, Y. Liu, L. Chen, B. He, Underwater image enhancement framework and its application on an autonomous underwater vehicle platform, Opt. Eng., 59 (2020), 083102. https://doi.org/10.1117/1.OE.59.8.083102 doi: 10.1117/1.OE.59.8.083102
![]() |
[29] |
K. Zhang, Z. Zhang, Z. Li, Y. Qiao, Joint face detection and alignment using multitask cascaded convolutional networks, IEEE Signal Proc. Let., 23 (2016), 1499–1503. https://doi.org/10.1109/LSP.2016.2603342 doi: 10.1109/LSP.2016.2603342
![]() |
[30] | J. Fu, J. Liu, H. Tian, Y. Li, Y. Bao, Z. Fang, et al., Dual attention network for scene segmentation, Proc. IEEE/CVF Conf. Comput. Vision Pattern Recognit. (CVPR), 2019, 3146–3154. |
[31] |
W. Sanghyun, H. Soonmin, I. S. Kweon, Stairnet: Top-down semantic aggregation for accurate one-shot detection, 2018 IEEE Winter Conf. Appl. Comput. Vision (WACV), 2018, 1093–1102. https://doi.org/10.1109/WACV.2018.00125 doi: 10.1109/WACV.2018.00125
![]() |
[32] | T. Ojala, M. Pietikäinen, T. Mäenpää, Gray scale and rotation invariant texture classification with local binary patterns, In: Computer Vision-ECCV 2000, Lecture Notes in Computer Science, Springer, 1842 (2000), 404–420. https://doi.org/10.1007/3-540-45054-8_27 |
[33] |
T. Ojala, M. Pietikainen, T. Maenpaa, Multiresolution gray-scale and rotation invariant texture classification with local binary patterns, IEEE T. Pattern Anal. Mach. Intell., 24 (2002), 971–987. https://doi.org/10.1109/TPAMI.2002.1017623 doi: 10.1109/TPAMI.2002.1017623
![]() |
[34] |
W. S. Noble, What is a support vector machine? Nat. Biotechnol., 24 (2006), 1565–1567. https://doi.org/10.1038/nbt1206-1565 doi: 10.1038/nbt1206-1565
![]() |
[35] | F. N. Iandola, S. Han, M. W. Moskewicz, K. Ashraf, W. J. Dally, K. Keutzer, SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and < 0.5 MB model size, arXiv, unpublished work. |
[36] | F. Chollet, Xception: Deep learning with depthwise separable convolutions, Proc. IEEE Conf. Comput. Vision Pattern Recognit. (CVPR), 2017, 1251–1258. |
[37] | M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, L. C. Chen, Mobilenetv2: Inverted residuals and linear bottlenecks, Proc. IEEE Conf. Comput. Vision Pattern Recognit. (CVPR), 2018, 4510–4520. |
[38] | A. Howard, M. Sandler, G. Chu, L. C. Chen, B. Chen, M. Tan, et al., Searching for mobilenetv3, Proc. IEEE/CVF Int. Conf. Comput. Vision (ICCV), 2019, 1314–1324. |
[39] | N. Ma, X. Zhang, H. Zheng, J. Sun, Shufflenet v2: Practical guidelines for efficient CNN architecture design, Proc. Eur. Conf. Comput. Vision (ECCV), 2018,116–131. |
[40] |
Z. Zhang, J. Yan, S. Liu, Z. Lei, D. Yi, S. Z. Li, A face antispoofing database with diverse attacks, 2012 5th IAPR Int. Conf. Biom. (ICB), 2012, 26–31. https://doi.org/10.1109/ICB.2012.6199754 doi: 10.1109/ICB.2012.6199754
![]() |
[41] |
A. Costa-Pazo, S. Bhattacharjee, E. Vazquez-Fernandez, Sebastien Marcel, The replay-mobile face presentation-attack database, 2016 Int. Conf. Biom. Spec. Interest Group (BIOSIG), 2016, 1–7. https://doi.org/10.1109/BIOSIG.2016.7736936 doi: 10.1109/BIOSIG.2016.7736936
![]() |
1. | Mengmei Xi, Fei Zhang, Xuejun Wang, Complete Moment Convergence and Lq Convergence for AANA Random Variables Under Sub-linear Expectations, 2024, 47, 0126-6705, 10.1007/s40840-023-01636-6 |