Processing math: 100%
Research article Special Issues

Audio-visual multi-modality driven hybrid feature learning model for crowd analysis and classification


  • The high pace emergence in advanced software systems, low-cost hardware and decentralized cloud computing technologies have broadened the horizon for vision-based surveillance, monitoring and control. However, complex and inferior feature learning over visual artefacts or video streams, especially under extreme conditions confine majority of the at-hand vision-based crowd analysis and classification systems. Retrieving event-sensitive or crowd-type sensitive spatio-temporal features for the different crowd types under extreme conditions is a highly complex task. Consequently, it results in lower accuracy and hence low reliability that confines existing methods for real-time crowd analysis. Despite numerous efforts in vision-based approaches, the lack of acoustic cues often creates ambiguity in crowd classification. On the other hand, the strategic amalgamation of audio-visual features can enable accurate and reliable crowd analysis and classification. Considering it as motivation, in this research a novel audio-visual multi-modality driven hybrid feature learning model is developed for crowd analysis and classification. In this work, a hybrid feature extraction model was applied to extract deep spatio-temporal features by using Gray-Level Co-occurrence Metrics (GLCM) and AlexNet transferrable learning model. Once extracting the different GLCM features and AlexNet deep features, horizontal concatenation was done to fuse the different feature sets. Similarly, for acoustic feature extraction, the audio samples (from the input video) were processed for static (fixed size) sampling, pre-emphasis, block framing and Hann windowing, followed by acoustic feature extraction like GTCC, GTCC-Delta, GTCC-Delta-Delta, MFCC, Spectral Entropy, Spectral Flux, Spectral Slope and Harmonics to Noise Ratio (HNR). Finally, the extracted audio-visual features were fused to yield a composite multi-modal feature set, which is processed for classification using the random forest ensemble classifier. The multi-class classification yields a crowd-classification accurac12529y of (98.26%), precision (98.89%), sensitivity (94.82%), specificity (95.57%), and F-Measure of 98.84%. The robustness of the proposed multi-modality-based crowd analysis model confirms its suitability towards real-world crowd detection and classification tasks.

    Citation: H. Y. Swathi, G. Shivakumar. Audio-visual multi-modality driven hybrid feature learning model for crowd analysis and classification[J]. Mathematical Biosciences and Engineering, 2023, 20(7): 12529-12561. doi: 10.3934/mbe.2023558

    Related Papers:

    [1] Ze-Miao Dai, Jia-Bao Liu, Kang Wang . Analyzing the normalized Laplacian spectrum and spanning tree of the cross of the derivative of linear networks. AIMS Mathematics, 2024, 9(6): 14594-14617. doi: 10.3934/math.2024710
    [2] Zhi-Yu Shi, Jia-Bao Liu . Topological indices of linear crossed phenylenes with respect to their Laplacian and normalized Laplacian spectrum. AIMS Mathematics, 2024, 9(3): 5431-5450. doi: 10.3934/math.2024262
    [3] Yinzhen Mei, Chengxiao Guo . The minimal degree Kirchhoff index of bicyclic graphs. AIMS Mathematics, 2024, 9(7): 19822-19842. doi: 10.3934/math.2024968
    [4] Ali Al Khabyah . Mathematical aspects and topological properties of two chemical networks. AIMS Mathematics, 2023, 8(2): 4666-4681. doi: 10.3934/math.2023230
    [5] Yuan Shan, Baoqing Liu . Existence and multiplicity of solutions for generalized asymptotically linear Schrödinger-Kirchhoff equations. AIMS Mathematics, 2021, 6(6): 6160-6170. doi: 10.3934/math.2021361
    [6] Giovanni G. Soares, Ernesto Estrada . Navigational bottlenecks in nonconservative diffusion dynamics on networks. AIMS Mathematics, 2024, 9(9): 24297-24325. doi: 10.3934/math.20241182
    [7] Zenan Du, Lihua You, Hechao Liu, Yufei Huang . The Sombor index and coindex of two-trees. AIMS Mathematics, 2023, 8(8): 18982-18994. doi: 10.3934/math.2023967
    [8] Ali N. A. Koam, Ali Ahmad, Yasir Ahmad . Computation of reverse degree-based topological indices of hex-derived networks. AIMS Mathematics, 2021, 6(10): 11330-11345. doi: 10.3934/math.2021658
    [9] Jahfar T K, Chithra A V . Central vertex join and central edge join of two graphs. AIMS Mathematics, 2020, 5(6): 7214-7233. doi: 10.3934/math.2020461
    [10] Yun-Ho Kim, Hyeon Yeol Na . Multiplicity of solutions to non-local problems of Kirchhoff type involving Hardy potential. AIMS Mathematics, 2023, 8(11): 26896-26921. doi: 10.3934/math.20231377
  • The high pace emergence in advanced software systems, low-cost hardware and decentralized cloud computing technologies have broadened the horizon for vision-based surveillance, monitoring and control. However, complex and inferior feature learning over visual artefacts or video streams, especially under extreme conditions confine majority of the at-hand vision-based crowd analysis and classification systems. Retrieving event-sensitive or crowd-type sensitive spatio-temporal features for the different crowd types under extreme conditions is a highly complex task. Consequently, it results in lower accuracy and hence low reliability that confines existing methods for real-time crowd analysis. Despite numerous efforts in vision-based approaches, the lack of acoustic cues often creates ambiguity in crowd classification. On the other hand, the strategic amalgamation of audio-visual features can enable accurate and reliable crowd analysis and classification. Considering it as motivation, in this research a novel audio-visual multi-modality driven hybrid feature learning model is developed for crowd analysis and classification. In this work, a hybrid feature extraction model was applied to extract deep spatio-temporal features by using Gray-Level Co-occurrence Metrics (GLCM) and AlexNet transferrable learning model. Once extracting the different GLCM features and AlexNet deep features, horizontal concatenation was done to fuse the different feature sets. Similarly, for acoustic feature extraction, the audio samples (from the input video) were processed for static (fixed size) sampling, pre-emphasis, block framing and Hann windowing, followed by acoustic feature extraction like GTCC, GTCC-Delta, GTCC-Delta-Delta, MFCC, Spectral Entropy, Spectral Flux, Spectral Slope and Harmonics to Noise Ratio (HNR). Finally, the extracted audio-visual features were fused to yield a composite multi-modal feature set, which is processed for classification using the random forest ensemble classifier. The multi-class classification yields a crowd-classification accurac12529y of (98.26%), precision (98.89%), sensitivity (94.82%), specificity (95.57%), and F-Measure of 98.84%. The robustness of the proposed multi-modality-based crowd analysis model confirms its suitability towards real-world crowd detection and classification tasks.



    Graph categories considered in this study are simple, finite, and linked. Allow the graph G to be made up of VG and edge set EG, i.e., G=(VG,EG). For more graph notations, readers should refer to [1].

    If and only if two neighbouring verticies i and j of G, the adjacency matrix A(G)=(aij) is a (0, 1)-matrix, aij=1. The diagonal degree matrix of G is

    DG=diag(d1,d2,,dn),

    where di represents the degree of vertex i in G. The difference between the degree matrix DG and adjacency matrix AG of G gives rise to the Laplace matrix, denoted as LG. The normalized Laplacian [2] is defined as

    L(G)=ID(G)12(D(G)1A(G))D(G)12=D(G)12L(G)D(G)12.

    The (m,n)th-entry of L(G), which is designated

    (L(G))mn={1,m=n;1dmdn,mn,vmisadjacenttovn;0,otherwise. (1.1)

    The distance between both vi and vj, known as dij=dG(vi,vj), represents the length of the smallest path in question. Wiener and Dobrynin [3,4] introduced the Wiener index for the first time in 1947. In addition, the Wiener index is denoted as

    W(G)=i<jdij.

    For further information on the Wiener index, please refer to [5,6,7,8,9].

    The Gutman index of a simple graph G is introduced [10] and denoted as

    Gut(G)=i<jdidjdij,

    taking into account the degree di of vertex vi.

    The Kirchhoff index[11,12] characterizes graph G by summing the resistance distances between every pair of vertices, similar to the Wiener index, namely

    Kf(G)=i<jrij.

    The multiplicative degree-Kirchhoff index, initially introduced by Chen and Zhang[13] in 2007. It is an extension of the traditional Kirchhoff index, which is expressed as

    Kf(G)=i<jdidjrij.

    The techniques of the Kirchhoff index and multiplication degree-Kirchhoff index can be found in [14,15,16,17,18]. The multiplication degree-Kirchhoff index has garnered significant attention due to its remarkable contributions in academia and practical applications in computer network science, epidemiology, social economics, and other fields. Further research results on the Kirchhoff index and multiplication degree-Kirchhoff index can be explored through [19,20,21,22,23].

    The spanning tree of a graph G, also known as complexity, denoted as τ(G), refers to the number of subgraphs that encompass all vertices in G. This measure serves as a crucial indicator for network stability and plays a significant role in assessing the structural characteristics of graphs. For further insights into related topics such as the count of spanning trees, interested readers are encouraged to consult [24,25,26].

    With the rapid advancement of scientific research and the successful application of topology in practical scenarios, topological theory has gained increasing recognition worldwide. The calculation problem concerning the phenylene Wiener index has been effectively resolved by Pavlović and Gutman [27]. Chen and Zhang [28] have developed a precise equation for predicting the Wiener index of random phenylene chains. Additionally, Liu et al. [29] have identified both the degree-Kirchhoff index and the number of spanning trees for Ln dicyclobutadieno derivatives of [n] phenylenes.

    Given two automorphic graphs S and K, we define the symbol SK to represent the strong product of these two graphs with V(S)×V(K), which is commonly referred to as the strong product in graph theory literature. Readers can refer to [30] for more comprehensive definitions and concepts. Recently, Pan et al.[25] utilized the resistance distance of a strong prism formed by Pn and Cn to determine the Kirchhoff index. Similarly, Li et al.[31] derived graph invariants and spanning trees from the strong prism of a star Sn. Motivated by [30,31,32,33], we obtain the pentagonal network Rn and its strong product P2n. The pentagonal network consists of numerous adjacent pentagons and quadrilaterals with each quadrangle having a maximum of two non-adjacent pentagons, as shown in Figure 1. The P2n is the strong product of Rn, as depicted in Figure 2. It obviously that

    |V(P2n)|=14nand|E(P2n)|=47n8.
    Figure 1.  Linear pentagonal network Rn.
    Figure 2.  The strong product P2n of the linear pentagonal.

    In this paper, we focus on the strong product of pentagonal networks, specifically examining the graph P2n with n1. The subsequent sections are organized as follows: Section 2 provides a comprehensive review of relevant research materials, presenting illustrations, concepts and lemmas. In Section 3, we derive the normalized Laplacian spectrum and present an explicit closed formula for the multiplicative degree-Kirchhoff index. Additionally, we calculate the complexity of P2n. In Section 4, we conclude the paper.

    In this section, let Rn represent the penagonal-quadrilateral networks, as illustrated in Figure 1. P2n is composed of Rn and its copy Rn, positioned one in front and one behind, as shown in Figure 2. Moreover,

    ΦA(x)=det(xInA)

    represents the characteristic polynomial of matrix A.

    The fact that

    π=(1o,2o,,no)(1,1)(2,2)((3n),(3n))

    is an automorphism deserves attention. Let

    V1={1o,2o,,no,u1,u2,,u3n,v1,,v3n},
    V2={1o,2o,,no,u1,u2,,u3n,v1,,v3n},
    |V(P2n)|=14nand|E(P2n)|=47n8.

    Subsequently, the normalized Laplacian matrix can be represented as a block matrix, that is

    L(P2n)=(LV1V1LV1V2LV2V1LV2V2),

    in which

    LV1V1=LV2V2,LV1V2=LV2V1.

    Let

    W=(12I6n12I6n12I6n12I6n),

    then,

    WL(P2n)W=(LA00LS),

    where

    LA=LV1V1+LV1V2andLS=LV1V1LV1V2.

    Observe that W and W are transpose matrices of each other.

    The characteristic polynomial of the matrix R, is denoted as

    Φ(R):=det(xIR).

    The decomposition theorem process for P2n is obtained in a similar manner to Pan and Li[34], thus we omit this proof and present it as follows:

    Lemma 2.1. [35] Assuming that the determination of LA and LS has been previously described, then

    ΦL(Ln)(x)=ΦLA(x)ΦLS(x).

    Lemma 2.2. [13] The graph G is an undirected connected graph with n vertices and m edges. Then

    Kf(G)=2mnk=21λk.

    Lemma 2.3. [2] The number of spanning trees in G, referred to as the graph's complexity, can be considered a fundamental measure in graph theory. Then

    τ(G)=12mni=1dinj=2λj.

    In this section, we explore the methodology for deriving an explicit analytical expression of the multiplicative Kirchhoff index by traversing the normalized Laplacian matrix. Meanwhile, we determine the computational complexity of P2n. Subsequently, employing the normalized Laplacian, we derive matrices of order 6n as

    LV1V1=(1135000150000135117000170000171000017000001135000170000135100001515000011350000170001351170000170001710000017000011350000150001351)

    and

    LV1V2=(15135000150000135171700017000017170000170000017135000170000135150000151500001513500001700013517170000170001717000001700001713500001500013515).

    Due to

    LA=LV1V1(P2n)+LV1V2(P2n)

    and

    LS=LV1V1(P2n)LV1V2(P2n),

    it can be convincingly argued that

    LA=2(2513500015000013537170001700001737000000000037135000170000135250000151500002513500001700013537170000170001737000001700003713500001500013525)

    and

    LS=diag(65,87,87,,87,65,65,87,87,,87,65).

    Utilizing Lemma 2.1, it is revealed that the P2n normalized Laplacian spectrum consists of the eigenvalues from LA and LS. It is established that the LS possesses eigenvalues 65 and 87 with multiplicities of 4 and (6n4), respectively.

    Let

    M=(2513500000135371700000173717000001737000000037170000017371350000013525)(3n)×(3n)

    and

    N=diag(15,17,17,17,17,,17,17,15),

    where the matrices M and N are both of order 3n.

    The matrices M and N are combined to form a block matrix, denoted as 12LA, in the following manner:

    12LA=(MNNM).

    Suppose that

    W=(12I3n12I3n12I3n12I3n)

    is a block matrix. Hence, we can obtain

    W(12LA)W=(M+N00MN).

    Let J=M+N and K=MN. Then,

    J=(1513500000135271700000172717000001727000000027170000017471350000013535)(3n)×(3n)

    and

    K=(3513500000135471700000174717000001747000000047170000017471350000013535)(3n)×(3n),

    in which the diagonal elements are

    (35,47,47,47,47,47,35).

    Based on Lemma 2.1, it is evident and demonstrable that the eigenvalues of 12LA are identical to those of J and K. Assume that the eigenvalues of J and K are σi and ςj(i,j=1,2,,3n), respectively, with

    σ1σ2σ3σ3n,ς1ς2ς3ς3n.

    We verify σ10 and ς10. In addition, it is easy to know that the normalized Laplacian spectrum of P2n is {2σ1,2σ2,,2σ3n,2ς1,2ς2,,2ς3n}. Note that

    |E(P2n)|=47n8,

    we can obtain Lemma 3.1 according to Lemma 2.2.

    Lemma 3.1. Assume that P2n is the strong product of the pentagonal network. Then,

    Kf(P2n)=2(47n8)(2×56+(6n2)78+123ni=21σi+123nj=11ςj)=(47n8)(63n16+3ni=21σi+3nj=11ςj).

    Subsequently, we partition the computation of the aforementioned equation into two distinct components and prioritize the initial calculation of 3ni=21σi.

    Lemma 3.2. Suppose that σi(i=1,2,,3n) is defined as described previously.

    3ni=21σi=1035n3+142n2+617n2(81n+490).

    Proof. Suppose that

    Φ(J)=x3n+a1x3n1++a3nx2+a3n+1x=x(x3n1+a1x3n2++a3n1x+a3n2).

    Then σ2,σ3,,σ3n fulfil the following equation

    x3n1+a1x3n2++a3n2x+a3n1=0,

    and we observe that 1σ2,1σ3,,1σ3n are the roots of the following equation

    a3n1x3n1+a3n2x3n2++a1x+1=0.

    By Vieta's Theorem, one has

    3ni=21σi=(1)3n2a3n2(1)3n1a3n1. (3.1)

    For each value of i from 1 to 3n+1, we consider Ji and assign ji as the determinant of Ji. We will derive the formula for ji, which can be utilized to calculate (1)3n2a3n2 and (1)3n1a3n1. Then one has

    j1=15,j2=135,j3=1245,j4=11715,j5=112005,j6=184035,j7=1588245,j8=14117715,

    and

    {j3i=27j3i1149j3i2,1in;j3i+1=27j3i149j3i1,0in1;j3i+2=27j3i+1149j3i,0in1.

    Through a straightforward computation, one can derive the following general formulas:

    {j3i=75(1343)i,1in;j3i+1=15(1343)i,0in1;j3i+2=135(1343)i,0in1. (3.2)

    According to Eq (3.1), we divide the numerator and denominator into two facts and reveal them later. For the sake of convenience, we represent the diagonal elements of J as lii in a simplified manner.

    Fact 3.3.

    (1)3n1a3n1=490+81n25(1343)n.

    Proof. Given that J is a left-right symmetric matrix, the sum of all principal minors can be represented by the number (1)3n1a3n1, where these minors correspond to rows and columns with indices equal to (3n1).

    (1)3n1a3n1=3ni=1detLA[i]=3ni=1det(Ji100J3ni)=3ni=1ji1j3ni, (3.3)

    where

    J3ni=(li+1,i+1000l3n1,3n11350135l3n,3n).

    By Eqs (3.2) and (3.3), we have

    (1)3n1a3n1=2j3n+1+nl=1j3(l1)+2j3(nl)+2+nl=1j3lj3(nl)+1+n1l=0j3l+1j3(nl)=985(1343)n+343n352(1343)n+7n25(1343)n+n(1343)n=490+81n25(1343)n.

    This is the completion of the proof.

    Fact 3.4.

    (1)3n2a3n2=1035n3+142n2617n50(1343)n.

    Proof. We note that the sum of all principal minors of order 3n2 in J can be expressed as (1)3n2a3n2. After that,

    (1)3n2a3n2=3n1i<j|Ji1000Z000J3nj|,1i<j3n2,

    where

    Z=(li+1,i+100lj1,j1)

    and

    J3nj=(lj+1,j+1000l3n1,3n11350135l3n,3n).

    Note that

    (1)3n2a3n2=3n1i<jdetJi1detZdetJ3nj=3n1i<jdetZsi1s3nj. (3.4)

    According to Eq (3.4), the determinant of Z varies depending on the values of i and j, as well as s and t. Consequently, we can categorize the primary scenarios into six distinct classifications.

    Case 1. i=3s, j=3t for 1s<tn, and

    detZ1=|27170000172717000017271700001727000000271700001727|(3s3t1)=21(st)(1343)st.

    Case 2. i=3s, j=3t+1 for 1stn, and

    detZ2=|27170000172717000017271700001727000000271700001727|(3s3t)=[3(st)+1](1343)st,

    or i=3s+2, j=3t for 0s<tn, and

    detZ3=|27170000172717000017271700001727000000271700001727|(3s3t3)=(3s3t2)(1245)st1.

    Case 3. In the same way, i=3s, j=3t+2 for 1stn, and

    detZ4=|2717000017271700001727170000172700000027135000013515|(3s3t+1)=49(3s3t+2)(1343)st+1,

    or i=3s+1, j=3t for 0s<tn, and

    detZ5=|27170000172717000017271700001727000000271700001727|(3s3t2)=49(3s3t1)(1343)st1.

    Case 4. Similarly, i=3s+1, j=3t+1 for 0s<tn, and

    detZ6=|27170000172717000017271700001727000000271700001727|(3s3t1)=21(st)(1343)st,

    or i=3s+2, j=3t+2 for 0k<ln, and

    detZ7=|2717000017271700001727170000172700000027135000013515|(3s3t1)=detZ6.

    Case 5. i=3s+1, j=3t+2 for 0stn, and

    detZ8=|2717000017271700001727170000172700000025135000013527|(3s3t)=(3s3t+1)(1343)st.

    Case 6. i=3s+2, j=3t+1 for 0k<ln, and

    detZ9=|27170000172717000017271700001727000000271700001727|(3s3t2)=49(3s3t1)(1343)st.

    Therefore, we can obtain

    (1)3n2a3n2=1i<j3ndetZjk1j3nl=1+2+3, (3.5)

    where

    1=1s<tndetZ1j3s1j3n3t+1stndetZ2j3s1j3n3t1+1stn1detZ4j3s1j3n3t+1sndetJ[3s,3n+2]j3s1=7n(n21)50(1343)n+(n2+2n)(n+1)2450(1343)n+n2(n1)2450(1343)n+n(3n+1)490(1343)n=345n3+17n2336n50(1343)n,
    2=1s<tndetZ5j3sj3n3t+2+1s<tndetZ6j3sj3n3t+1+1stn1detZ8j3sj3n3t+1sndetJ[3s+1,3n+2]j3s+1sndetJ[1,3t]j3n3t+2+1tndetS[1,3t+1]s3n3t+1+0tn1detJ[1,3t+2]j3n3t+detJ[1,3n+2]=4950(n3n22n+2)(1343)n1+49n(n21)50(1343)n+49n(n21)50(1343)n+7n(3n1)10(1343)n+7n(3n+1)10(1343)n1+21n(n+1)10(1343)n+7n(3n1)10(1343)n+n(3n+1)(1343)n=345n3+73n292n50(1343)n

    and

    3=0s<tndetZ3j3s+1j3n3t+2+0s<tn1detZ7j3s+1j3n3t+0s<tndetZ9j3s+1j3n3t+1+0sn1detJ[3s+2,3n+2]j3s+1=49n(n2+n4)50(1343)n1+n(n21)50(1343)n+49n(n2+2n1)50(1343)n+21n(n+1)10(1343)n1=345n3+52n2189n50(1343)n.

    By substituting 1, 2, and 3 into Eq (3.5), the desired outcome can be deduced.

    (1)3n2a3n2=1+2+3=1035n3+142n2617n50(1343)n.

    This is the completion of the proof. Let

    0=ς1<ς2ς3ς3n

    represent the eigenvalues of J. By Facts 3.3 and 3.4, we can further investigate Lemma 3.2. According to Eq (3.1), it is evident that

    3ni=21σi=(1)3n2a3n2(1)3n1a3n1=1035n3+142n2+617n2(490+81n).

    Considering Lemma 3.1, we will focus on the calculations of 3nj=11ςj. Let

    δ(n)=10290n(11+230)+3600+124303060000

    and

    ξ(n)=10290n(11230)+3600124303060000.

    Lemma 3.5. The variable ςj (where j ranges from 1 to 3n+2) is assumed to be defined as previously described. One has

    3nj=11ςj=(1)3n1b3n1detK,

    where

    detK=45+1130125(11+430343)n+451130125(11430343)n

    and

    (1)3n1b3n1=δ(n)(11+430343)nξ(n)(11430343)n.

    Proof. The representation of Φ(K) can be expressed as

    y3n+b1y3n1++b3n2y2+b3n1y=y(y3n+1+b1y3n++b3n2y+b3n1),

    where ς1,ς2,,ς3n represent the roots of the equation.

    y3n1+b1y3n2++b3n2y+b3n1=0,

    and the equation is determined to possess 1ς1,1ς2,,1ς3n as its solutions

    b3n1y3n1+b3n2y3n2++b1y+1=0.

    By Vieta's Theorem, one holds that

    3nj=11ςj=(1)3n1b3n1detK. (3.6)

    To simplify the analysis, let Rp denote the p-th order principal minors of matrix K and kp represent the determinant of Rp. We will derive an equation for kp that can be utilized to calculate (1)3n1b3n1 and detK for values of p ranging from 1 to 3n1. Subsequently, we arrive at

    k1=15,k2=135,k3=1245,k4=11715,k5=112005,k6=184035,k7=1588245,k8=14117715

    and

    {k3p=27k3p1149k3p2,1pn;k3p+1=27k3p149k3p1,1pn;k3p+2=27k3p+1149k3p,0pn1.

    After a straightforward computation, the following general formulas can be derived

    {k3p=105+1430150(11+430343)p+1051430150(11430343)p,1pn;k3p+1=45+830150(11+430343)p+45830150(11430343)p,1pn;k3p+2=11+23070(11+430343)p+1123070(11430343)p,0pn1.

    Subsequently, we proceed by examining the following Facts:

    Fact 3.6.

    detK=45+1130125(11+430343)n+451130125(11430343)n.

    Proof. By expanding detK with respect to its last row, we obtain

    detK=35k3n+1135k3n=45+1130125(11+430343)n+451130125(11430343)n.

    The desired outcome has been achieved.

    Fact 3.7.

    (1)3n1b3n1=δ(n)(11+430343)nξ(n)(11430343)n,

    where

    δ(n)=10290n(11+430)+3600+124303060000

    and

    ξ(n)=10290n(11430)+3600124303060000.

    Proof. Considering that K has a (3n1)-row and (3n1)-column structure, the sum of all principal minors can be expressed as (1)3n1b3n1. Here, lii represents the diagonal entries of K. It is noteworthy that K exhibits bilateral symmetry, which enables us to derive specific information.

    (1)3n1b3n1=3ni=1detK[i]=3ni=1det(Ri100R3ni)=3ni=1ri1r3ni, (3.7)

    where

    R3ni=(gi+1,i+1000g3n1,3n11350135g3n,3n).

    In line with Eq (3.7), we have

    (1)3n1b3n1=2k3n1+nl=1detK[3l]+n1l=0detK[3l+1]+n1l=0detK[3l+2]=2k3n1+nl=1k3(l1)+2k3(nl)+2+n1l=0k3lk3(nl)+1+n1l=0k3l+1k3(nl). (3.8)

    Immediately, we can get

    nl=1k3(l1)+2k3(nl)+2=245n20(11+430343)n+1(11430343)n+1+30600(11+430343)n30600(11430343)n, (3.9)
    nl=0k3lk3(nl)+1=2391n300(11+430343)n+1(11430343)n+1+1373060000(11+430343)n1373060000(11430343)n, (3.10)
    n1l=0k3l+1k3(nl)=2391n300(11+430343)n+1(11430343)n+1+12713060000(11+430343)n12713060000(11430343)n (3.11)

    and

    2k3n1=90+163075(11+430343)n+90163075(11430343)n. (3.12)

    By incorporating Eqs (3.9)–(3.12) into Eq (3.8), we can achieve Lemma 3.7. Utilizing Eq (3.6), in conjunction with Facts 3.6 and 3.7, Lemma 3.5 can be promptly derived.

    By integrating Lemmas 3.1, 3.2 and 3.5, we can readily deduce the Theorems 3.8 and 3.9.

    Theorem 3.8. Assume that P2n is the strong product of the pentagonal network. One has

    Kf(P2n)=1029n3+5736n2+4275n+8503+(57n+30)(1)3n1b3n1detK,

    where

    (1)3n1b3n1=δ(n)(11+430343)nξ(n)(11430343)n,detK=45+1130125(11+430343)n+451130125(11430343)n,

    additionally,

    δ(n)=10290n(11+230)+3600+124303060000

    and

    ξ(n)=10290n(11230)+3600124303060000.

    Theorem 3.9. Let P2n be the strong product of pentagonal network. Then

    τ(P2n)=35232n+75((45+1130)(11+430)n+(451130)(11430)n).

    Proof. Based on the proof of Lemma 2.2, it is evident that σ1,σ2,σ2n+1 constitute the roots of the equation

    x2n+a1x2n1++a2n1x+a2n=0.

    Accordingly, one has

    3ni=2σi=(1)3n1a3n1.

    By Fact 3.3, we have

    3ni=2σi=12+23n25(1343)n.

    By the same method,

    3nj=1ςj=detK=45+1130375(11+430343)n+451130375((11430)343)n.

    Note that

    vVP2nd(P2n)=54716n4

    and

    |E(P2n)|=48n7.

    In conjunction with Lemma 2.3, we get

    τ(P2n)=12|E(P2n)|((65)2(87)6n23ni=22σi3nj=12ςjvVP2nd(P2n))=35232n+75((45+1130)(11+430)n+(451130)(11430)n).

    This is the completion of the proof.

    In this study, we have derived explicit expressions for the multiplicative degree-Kirchhoff index and complexity of P2n based on the spectrum of the Laplacian matrix, where P2n=RnRn. These two fundamental calculations serve as simple yet reliable graph invariants that effectively capture the stability of diverse networks. Future research should focus on applying our methodology to determine spectra for strong products of automorphic and symmetric networks.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors would like to express their sincere gratitude to the editor and anonymous referees for providing valuable suggestions, which significantly contributed to the substantial improvement of the original manuscript. This work was supported by Anhui Jianzhu University Funding under Grant KJ212005, and by Natural Science Fund of Education Department of Anhui Province under Grant KJ2020A0478.

    No potential confilcts of interest were reported by the authors.



    [1] J. C. S. Jacques Junior, S. R. Musse, C. R. Jung, Crowd analysis using computer vision techniques, IEEE Signal Process. Mag., 27 (2010), 66–77. https://doi.org/10.1109/MSP.2010.937394 doi: 10.1109/MSP.2010.937394
    [2] Z. Jie, D. Gao, D. Zhang, Moving vehicle detection for automatic traffic monitoring, vehicular technology, IEEE Trans. Veh. Technol., 56 (2007), 51–59. https://doi.org/10.1109/TVT.2006.883735 doi: 10.1109/TVT.2006.883735
    [3] H. Qiu, X. Liu, S. Rallapalli, A. J. Bency, K. Chan, R. Urgaonkar, et al., Kestrel: Video analytics for augmented multi-camera vehicle tracking, in 2018 IEEE/ACM Third International Conference on Internet-of-Things Design and Implementation (IoTDI), IEEE, Orlando, FL, (2018), 48–59. https://doi.org/10.1109/IoTDI.2018.00015
    [4] R. Mehran, A. Oyama, M. Shah, Abnormal crowd behavior detection using social force model, in 2009 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, (2009), 935–942. https://doi.org/10.1109/CVPR.2009.5206641
    [5] D. Y. Chen, P. C. Huang, Motion-based unusual event detection in human crowds, J. Visual Commun. Image Represent., 22 (2011), 178–186. https://doi.org/10.1016/j.jvcir.2010.12.004 doi: 10.1016/j.jvcir.2010.12.004
    [6] C. C. Loy, T. Xiang, S. Gong, Detecting and discriminating behavioural anomalies, Pattern Recognit., 44 (2011), 117–132. https://doi.org/10.1016/j.patcog.2010.07.023 doi: 10.1016/j.patcog.2010.07.023
    [7] P. C. Ribeiro, R. Audigier, Q. C. Pham, RIMOC, a feature to discriminate unstructured motions: Application to violence detection for video-surveillance, Comput. Vision Image Understanding, (2016), 1–23. https://doi.org/10.1016/j.cviu.2015.11.001
    [8] Y. Benabbas, N. Ihaddadene, C. Djeraba, Motion pattern extraction and event detection for automatic visual surveillance, EURASIP J. Image Video Process., 2011 (2011), 1–7. https://doi.org/10.1155/2011/163682 doi: 10.1155/2011/163682
    [9] B. Krausz, C. Bauckhage, Loveparade 2010: Automatic video analysis of a crowd disaster, Comput. Vision Image Understanding, 116 (2012), 307–319. https://doi.org/10.1016/j.cviu.2011.08.006 doi: 10.1016/j.cviu.2011.08.006
    [10] V. Kaltsa, A. Briassouli, I. Kompatsiaris, M. G. Strintzis, Timely, robust crowd event characterization, in 19th IEEE International Conference on Image Processing, IEEE, (2012), 2697–2700. https://doi.org/10.1109/ICIP.2012.6467455
    [11] Y. Zhang, L. Qiny, H. Yao, P. Xu, Q. Huang, Beyond particle flow: Bag of trajectory graphs for dense crowd event recognition, in IEEE International Conference on Image Processing, ICIP, 2013. https://doi.org/10.1109/ICIP.2013.6738737
    [12] W. Choi, S. Savarese, A unified framework for multi-target tracking and collective activity recognition, in Computer Vision-ECCV 2012, (2012), 215–230. https://doi.org/10.1007/978-3-642-33765-9_16
    [13] M. Hu, S. Ali, M. Shah, Learning motion patterns in crowded scenes using motion flow field, in 2008 19th International Conference on Pattern Recognition, (2008), 1–5. https://doi.org/10.1109/ICPR.2008.4761183
    [14] S. Ali, M. Shah, A lagrangian particle dynamics approach for crowd flow segmentation and stability analysis, in 2007 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2007. https://doi.org/10.1109/CVPR.2007.382977
    [15] C. C. Loy, T. Xiang, S. Gong, Modelling multi-object activity by gaussian processes, in Proceedings of the British Machine Vision Conference, BMVC, 2009.
    [16] A. S. Rao, J. Gubbi, S. Marusic, M. Palaniswami, Crowd event detection on optical flow manifolds, IEEE Trans. Cybern., 46 (2016), 1524–1537. https://doi.org/10.1109/TCYB.2015.2451136 doi: 10.1109/TCYB.2015.2451136
    [17] H. Mousavi, S. Mohammadi, A. Perina, R. Chellali, V. Murino, Analyzing tracklets for the detection of abnormal crowd behavior, in IEEE Winter Conference on Applications of Computer Vision, (2015), 148–155. https://doi.org/10.1109/WACV.2015.27
    [18] H. Fradi, J. L. Dugelay, Spatial and temporal variations of feature tracks for crowd behavior analysis, J. Multimodal User Interface, 10 (2016), 307–317. https://doi.org/10.1007/s12193-015-0179-2 doi: 10.1007/s12193-015-0179-2
    [19] Y. Zhao, Z. Li, X. Chen, Y. Chen, Mobile crowd sensing service platform for social security incidents in edge computing, in 2018 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), (2018), 568–574. https://doi.org/10.1109/Cybermatics_2018.2018.00118
    [20] Y. Yuan, J. Fang, Q. Wang, Online anomaly detection in crowd scenes via structure analysis, IEEE Trans. Cybern., 45 (2015), 548–561. https://doi.org/10.1109/TCYB.2014.2330853 doi: 10.1109/TCYB.2014.2330853
    [21] S. Wu, B. E. Moore, M. Shah, Chaotic invariants of Lagrangian particle trajectories for anomaly detection in crowded scenes, in 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (2010), 2054–2060. https://doi.org/10.1109/CVPR.2010.5539882
    [22] X. Cui, Q. Liu, M. Gao, D. N. Metaxas, Abnormal detection using interaction energy potentials, in Conference on Computer Vision and Pattern Recognition, CVPR, 2011, (2011), 3161–3167. https://doi.org/10.1109/CVPR.2011.5995558
    [23] S. Cho, H. Kang, Abnormal behavior detection using hybrid agents in crowded scenes, Pattern Recognit. Lett., 44 (2014), 64–70. https://doi.org/10.1016/j.patrec.2013.11.017 doi: 10.1016/j.patrec.2013.11.017
    [24] E. de-la-Calle-Silos, I. González-Díaz, E. Díaz-de-María, Mid-level feature set for specific event and anomaly detection in crowded scenes, in 2013 IEEE International Conference on Image Processing, Melbourne, (2013), 4001–4005. https://doi.org/10.1109/ICIP.2013.6738824
    [25] N. Ihaddadene, C. Djeraba, Real-time crowd motion analysis, in 2008 19th International Conference on Pattern Recognition, (2008), 1–4. https://doi.org/10.1109/ICPR.2008.4761041
    [26] A. N. Shuaibu, A. S. Malik, I. Faye, Behavior representation in visual crowd scenes using space-time features, in 2016 6th International Conference on Intelligent and Advanced Systems (ICIAS), (2016), 1–6. https://doi.org/10.1109/ICIAS.2016.7824073
    [27] R. Hu, Q. Mo, Y. Xie, Y. Xu, J. Chen, Y. Yang, et al., AVMSN: An audio-visual two stream crowd counting framework under low-quality conditions, IEEE Access, 9 (2021), 80500–80510. https://doi.org/10.1109/ACCESS.2021.3074797
    [28] U. Sajid, X. Chen, H. Sajid, T. Kim, G. Wang, Audio-visual transformer based crowd counting, in 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), (2021), 2249–2259. https://doi.org/10.1109/ICCVW54120.2021.00254
    [29] A. N. Shuaibu, A. S. Malik, I. Faye, Adaptive feature learning CNN for behavior recognition in crowd scene, in 2017 IEEE International Conference on Signal and Image Processing Applications (ICSIPA), Malaysia, (2017), 357–361. https://doi.org/10.1109/ICSIPA.2017.8120636
    [30] K. Zheng, W. Q. Yan, P. Nand, Video dynamics detection using deep neural networks, IEEE Trans. Emerging Top. Comput. Intell., 2 (2018), 224–234. https://doi.org/10.1109/TETCI.2017.2778716 doi: 10.1109/TETCI.2017.2778716
    [31] Y. Li, A deep spatio-temporal perspective for understanding crowd behavior, IEEE Trans. Multimedia, 20 (2018), 3289–3297. https://doi.org/10.1109/TMM.2018.2834873 doi: 10.1109/TMM.2018.2834873
    [32] L. F. Borja-Borja, M. Saval-Calvo, J. Azorin-Lopez, A short review of deep learning methods for understanding group and crowd activities, in 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, (2018), 1–8. https://doi.org/10.1109/IJCNN.2018.8489692
    [33] B. Mandal, J. Fajtl, V. Argyriou, D. Monekosso, P. Remagnino, Deep residual network with subclass discriminant analysis for crowd behavior recognition, in 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, (2018), 938–942. https://doi.org/10.1109/ICIP.2018.8451190
    [34] A. N. Shuaibu, A. S. Malik, I. Faye, Y. S. Ali, Pedestrian group attributes detection in crowded scenes, in 2017 International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Fez, (2017), 1–5. https://doi.org/10.1109/ATSIP.2017.8075584
    [35] F. Solera, S. Calderara, R. Cucchiara, Socially constrained structural learning for groups detection in crowd, IEEE Trans. Pattern Anal. Mach. Intell., 38 (2016), 995–1008. https://doi.org/10.1109/TPAMI.2015.2470658 doi: 10.1109/TPAMI.2015.2470658
    [36] S. Kim, S. Kwak, B. C. Ko, Fast pedestrian detection in surveillance video based on soft target training of shallow random forest, IEEE Access, 7 (2019), 12415–12426. https://doi.org/10.1109/ACCESS.2019.2892425 doi: 10.1109/ACCESS.2019.2892425
    [37] S. Din, A. Paul, A. Ahmad, B. B. Gupta, S. Rho, Service orchestration of optimizing continuous features in industrial surveillance using Big Data based fog-enabled Internet of Things, IEEE Access, 6 (2018), 21582–21591. https://doi.org/10.1109/ACCESS.2018.2800758 doi: 10.1109/ACCESS.2018.2800758
    [38] M. U. K. Khan, H. Park, C. Kyung, Rejecting motion outliers for efficient crowd anomaly detection, IEEE Trans. Inf. Forensics Secur., 14 (2018), 541–556. https://doi.org/10.1109/TIFS.2018.2856189 doi: 10.1109/TIFS.2018.2856189
    [39] A. N. Shuaibu, I. Faye, Y. Salih Ali, N. Kamel, M. N. Saad, A. S. Malik, Sparse representation for crowd attributes recognition, IEEE Access, 5 (2017), 10422–10433. https://doi.org/10.1109/ACCESS.2017.2708838 doi: 10.1109/ACCESS.2017.2708838
    [40] C. Riachy, F. Khelifi, A. Bouridane, Video-based person re-identification using unsupervised tracklet matching, IEEE Access, 7 (2019), 20596–20606. https://doi.org/10.1109/ACCESS.2019.2896779 doi: 10.1109/ACCESS.2019.2896779
    [41] H. Fradi, J. Dugelay, Sparse feature tracking for crowd change detection and event recognition, in 2014 22nd International Conference on Pattern Recognition, ICPR, (2014), 4116–4121. https://doi.org/10.1109/ICPR.2014.705
    [42] B. Yogameena, S. Saravana Perumal, N. Packiyaraj, P. Saravanan, Ma-Th algorithm for people count in a dense crowd and their behaviour classification, in 2012 International Conference on Machine Vision and Image Processing (MVIP), Taipei, (2012), 17–20. https://doi.org/10.1109/MVIP.2012.6428750
    [43] R. Mehran, A. Oyama, M. Shah, Abnormal crowd behavior detection using social force model, in 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, (2009), 935–942. https://doi.org/10.1109/ICIP.2012.6467453
    [44] X. Zhou, L. Zhang, Abnormal event detection using recurrent neural network, in 2015 International Conference on Computer Science and Applications (CSA), Wuhan, (2015), 222–226. https://doi.org/10.1109/CSA.2015.64
    [45] S. S. Pathan, A. Al-Hamadi, B. Michaelis, Crowd behavior detection by statistical modeling of motion patterns, in 2010 International Conference of Soft Computing and Pattern Recognition, Paris, (2010), 81–86. https://doi.org/10.1109/SOCPAR.2010.5686403
    [46] S. Wang, Z. Miao, Anomaly detection in crowd scene using historical information, in 2010 International Symposium on Intelligent Signal Processing and Communication Systems, Chengdu, (2010), 1–4. https://doi.org/10.1109/ISPACS.2010.5704770
    [47] P. Wang, W. Hao, Z. Sun, S. Wang, E. Tan, L. Li, et al., Regional detection of traffic congestion using in a large-scale surveillance system via deep residual TrafficNet, IEEE Access, 6 (2018), 68910–68919. https://doi.org/10.1109/ACCESS.2018.2879809 doi: 10.1109/ACCESS.2018.2879809
    [48] T. P. Nguyen, C. C. Pham, S. V. Ha, J. W. Jeon, Change detection by training a triplet network for motion feature extraction, IEEE Trans. Circuits Syst. Video Technol., 29 (2019), 433–446. https://doi.org/10.1109/TCSVT.2018.2795657 doi: 10.1109/TCSVT.2018.2795657
    [49] C. Riachy, F. Khelifi, A. Bouridane, Video-based person re-identification using unsupervised tracklet matching, IEEE Access, 7 (2019), 20596–20606. https://doi.org/10.1109/ACCESS.2019.2896779 doi: 10.1109/ACCESS.2019.2896779
    [50] G. Olague, D. E. Hernández, E. Clemente, M. Chan-Ley, Evolving head tracking routines with brain programming, IEEE Access, 6 (2018), 26254–26270. https://doi.org/10.1109/ACCESS.2018.2831633 doi: 10.1109/ACCESS.2018.2831633
    [51] Y. Lee, S. Chen, J. Hwang, Y. Hung, An ensemble of invariant features for person reidentification, IEEE Trans. Circuits Syst. Video Technol., 27 (2017), 470–483. https://doi.org/10.1109/TCSVT.2016.2637818 doi: 10.1109/TCSVT.2016.2637818
    [52] S. Yi, H. Li, X. Wang, Pedestrian behavior modeling from stationary crowds with applications to intelligent surveillance, IEEE Trans. Image Process., 25 (2016), 4354–4368. https://doi.org/10.1109/TIP.2016.2590322 doi: 10.1109/TIP.2016.2590322
    [53] Bera, S. Kim, D. Manocha, Interactive crowd-behavior learning for surveillance and training, IEEE Comput. Graphics Appl., 36 (2016), 37–45. https://doi.org/10.1109/MCG.2016.113 doi: 10.1109/MCG.2016.113
    [54] H. Fradi, B. Luvison, Q. C. Pham, Crowd behavior analysis using local mid-level visual descriptors, IEEE Trans. Circuits Syst. Video Technol., 27 (2017), 589–602. https://doi.org/10.1109/TCSVT.2016.2615443 doi: 10.1109/TCSVT.2016.2615443
    [55] M. Abdar, N. Y. Yen, Design of a universal user model for dynamic crowd preference sensing and decision-making behavior analysis, IEEE Access, 5 (2017), 24842–24852. https://doi.org/10.1109/ACCESS.2017.2735242 doi: 10.1109/ACCESS.2017.2735242
    [56] X. Zhang, Q. Yu, H. Yu, Physics inspired methods for crowd video surveillance and analysis: A survey, IEEE Access, 6 (2018), 66816–66830. https://doi.org/10.1109/ACCESS.2018.2878733 doi: 10.1109/ACCESS.2018.2878733
    [57] Y. Zhang, L. Qin, R. Ji, S. Zhao, Q. Huang, J. Luo, Exploring coherent motion patterns via structured trajectory learning for crowd mood modeling, IEEE Trans. Circuits Syst. Video Technol., 27 (2017), 635–648. https://doi.org/10.1109/TCSVT.2016.2593609 doi: 10.1109/TCSVT.2016.2593609
    [58] W. Liu, R. W. H. Lau, X. Wang, D. Manocha, Exemplar-AMMs: Recognizing Crowd movements from pedestrian trajectories, IEEE Trans. Multimedia, 18 (2016), 2398–2406. https://doi.org/10.1109/TMM.2016.2598091 doi: 10.1109/TMM.2016.2598091
    [59] C. Chen, Y. Shao, X. Bi, Detection of anomalous crowd behavior based on the acceleration feature, IEEE Sensors J., 15 (2015), 7252–7261. https://doi.org/10.1109/JSEN.2015.2472960 doi: 10.1109/JSEN.2015.2472960
    [60] V. J. Kok, C. S. Chan, GrCS: Granular computing-based crowd segmentation, IEEE Trans. Cybern., 47 (2017), 1157–11680. https://doi.org/10.1109/TCYB.2016.2538765 doi: 10.1109/TCYB.2016.2538765
    [61] C. Chen, Y. Shao, Crowd escape behavior detection and localization based on divergent centers, IEEE Sensors J., 15 (2015), 2431–2439. https://doi.org/10.1109/JSEN.2014.2381260 doi: 10.1109/JSEN.2014.2381260
    [62] Y. Zhang, L. Qin, R. Ji, H. Yao, Q. Huang, Social attribute-aware force model: Exploiting richness of interaction for abnormal crowd detection, IEEE Trans. Circuits Syst. Video Technol., 25 (2015), 1231–1245. https://doi.org/10.1109/TCSVT.2014.2355711 doi: 10.1109/TCSVT.2014.2355711
    [63] H. Yao, A. Cavallaro, T. Bouwmans, Z. Zhang, Guest Editorial Introduction to the Special Issue on group and crowd behavior analysis for intelligent multicamera video surveillance, IEEE Trans. Circuits Syst. Video Technol., 27 (2017), 405–408. https://doi.org/10.1109/TCSVT.2017.2669658 doi: 10.1109/TCSVT.2017.2669658
    [64] A. S. Keçeli, A. Kaya, Violent activity detection with transfer learning method, Electron. Lett., 53 (2017), 1047–1048. https://doi.org/10.1049/el.2017.0970 doi: 10.1049/el.2017.0970
    [65] V. Kaltsa, A. Briassouli, I. Kompatsiaris, L. J. Hadjileontiadis, M. G. Strintzis, Swarm intelligence for detecting interesting events in crowded environments, IEEE Trans. Image Process., 24 (2015), 2153–2166. https://doi.org/10.1109/TIP.2015.2409559 doi: 10.1109/TIP.2015.2409559
    [66] V. H. Roldão Reis, S. J. F. Guimarães, Z. K. Gonçalves do Patrocínio, Dense crowd counting with capsule networks, in 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), (2020), 267–272. https://doi.org/10.1109/IWSSIP48289.2020.9145163
    [67] P. Foggia, A. Saggese, N. Strisciuglio, M. Vento, N. Petkov, Car crashes detection by audio analysis in crowded roads, in 2015 12th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), (2015), 1–6. https://doi.org/10.1109/AVSS.2015.7301731
    [68] S. Duan, X. Wang, X. Yu, Crowded abnormal detection based on mixture of kernel dynamic texture, in 2014 International Conference on Audio, Language and Image Processing, (2014), 931–936. https://doi.org/10.1109/ICALIP.2014.7009931
    [69] Ç. Okuyucu, M. Sert, A. Yazici, Audio feature and classifier analysis for efficient recognition of environmental sounds, in 2013 IEEE International Symposium on Multimedia, (2013), 125–132. https://doi.org/10.1109/ISM.2013.29
    [70] M. Baillie, J. M. Jose, An audio-based sports video segmentation and event detection algorithm, in 2004 Conference on Computer Vision and Pattern Recognition Workshop, (2004), 110. https://doi.org/10.1109/CVPR.2004.298
    [71] M. A. Hossan, S. Memon, M. A. Gregory, A novel approach for MFCC feature extraction, in 2010 4th International Conference on Signal Processing and Communication Systems, (2011), 1–5. https://doi.org/10.1109/ICSPCS.2010.5709752
    [72] T. D. Ganchev, Speaker Recognition, PhD thesis, University of Patras, 2005.
    [73] H. Y. Swathi, G. Shivakumar, Evolutionary computing assisted neural network for crowd behaviour classification, Neuroquantology, 20 (2022), 2848–2855. https://doi.org/10.14704/nq.2022.20.9.NQ44331 doi: 10.14704/nq.2022.20.9.NQ44331
    [74] http://mha.cs.umn.edu/Movies/Crowd-Activity-All.avi
    [75] PETS09 dataset, Available from: http://www.cvg.reading.ac.uk/PETS2009/a.html.
    [76] M. Marsden, K. McGuinness, S. Little, N. E. O'Connor, Holistic features for real-time crowd behaviour anomaly detection, in 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, (2016), 918–922. https://doi.org/10.1109/ICIP.2016.7532491
    [77] S. Mohammadi, A. Perina, V. Murino, Violence detection in crowded scenes using substantial derivative, in 2015 12th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), 2015. https://doi.org/10.1109/AVSS.2015.7301787
    [78] H. Mousavi, S. Mohammadi, A. Perina, R. Chellali, V. Murino, Analyzing tracklets for the detection of abnormal crowd behavior, in 2015 IEEE Winter Conference on Applications of Computer Vision, IEEE, (2015), 148–155. https://doi.org/10.1109/WACV.2015.27
    [79] Y. Itcher, T. Hassner, O, Kliper-Gross, Violent flows: Real-time detection of violent crowd behavior, in 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 2012. https://doi.org/10.1109/CVPRW.2012.6239348
    [80] H. Mousavi, M. Nabi, H. Kiani, A. Perina, V Murino, Crowd motion monitoring using tracklet-based commotion measure, in 2015 IEEE International Conference on Image Processing (ICIP), IEEE, 2015. https://doi.org/10.1109/ICIP.2015.7351223
    [81] L. Lu, J. He, Z. Xu, Y. Xu, C. Zhang, J. Wang, et al., Crowd behavior understanding through SIOF feature analysis, in 2017 23rd International Conference on Automation and Computing (ICAC), Huddersfield, (2017), 1–6. https://doi.org/10.23919/IConAC.2017.8082086
  • This article has been cited by:

    1. Muhammad Shoaib Sardar, Shou-Jun Xu, Resistance Distance and Kirchhoff Index in Windmill Graphs, 2025, 22, 15701794, 159, 10.2174/0115701794299562240606054510
    2. Md. Abdus Sahir, Sk. Md. Abu Nayeem, Characterizing the Degree-Kirchhoff, Gutman, and Schultz Indices in Pentagonal Cylinders and Möbius Chains, 2025, 0278-081X, 10.1007/s00034-025-03130-9
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(12701) PDF downloads(84) Cited by(1)

Figures and Tables

Figures(6)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog