Original Features | ||||
1 | 18 | 0.9429 | 0.9693 | 0.4797 |
2 | 46 | 0.9782 | 0.9877 | 0.7718 |
3 | 108 | 0.9907 | 0.9939 | 0.9076 |
4 | 192 | 1 | 1 | 0.9490 |
To investigate potentially dependent lifetimes, it is necessary to extend the α-quantile inactivity time to bivariate and multivariate frameworks. To extend this measure to a dynamic multivariate framework, all possible trajectories at time t are considered. The behavior of the extended α-quantile of inactivity time was investigated in relation to the corresponding multivariate hazard rate function. The α-quantile of the inactivity order is defined and discussed for the multivariate case. The difference between the two bivariate α-quantile functions of inactivity, which is an important measure for studying the effect of treatment on lifespan, was also investigated. This measure was used to analyze the effect of laser treatment on the delay of blindness. Two bootstrap approaches were implemented to construct confidence bounds for the difference measure.
Citation: Mohamed Kayid. Dynamic multivariate quantile inactivity time and applications in investigation of a treatment effect[J]. AIMS Mathematics, 2024, 9(11): 30000-30014. doi: 10.3934/math.20241449
[1] | Wenxue Huang, Xiaofeng Li, Yuanyi Pan . Increase statistical reliability without losing predictive power by merging classes and adding variables. Big Data and Information Analytics, 2016, 1(4): 341-348. doi: 10.3934/bdia.2016014 |
[2] | Jian-Bing Zhang, Yi-Xin Sun, De-Chuan Zhan . Multiple-instance learning for text categorization based on semantic representation. Big Data and Information Analytics, 2017, 2(1): 69-75. doi: 10.3934/bdia.2017009 |
[3] | Dongyang Yang, Wei Xu . Statistical modeling on human microbiome sequencing data. Big Data and Information Analytics, 2019, 4(1): 1-12. doi: 10.3934/bdia.2019001 |
[4] | Wenxue Huang, Qitian Qiu . Forward Supervised Discretization for Multivariate with Categorical Responses. Big Data and Information Analytics, 2016, 1(2): 217-225. doi: 10.3934/bdia.2016005 |
[5] | Jinyuan Zhang, Aimin Zhou, Guixu Zhang, Hu Zhang . A clustering based mate selection for evolutionary optimization. Big Data and Information Analytics, 2017, 2(1): 77-85. doi: 10.3934/bdia.2017010 |
[6] | Ricky Fok, Agnieszka Lasek, Jiye Li, Aijun An . Modeling daily guest count prediction. Big Data and Information Analytics, 2016, 1(4): 299-308. doi: 10.3934/bdia.2016012 |
[7] | Minlong Lin, Ke Tang . Selective further learning of hybrid ensemble for class imbalanced increment learning. Big Data and Information Analytics, 2017, 2(1): 1-21. doi: 10.3934/bdia.2017005 |
[8] | David E. Bernholdt, Mark R. Cianciosa, David L. Green, Kody J.H. Law, Alexander Litvinenko, Jin M. Park . Comparing theory based and higher-order reduced models for fusion simulation data. Big Data and Information Analytics, 2018, 3(2): 41-53. doi: 10.3934/bdia.2018006 |
[9] | Cai-Tong Yue, Jing Liang, Bo-Fei Lang, Bo-Yang Qu . Two-hidden-layer extreme learning machine based wrist vein recognition system. Big Data and Information Analytics, 2017, 2(1): 59-68. doi: 10.3934/bdia.2017008 |
[10] | Yaru Cheng, Yuanjie Zheng . Frequency filtering prompt tuning for medical image semantic segmentation with missing modalities. Big Data and Information Analytics, 2024, 8(0): 109-128. doi: 10.3934/bdia.2024006 |
To investigate potentially dependent lifetimes, it is necessary to extend the α-quantile inactivity time to bivariate and multivariate frameworks. To extend this measure to a dynamic multivariate framework, all possible trajectories at time t are considered. The behavior of the extended α-quantile of inactivity time was investigated in relation to the corresponding multivariate hazard rate function. The α-quantile of the inactivity order is defined and discussed for the multivariate case. The difference between the two bivariate α-quantile functions of inactivity, which is an important measure for studying the effect of treatment on lifespan, was also investigated. This measure was used to analyze the effect of laser treatment on the delay of blindness. Two bootstrap approaches were implemented to construct confidence bounds for the difference measure.
In high dimensional, large sample categorical data analysis, feature selection or dimension reduction is usually involved. Existing feature selection procedures either call upon the original variables or rely on linear models (or generalized linear models). Linear models are restrained by certain assumptions of multivariate distribution of data. Some categories in the original categorical explanatory variables may not be informative enough, redundant or even irrelevant with the response variable. Besides, a regular feature selection's statistical reliability may be jeopardized if it picks up variables with a large domains. So we propose a category-based probabilistic approach for feature selection.
One can refer to [10,8] for more introductions to various data types and algorithms in feature selection. The reliability was introduced by variance in [9,3] and by proportion of categories in [5]. The reliability measure used here was proposed in [6], denoted as
As in [6], we propose a category based feature selection method in this article to improve the statistical reliability and to increase the overall point-hit accuracy by merging or removing the less-informative or redundant categories in the categorical explanatory variables. Instead of doing as in [6], we first transform each original categorical explanatory variable into multiple dummy variables, then select the more informative ones by a stepwise forward feature selection approach, and then merge the unselected categories. The merging process in [6], on the other hand, is to find less informative categories within those pre-selected original explanatory variables and merge. Our proposed approach here can compare not only the categories within one explanatory variable one another, but also among the categories in other explanatory variables. Certain introductions and applications to dummy variable can be found in [1,2].
The rest of this article is organized as follows. Section 2 introduces the association measures and the reliability measures; Section 3 introduces the dummy variable approach, proves two theorems, and proposes the detailed feature selection steps; two experiments are conducted in Section 4; we briefly summarize results of this article in the last section.
Assume we are given a data set with one categorical explanatory variable
The GK-lambda (denoted as
λ=∑xρxm−ρ⋅m1−ρ⋅m, |
where
ρ⋅m=maxyρ⋅y=maxyp(Y=y), ρxm=maxyρxy=maxyp(X=x;Y=y). |
Please note that
τ=∑x∑yρxy2/ρx⋅−∑yρ⋅y21−∑yρ⋅y2, |
where
ρx⋅=p(X=x). |
Both
Given a categorical data set with two variables
E(Gini(X|Y))=1−nX∑i=1nY∑j=1p(X=i|Y=j)2p(Y=j). |
Notice that
0≤E(Gini(X|Y))≤1−1|Domain(X)|≤1−1|Domain(X,Y)|; |
and the smaller
We transform
Proposition 3.1.
τ(Y|X1,X2,⋯,XnX)=τ(Y|X). |
Proof. Since
ωY|X=∑i,sp(Y=s|X=i)2p(X=i). |
Thus we only need to prove that
ωY|X1,X2,⋯,XnX=ωY|X |
Since
ωY|X1,X2,⋯,XnX=nY∑j=11∑i1=01∑i2=0⋯1∑inX=0p(X1=i1,X2=i1,⋯,XnX=inX,Y=j)2p(X1=i1,X2=i2,⋯,Xn=in) |
and
Xi=1 when and only when Xs=0, for all s≠i,s=1,2,⋯,nX, |
we have
p(X1=0,X2=0,⋯,Xi=1,⋯,XnX=0,Y=j)=p(X=i,Y=j), |
j=1,2,⋯,nY. |
So
ωY|X1,X2,⋯,XnX=nY∑j=1nnX∑i=1p(X1=0,X2=0,⋯,Xi=1,⋯,XnX=0,Y=j)2p(X1=0,X2=0,⋯,Xi=1,⋯,Xn=0)=nY∑j=1nX∑i=1p(X=i,Y=j)2p(X=i)=ωY|X, |
that is,
ωY|X1,⋯,XnX=ωY|X⇔τ(Y|X1,⋯,XnX)=τ(Y|X). |
It is of interest to introduce the following notion.
Definition 3.1. For a given categorical response variable
The next proposition tells that merging two categories in
Proposition 3.2. If two categories,
τ(Y|X′)≤τ(Y|X), |
where
Proof. Notice that this proposition is equivalent to the following inequality.
ωY|X′≤ωY|X. |
Let
p(X=s;Y=j)p(X=s)=bs,p(X=t;Y=j)p(X=t)=bt, for Y=1,2,⋯,nY. |
We have
p(X=s;Y=j)=p(X=s)bs,p(X=t;Y=j)=p(X=t)bt, and ωY|X′=nY∑j=1∑i≠s,tp(X=i;Y=j)2p(X=i)+nY∑j=1p(X=m;Y=j)2p(X=m), |
where
nY∑j=1p(X=m;Y=j)2p(X=m)=nY∑j=1(p(X=s;Y=j)+p(X=t;Y=j))2p(X=s)+p(X=t)=nY∑j=1(p(X=s)bs+p(X=t)bt)2p(X=s)+p(X=t), | (1) |
and
nY∑j=1(p(X=s;Y=j)2p(X=s)+p(X=t;Y=j)2p(X=t))=nY∑j=1(b2sp(X=s)+b2tp(X=t)). | (2) |
Multiplying
(1)=nY∑j=1(b2sp(X=s)2+b2tp(X=t)2+2bsbtp(X=s)p(X=t)),(2)=nY∑j=1(b2sp(X=s)2+b2tp(X=t)2+(b2s+b2t)p(X=s)p(X=t)). |
Since
ωY|X′≤ωY|X⇔τ(Y|X′)≤τ(Y|X); |
and the equality holds if and only if
In actual high dimensional data analysis projects, there are usually some categories in some explanatory variables that can be merged such that the association degree decrease is ignorable while the merge raises significantly selected features' statistical reliability. This is especially the case when the data set is high dimensional and many explanatory variables have many categories. Two experiments are conducted in the next section to support this supposition by showing that with similar reliability, merging categories can significantly increase the statistical reliability while not reducing association degree significantly.
A feature selection procedure usually follows a stepwise forward variable selection scheme, in which explanatory variables are selected one by one until certain pre-assigned threshold is hit. A reasonable threshold to stop the process is set by an acceptable association degree and statistical reliability. Specifically, for a given set of explanatory variables
1. identify a subset of explanatory variables, denoted as
D1={Xh∈X|τ(Y|{Xh}∪D0)=maxXi∈X∖D0τ(Y|{Xi}∪D0)} |
where
2. select the one in
Xi1={Xk|E(Gini({Xk}∪D0|Y))=minXh∈D1E(Gini({Xh}∪D0|Y))}; |
3. define the new set of selected variables as follows.
D2={Xi1}∪D0 |
4. repeat the previous steps until the stopping criterion is met.
Thus the idea of this general feature selection process is to select a set of variables with the highest association degree and the reliability from the previous step, or with the association degree from the previous step and the highest statistical reliability. More detailed explanations and similar procedures can be found in [8].
The category based version of the previous procedure is to transform all the original (non-binary categorical) explanatory variables into dummy variables before the general steps. The unselected categories are then merged into a new category in each original variable as described below.
1. Transform each original variable
2. Follow the steps in Section (3.2) to select
3. Merge the remaining
Notice that despite the genuine advantage of the category-based forward selection process, it has a higher time cost than the corresponding original variable based approach. It has to go through more loops to reach the same target given more features to be scanned. Generally, a complexity analysis needs to be related to a specifically designed and implemented algorithm. However, it is not this article's objective to discuss this subject in detail thus a brief discussion is carried out as follows.
Assume that the time cost for one variable set's association is a constant
The experiment's purpose is to evaluate the association and the reliability differences between the category based and the original variable based feature selection processes. The first experiment uses the mushroom data set from UCI Machine Learning Repository[13]. It has
The mushroom's type is chosen as the response variable while the other 21 variables are the explanatory ones. We are going to compare the feature selection result by the original variables and that by the transformed dummy variables. Please note that
Original Features | ||||
1 | 18 | 0.9429 | 0.9693 | 0.4797 |
2 | 46 | 0.9782 | 0.9877 | 0.7718 |
3 | 108 | 0.9907 | 0.9939 | 0.9076 |
4 | 192 | 1 | 1 | 0.9490 |
Merged Features | ||||
4 | 16 | 0.9445 | 0.9693 | 0.2098 |
4 | 24 | 0.9908 | 0.9939 | 0.2143 |
5 | 30 | 0.9962 | 0.9979 | 0.4669 |
6 | 38 | 1 | 1 | 0.6638 |
As described in Table 1, there can only be four variables by the feature selection through the original variables with the final association which reaches the maximum (data-based) association degree as
The category-based feature selection always gives rise to remarkably better reliability (
It is also see from these two tables that we have in both experiments a higher association by the categories than that by the original variables given the same reliability threshold: for an almost equal reliability, say,
In this experiment, variable HouseholdType is chosen as the response variable and all other
OrigVarFeatures | ||||
1 | 66 | 0.3005 | 0.3444 | 0.8201 |
2 | 252 | 0.3948 | 0.4391 | 0.9046 |
3 | 1830 | 0.4383 | 0.4648 | 0.9833 |
Merged Features | ||||
2 | 24 | 0.3242 | 0.3934 | 0.5491 |
2 | 36 | 0.3573 | 0.4165 | 0.6242 |
2 | 48 | 0.3751 | 0.4234 | 0.6388 |
3 | 96 | 0.3901 | 0.4234 | 0.7035 |
4 | 186 | 0.4017 | 0.4269 | 0.7774 |
4 | 282 | 0.4121 | 0.4317 | 0.8066 |
5 | 558 | 0.4221 | 0.4548 | 0.8782 |
6 | 966 | 0.4314 | 0.4768 | 0.8968 |
7 | 1716 | 0.4436 | 0.4856 | 0.9135 |
One can see from these two tables that the category based approaches produces an association degree
By transforming the categorical explanatory variables into their dummy forms and applying feature selection procedure to the transformed variables, we can select the informative categories and merge the less informative or redundant categories in the explanatory variables in order to increase the association and raise the reliability.
[1] |
D. C. Schmittlein, D. G. Morrison, The median residual lifetime: A characterization theorem and an application, Oper. Res., 29 (1981), 392–399. https://doi.org/10.1287/opre.29.2.392 doi: 10.1287/opre.29.2.392
![]() |
[2] |
R. E. Barlow, A. W. Marshall, F. Proschan, Properties of probability distributions with monotone hazard rate, Ann. Math. Statist., 34 (1963), 375–389. https://doi.org/10.1214/aoms/1177704147 doi: 10.1214/aoms/1177704147
![]() |
[3] |
H. W. Block, T. H. Savits, H. Singh, The reversed hazard rate function, Probab. Eng. Inf. Sci., 12 (1998), 69–90. https://doi.org/10.1017/S0269964800005064 doi: 10.1017/S0269964800005064
![]() |
[4] |
A. D. Crescenzo, Some results on the proportional reversed hazards model, Stat. Probab. Lett., 50 (2000), 313–321. https://doi.org/10.1016/S0167-7152(00)00127-9 doi: 10.1016/S0167-7152(00)00127-9
![]() |
[5] |
N. K. Chandra, D. Roy, Some results on reversed hazard rate, Probab. Eng. Inf. Sci., 15 (2001), 95–102. https://doi.org/10.1017/S0269964801151077 doi: 10.1017/S0269964801151077
![]() |
[6] |
M. S. Finkelstein, On the reversed hazard rate, Reliab. Eng. Syst. Saf., 78 (2002), 71–75. https://doi.org/10.1016/S0951-8320(02)00113-8 doi: 10.1016/S0951-8320(02)00113-8
![]() |
[7] |
C. Kundu, A. K. Nanda, T. Hu, A note on reversed hazard rate of order statistics and record values, J. Stat. Plan. Infer., 139 (2009), 1257–1265. https://doi.org/10.1016/j.jspi.2008.08.002 doi: 10.1016/j.jspi.2008.08.002
![]() |
[8] |
X. Li, D. Da, P. Zhao, On reversed hazard rate in general mixture models, Stat. Probab. Lett., 80 (2010), 654–661. https://doi.org/10.1016/j.spl.2009.12.023 doi: 10.1016/j.spl.2009.12.023
![]() |
[9] |
M. Burkschat, N. Torrado, On the reversed hazard rate of sequential order statistics, Stat. Probab. Lett., 85 (2014), 106–113. https://doi.org/10.1016/j.spl.2013.11.015 doi: 10.1016/j.spl.2013.11.015
![]() |
[10] |
M. Esna-Ashari, N. Balakrishnan, M. Alimohammadi, HR and RHR orderings of generalized order statistics, Metrika, 86 (2023), 131–148. https://doi.org/10.1007/s00184-022-00865-2 doi: 10.1007/s00184-022-00865-2
![]() |
[11] |
J. E. Contreras-Reyes, D. I. Gallardo, O. Kharazmi, Time-dependent residual Fisher information and distance for some special continuous distributions, Commun. Stat.-Simul. Comput., 53 (2022), 4331–4351. https://doi.org/10.1080/03610918.2022.2146136 doi: 10.1080/03610918.2022.2146136
![]() |
[12] |
N. Unnikrishnan, B. Vineshkumar, Reversed percentile residual life and related concepts, J. Korean Stat. Soc., 40 (2011), 85–92. https://doi.org/10.1016/j.jkss.2010.06.001 doi: 10.1016/j.jkss.2010.06.001
![]() |
[13] |
M. Mahdy, Further results involving percentile inactivity time order and its inference, Metron, 72 (2014), 269–282. https://doi.org/10.1007/s40300-013-0017-9 doi: 10.1007/s40300-013-0017-9
![]() |
[14] |
M. Shafaei Noughabi, Solving a functional equation and characterizing distributions by quantile past lifetime functions, Econ. Qual. Control, 31 (2016), 55–58. https://doi.org/10.1515/eqc-2015-0017 doi: 10.1515/eqc-2015-0017
![]() |
[15] |
M. Shafaei Noughabi, S. Izadkhah, . On the quantile past lifetime of the components of the parallel systems, Commun. Stat.-Theory Methods, 45 (2016), 2130–2136. https://doi.org/10.1080/03610926.2013.875573 doi: 10.1080/03610926.2013.875573
![]() |
[16] |
L. Balmert, J. H. Jeong, Nonparametric inference on quantile lost lifespan, Biometrics, 73 (2017), 252–259. https://doi.org/10.1111/biom.12555 doi: 10.1111/biom.12555
![]() |
[17] |
L. C. Balmert, R. Li, L. Peng, J. H. Jeong, Quantile regression on inactivity time, Stat. Methods Med. Res., 30 (2021), 1332–1346. https://doi.org/10.1177/0962280221995977 doi: 10.1177/0962280221995977
![]() |
[18] |
M. Kayid, Statistical inference of an α-quantile past lifetime function with applications, AIMS Mathematics, 9 (2024), 15346–15360. https://doi.org/10.3934/math.2024745 doi: 10.3934/math.2024745
![]() |
[19] | A. P. Basu, Bivariate failure rate, J. Am. Stat. Assoc., 66 (1971), 103–104. |
[20] |
N. L. Johnson, S. Kotz, A vector multivariate hazard rate, J. Multivar. Anal., 5 (1975), 53–66. https://doi.org/10.1016/0047-259X(75)90055-X doi: 10.1016/0047-259X(75)90055-X
![]() |
[21] |
K. R. Nair, N. U. Nair, Bivariate mean residual life, IEEE Trans. Reliab., 38 (1989), 362–364. https://doi.org/10.1109/24.44183 doi: 10.1109/24.44183
![]() |
[22] |
M. Shaked, J. G. Shanthikumar, Dynamic multivariate mean residual life functions, J. Appl. Probab., 28 (1991), 613–629. https://doi.org/10.2307/3214496 doi: 10.2307/3214496
![]() |
[23] |
M. Shafaei Noughabi, M. Kayid, A. M. Abouammoh, Dynamic multivariate quantile residual life in reliability theory, Math. Probl. Eng., 2018 (2018), 1245656. https://doi.org/10.1155/2018/1245656 doi: 10.1155/2018/1245656
![]() |
[24] |
M. Shafaei Noughabi, M. Kayid, Bivariate quantile residual life: a characterization theorem and statistical properties, Stat. Pap., 60 (2019), 2001–2012. https://doi.org/10.1007/s00362-017-0905-9 doi: 10.1007/s00362-017-0905-9
![]() |
[25] | M. Kayid, Multivariate mean inactivity time functions with reliability applications, Int. J. Reliab. Appl., 7 (2006), 127–140. |
[26] |
F. Buono, E. De Santis, M. Longobardi, S. Fabio, Multivariate reversed hazard rates and inactivity times of systems, Methodol. Comput. Appl. Probab., 24 (2022), 1987–2008. https://doi.org/10.1007/s11009-021-09905-2 doi: 10.1007/s11009-021-09905-2
![]() |
[27] | R. B. Nelsen, Archimedean Copulas, In: An introduction to copulas, New York: Springer, 2006,109–155. |
[28] | M. Shaked, J. G. Shanthikumar, Stochastic orders, New York: Springer, 2007. https://doi.org/10.1007/978-0-387-34675-5 |
[29] |
M. Shaked, J. G. Shanthikumar, Multivariate stochastic orderings and positive dependence in reliability theory, Math. Oper. Res., 15 (1990), 545–552. https://doi.org/10.1287/moor.15.3.545 doi: 10.1287/moor.15.3.545
![]() |
[30] | M. Shaked, J. G. Shanthikumar, Multivariate conditional hazard rate and mean residual life functions and their applications, In: Reliability and Decision Making, New York: Chapman, 1993,137–155. |
Original Features | ||||
1 | 18 | 0.9429 | 0.9693 | 0.4797 |
2 | 46 | 0.9782 | 0.9877 | 0.7718 |
3 | 108 | 0.9907 | 0.9939 | 0.9076 |
4 | 192 | 1 | 1 | 0.9490 |
Merged Features | ||||
4 | 16 | 0.9445 | 0.9693 | 0.2098 |
4 | 24 | 0.9908 | 0.9939 | 0.2143 |
5 | 30 | 0.9962 | 0.9979 | 0.4669 |
6 | 38 | 1 | 1 | 0.6638 |
OrigVarFeatures | ||||
1 | 66 | 0.3005 | 0.3444 | 0.8201 |
2 | 252 | 0.3948 | 0.4391 | 0.9046 |
3 | 1830 | 0.4383 | 0.4648 | 0.9833 |
Merged Features | ||||
2 | 24 | 0.3242 | 0.3934 | 0.5491 |
2 | 36 | 0.3573 | 0.4165 | 0.6242 |
2 | 48 | 0.3751 | 0.4234 | 0.6388 |
3 | 96 | 0.3901 | 0.4234 | 0.7035 |
4 | 186 | 0.4017 | 0.4269 | 0.7774 |
4 | 282 | 0.4121 | 0.4317 | 0.8066 |
5 | 558 | 0.4221 | 0.4548 | 0.8782 |
6 | 966 | 0.4314 | 0.4768 | 0.8968 |
7 | 1716 | 0.4436 | 0.4856 | 0.9135 |
Original Features | ||||
1 | 18 | 0.9429 | 0.9693 | 0.4797 |
2 | 46 | 0.9782 | 0.9877 | 0.7718 |
3 | 108 | 0.9907 | 0.9939 | 0.9076 |
4 | 192 | 1 | 1 | 0.9490 |
Merged Features | ||||
4 | 16 | 0.9445 | 0.9693 | 0.2098 |
4 | 24 | 0.9908 | 0.9939 | 0.2143 |
5 | 30 | 0.9962 | 0.9979 | 0.4669 |
6 | 38 | 1 | 1 | 0.6638 |
OrigVarFeatures | ||||
1 | 66 | 0.3005 | 0.3444 | 0.8201 |
2 | 252 | 0.3948 | 0.4391 | 0.9046 |
3 | 1830 | 0.4383 | 0.4648 | 0.9833 |
Merged Features | ||||
2 | 24 | 0.3242 | 0.3934 | 0.5491 |
2 | 36 | 0.3573 | 0.4165 | 0.6242 |
2 | 48 | 0.3751 | 0.4234 | 0.6388 |
3 | 96 | 0.3901 | 0.4234 | 0.7035 |
4 | 186 | 0.4017 | 0.4269 | 0.7774 |
4 | 282 | 0.4121 | 0.4317 | 0.8066 |
5 | 558 | 0.4221 | 0.4548 | 0.8782 |
6 | 966 | 0.4314 | 0.4768 | 0.8968 |
7 | 1716 | 0.4436 | 0.4856 | 0.9135 |