Divide-and-Conquer Strategy Kernel Ridge Regression Nonparametric Inference Simulations Semi-Nonparametric Inferences for Massive Data Guang Cheng 1 Department of Statistics Purdue University Statistics Seminar at NCSU October, 2015 1 Acknowledge NSF, Simons Foundation and ONR. A Joint Work with Tianqi Zhao and Han Liu
99
Embed
Semi-Nonparametric Inferences for Massive Datachengg/Massive_Heterogeneous_Data.pdf · Recent News on Big Data On August 6, 2014, Nature2 released news: \US Big-Data Health Network
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
On August 6, 2014, Nature2 released news: “US Big-DataHealth Network Launches Aspirin Study.”
In this $10-million pilot study, the use of aspirin to preventheart disease will be investigated;
Participants will take daily doses of aspirin that fall withinthe range typically prescribed for heart disease, and bemonitored to determine whether one dosage works betterthan the others;
The health-care data such as insurance claims, blood testsand medical histories will be collected from as many as 30million people in the United States through PCORnet3;
On August 6, 2014, Nature2 released news: “US Big-DataHealth Network Launches Aspirin Study.”
In this $10-million pilot study, the use of aspirin to preventheart disease will be investigated;
Participants will take daily doses of aspirin that fall withinthe range typically prescribed for heart disease, and bemonitored to determine whether one dosage works betterthan the others;
The health-care data such as insurance claims, blood testsand medical histories will be collected from as many as 30million people in the United States through PCORnet3;
On August 6, 2014, Nature2 released news: “US Big-DataHealth Network Launches Aspirin Study.”
In this $10-million pilot study, the use of aspirin to preventheart disease will be investigated;
Participants will take daily doses of aspirin that fall withinthe range typically prescribed for heart disease, and bemonitored to determine whether one dosage works betterthan the others;
The health-care data such as insurance claims, blood testsand medical histories will be collected from as many as 30million people in the United States through PCORnet3;
On August 6, 2014, Nature2 released news: “US Big-DataHealth Network Launches Aspirin Study.”
In this $10-million pilot study, the use of aspirin to preventheart disease will be investigated;
Participants will take daily doses of aspirin that fall withinthe range typically prescribed for heart disease, and bemonitored to determine whether one dosage works betterthan the others;
The health-care data such as insurance claims, blood testsand medical histories will be collected from as many as 30million people in the United States through PCORnet3;
PCORnet will connect multiple smaller networks, givingresearchers access to records at a large number ofinstitutions without creating a central data repository;
This decentralization creates one of the greatest challengeson how to merge and standardize data from differentnetworks to enable accurate comparison;
The many types of data – scans from medical imaging,vital-signs records and, eventually, genetic information canbe messy, and record-keeping systems vary amonghealth-care institutions.
PCORnet will connect multiple smaller networks, givingresearchers access to records at a large number ofinstitutions without creating a central data repository;
This decentralization creates one of the greatest challengeson how to merge and standardize data from differentnetworks to enable accurate comparison;
The many types of data – scans from medical imaging,vital-signs records and, eventually, genetic information canbe messy, and record-keeping systems vary amonghealth-care institutions.
PCORnet will connect multiple smaller networks, givingresearchers access to records at a large number ofinstitutions without creating a central data repository;
This decentralization creates one of the greatest challengeson how to merge and standardize data from differentnetworks to enable accurate comparison;
The many types of data – scans from medical imaging,vital-signs records and, eventually, genetic information canbe messy, and record-keeping systems vary amonghealth-care institutions.
As far as we are aware, the statistical studies of the D&Cmethod focus on either parametric inferences, e.g.,Bootstrap (Kleiner et al, 2014, JRSS-B) and Bayesian(Wang and Dunson, 2014, Arxiv), or nonparametricminimaxity (Zhang et al, 2014, Arxiv);
Semi/nonparametric inferences for massive data stillremain untouched (although they are crucially importantin evaluating reproducibility in modern scientific studies).
As far as we are aware, the statistical studies of the D&Cmethod focus on either parametric inferences, e.g.,Bootstrap (Kleiner et al, 2014, JRSS-B) and Bayesian(Wang and Dunson, 2014, Arxiv), or nonparametricminimaxity (Zhang et al, 2014, Arxiv);
Semi/nonparametric inferences for massive data stillremain untouched (although they are crucially importantin evaluating reproducibility in modern scientific studies).
In theory, we want to derive a theoretical upper bound fors under which the following oracle rule holds:“the nonparametric inferences constructed based on fN are(asymp.) the same as those on the oracle estimator fN .”
Meanwhile, we want to know how to choose the smoothingparameter in each sub-sample;
Allowing s→∞ significantly complicates the traditionaltheoretical analysis.
In theory, we want to derive a theoretical upper bound fors under which the following oracle rule holds:“the nonparametric inferences constructed based on fN are(asymp.) the same as those on the oracle estimator fN .”
Meanwhile, we want to know how to choose the smoothingparameter in each sub-sample;
Allowing s→∞ significantly complicates the traditionaltheoretical analysis.
In theory, we want to derive a theoretical upper bound fors under which the following oracle rule holds:“the nonparametric inferences constructed based on fN are(asymp.) the same as those on the oracle estimator fN .”
Meanwhile, we want to know how to choose the smoothingparameter in each sub-sample;
Allowing s→∞ significantly complicates the traditionaltheoretical analysis.
Theorem 1. Suppose regularity conditions on ε, K(·, ·) andφj(·) hold, e.g., tail condition on ε and supj ‖φj‖∞ ≤ Cφ. Giventhat H is not too large (in terms of its packing entropy), wehave for any fixed x0 ∈ X ,
√Nh(fN (x0)− f0(x0))
d−→ N(0, σ2x0), (1)
where h = h(λ) = r(λ)−1 and r(λ) ≡∑∞
i=11 + λ/µi−1.
An important consequence is that the rate√Nh and variance
σ2x0 are the same as those of fN (based on the entire dataset).Hence, the oracle property of the local confidence interval holdsunder the above conditions that determine s and λ.
5Simultaneous confidence band result delivers similar theoretical insights
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
The oracle property of local confidence interval holds under thefollowing conditions on λ and s:
Finite Rank (with a rank r):
λ = o(N−1/2) and log(λ−1) = o(log2N);
Exponential Decay (with a power p):
λ = o((logN)1/(2p)/√N) and log(λ−1) = o(log2(N));
Polynomial Decay (with a power m > 1/2):
λ N−d for some 2m/(4m+ 1) < d < 4m2/(8m− 1).
Choose λ as if working on the entire dataset with samplesize N . Hence, the standard generalized cross validationmethod (applied to each subsample) fails in this case.
Theorem 2. We prove that PLRTN,λ and PLRTN,λ are bothconsistent under some upper bound of s, but the latter isminimax optimal (Ingster, 1993) when choosing some s strictlysmaller than the above upper bound required for consistency.
An additional big data insight: we have to sacrifice certainamount of computational efficiency (avoid choosing thelargest possible s) for obtaining the optimality.
Theorem 2. We prove that PLRTN,λ and PLRTN,λ are bothconsistent under some upper bound of s, but the latter isminimax optimal (Ingster, 1993) when choosing some s strictlysmaller than the above upper bound required for consistency.
An additional big data insight: we have to sacrifice certainamount of computational efficiency (avoid choosing thelargest possible s) for obtaining the optimality.
Oracle rule holds when s does not grow too fast;choose the smoothing parameter as if not splitting the data;sacrifice computational efficiency for obtaining optimality.
Key technical tool: Functional Bahadur Representation inShang and C. (2013, AoS).
Oracle rule holds when s does not grow too fast;choose the smoothing parameter as if not splitting the data;sacrifice computational efficiency for obtaining optimality.
Key technical tool: Functional Bahadur Representation inShang and C. (2013, AoS).
Oracle rule holds when s does not grow too fast;choose the smoothing parameter as if not splitting the data;sacrifice computational efficiency for obtaining optimality.
Key technical tool: Functional Bahadur Representation inShang and C. (2013, AoS).
Oracle rule holds when s does not grow too fast;choose the smoothing parameter as if not splitting the data;sacrifice computational efficiency for obtaining optimality.
Key technical tool: Functional Bahadur Representation inShang and C. (2013, AoS).
Oracle rule holds when s does not grow too fast;choose the smoothing parameter as if not splitting the data;sacrifice computational efficiency for obtaining optimality.
Key technical tool: Functional Bahadur Representation inShang and C. (2013, AoS).
Figure: Mean-square errors of fN under different choices of N and s
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Part II: Heterogeneous Data
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Outline
1 A Partially Linear Modelling
2 Efficiency Boosting
3 Heterogeneity Testing
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Revisit US Health Data
Let us revisit the news on US Big-Data Health Network.
Different networks such as US hospitals conduct the sameclinical trial on the relation between a response variable Yi.e., heart disease, and a set of predictors Z,X1, X2, . . . , Xp
including the dosage of aspirin;
Medical knowledge suggests that the relation between Yand Z (e.g., blood pressure) should be homogeneous for allhuman;
However, for the other covariates X1, X2, . . . , Xp (e.g.,certain genes), we allow their (linear) relations with Y topotentially vary in different networks (located in differentareas). For example, the genetic functionality of differentraces might be heterogenous;
The linear relation is assumed here for simplicity, andparticularly suitable when the covariates are discrete suchas the dosage of aspirin, e.g., 1 or 2 tablets each day.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Revisit US Health Data
Let us revisit the news on US Big-Data Health Network.
Different networks such as US hospitals conduct the sameclinical trial on the relation between a response variable Yi.e., heart disease, and a set of predictors Z,X1, X2, . . . , Xp
including the dosage of aspirin;
Medical knowledge suggests that the relation between Yand Z (e.g., blood pressure) should be homogeneous for allhuman;
However, for the other covariates X1, X2, . . . , Xp (e.g.,certain genes), we allow their (linear) relations with Y topotentially vary in different networks (located in differentareas). For example, the genetic functionality of differentraces might be heterogenous;
The linear relation is assumed here for simplicity, andparticularly suitable when the covariates are discrete suchas the dosage of aspirin, e.g., 1 or 2 tablets each day.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Revisit US Health Data
Let us revisit the news on US Big-Data Health Network.
Different networks such as US hospitals conduct the sameclinical trial on the relation between a response variable Yi.e., heart disease, and a set of predictors Z,X1, X2, . . . , Xp
including the dosage of aspirin;
Medical knowledge suggests that the relation between Yand Z (e.g., blood pressure) should be homogeneous for allhuman;
However, for the other covariates X1, X2, . . . , Xp (e.g.,certain genes), we allow their (linear) relations with Y topotentially vary in different networks (located in differentareas). For example, the genetic functionality of differentraces might be heterogenous;
The linear relation is assumed here for simplicity, andparticularly suitable when the covariates are discrete suchas the dosage of aspirin, e.g., 1 or 2 tablets each day.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Revisit US Health Data
Let us revisit the news on US Big-Data Health Network.
Different networks such as US hospitals conduct the sameclinical trial on the relation between a response variable Yi.e., heart disease, and a set of predictors Z,X1, X2, . . . , Xp
including the dosage of aspirin;
Medical knowledge suggests that the relation between Yand Z (e.g., blood pressure) should be homogeneous for allhuman;
However, for the other covariates X1, X2, . . . , Xp (e.g.,certain genes), we allow their (linear) relations with Y topotentially vary in different networks (located in differentareas). For example, the genetic functionality of differentraces might be heterogenous;
The linear relation is assumed here for simplicity, andparticularly suitable when the covariates are discrete suchas the dosage of aspirin, e.g., 1 or 2 tablets each day.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Revisit US Health Data
Let us revisit the news on US Big-Data Health Network.
Different networks such as US hospitals conduct the sameclinical trial on the relation between a response variable Yi.e., heart disease, and a set of predictors Z,X1, X2, . . . , Xp
including the dosage of aspirin;
Medical knowledge suggests that the relation between Yand Z (e.g., blood pressure) should be homogeneous for allhuman;
However, for the other covariates X1, X2, . . . , Xp (e.g.,certain genes), we allow their (linear) relations with Y topotentially vary in different networks (located in differentareas). For example, the genetic functionality of differentraces might be heterogenous;
The linear relation is assumed here for simplicity, andparticularly suitable when the covariates are discrete suchas the dosage of aspirin, e.g., 1 or 2 tablets each day.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
A Partially Linear Modelling
Assume that there exist s heterogeneous subpopulations:P1, . . . , Ps (with equal sample size n = N/s);
In the j-th subpopulation, we assume
Y = XTβ(j)0 + f0(Z) + ε, (1)
where ε has a sub-Gaussian tail and V ar(ε) = σ2;
We call β(j) as the heterogeneity and f as the commonalityof the massive data in consideration;
(1) is a typical semi-nonparametric model (see C. andShang, 2015, AoS) since β(j) and f are both of interest.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
A Partially Linear Modelling
Assume that there exist s heterogeneous subpopulations:P1, . . . , Ps (with equal sample size n = N/s);
In the j-th subpopulation, we assume
Y = XTβ(j)0 + f0(Z) + ε, (1)
where ε has a sub-Gaussian tail and V ar(ε) = σ2;
We call β(j) as the heterogeneity and f as the commonalityof the massive data in consideration;
(1) is a typical semi-nonparametric model (see C. andShang, 2015, AoS) since β(j) and f are both of interest.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
A Partially Linear Modelling
Assume that there exist s heterogeneous subpopulations:P1, . . . , Ps (with equal sample size n = N/s);
In the j-th subpopulation, we assume
Y = XTβ(j)0 + f0(Z) + ε, (1)
where ε has a sub-Gaussian tail and V ar(ε) = σ2;
We call β(j) as the heterogeneity and f as the commonalityof the massive data in consideration;
(1) is a typical semi-nonparametric model (see C. andShang, 2015, AoS) since β(j) and f are both of interest.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
A Partially Linear Modelling
Assume that there exist s heterogeneous subpopulations:P1, . . . , Ps (with equal sample size n = N/s);
In the j-th subpopulation, we assume
Y = XTβ(j)0 + f0(Z) + ε, (1)
where ε has a sub-Gaussian tail and V ar(ε) = σ2;
We call β(j) as the heterogeneity and f as the commonalityof the massive data in consideration;
(1) is a typical semi-nonparametric model (see C. andShang, 2015, AoS) since β(j) and f are both of interest.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Estimation Procedure
Individual estimation in the j-th subpopulation:
(β(j)n , f (j)n )
= argmin(β,f)∈Rp×H
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − f(Z
(j)i ))2
+ λ‖f‖2H
;
Aggregation: fN = (1/s)∑s
j=1 f(j)n ;
A plug-in estimate for the j-th heterogeneity parameter:
β(j)n = argmin
β∈Rp
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
;
Our final estimate is (β(j)n , fN ).
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Estimation Procedure
Individual estimation in the j-th subpopulation:
(β(j)n , f (j)n )
= argmin(β,f)∈Rp×H
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − f(Z
(j)i ))2
+ λ‖f‖2H
;
Aggregation: fN = (1/s)∑s
j=1 f(j)n ;
A plug-in estimate for the j-th heterogeneity parameter:
β(j)n = argmin
β∈Rp
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
;
Our final estimate is (β(j)n , fN ).
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Estimation Procedure
Individual estimation in the j-th subpopulation:
(β(j)n , f (j)n )
= argmin(β,f)∈Rp×H
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − f(Z
(j)i ))2
+ λ‖f‖2H
;
Aggregation: fN = (1/s)∑s
j=1 f(j)n ;
A plug-in estimate for the j-th heterogeneity parameter:
β(j)n = argmin
β∈Rp
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
;
Our final estimate is (β(j)n , fN ).
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Estimation Procedure
Individual estimation in the j-th subpopulation:
(β(j)n , f (j)n )
= argmin(β,f)∈Rp×H
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − f(Z
(j)i ))2
+ λ‖f‖2H
;
Aggregation: fN = (1/s)∑s
j=1 f(j)n ;
A plug-in estimate for the j-th heterogeneity parameter:
β(j)n = argmin
β∈Rp
1
n
n∑i=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
;
Our final estimate is (β(j)n , fN ).
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Relation to Homogeneous Data
The major concern of homogeneous data is the extremelyhigh computational cost. Fortunately, this can be dealt bythe divide-and-conquer approach;
However, when analyzing heterogeneous data, our majorinterest1 is about how to efficiently extract commonfeatures across many subpopulations while exploringheterogeneity of each subpopulation as s→∞;
Therefore, comparisons between (β(j)n , fN ) and oracle
estimate (in terms of both risk and limit distribution)would be needed.
1D&C can be applied to the sub-population with large sample size.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Relation to Homogeneous Data
The major concern of homogeneous data is the extremelyhigh computational cost. Fortunately, this can be dealt bythe divide-and-conquer approach;
However, when analyzing heterogeneous data, our majorinterest1 is about how to efficiently extract commonfeatures across many subpopulations while exploringheterogeneity of each subpopulation as s→∞;
Therefore, comparisons between (β(j)n , fN ) and oracle
estimate (in terms of both risk and limit distribution)would be needed.
1D&C can be applied to the sub-population with large sample size.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Relation to Homogeneous Data
The major concern of homogeneous data is the extremelyhigh computational cost. Fortunately, this can be dealt bythe divide-and-conquer approach;
However, when analyzing heterogeneous data, our majorinterest1 is about how to efficiently extract commonfeatures across many subpopulations while exploringheterogeneity of each subpopulation as s→∞;
Therefore, comparisons between (β(j)n , fN ) and oracle
estimate (in terms of both risk and limit distribution)would be needed.
1D&C can be applied to the sub-population with large sample size.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Oracle Estimate
We define the oracle estimate for f as if the heterogeneityinformation βj were known:
for = argminf∈H
1
N
n,s∑i,j=1
(Y
(j)i − (β
(j)0 )TX
(j)i − f(Z
(j)i ))2
+ λ‖f‖2H
.
The oracle estimate for βj can be defined similarly:
β(j)or = argmin
β
1
n
n∑i=1
(Y
(j)i − (β(j))TX
(j)i − f0(Z
(j)i ))2
+ λ‖f‖2H
.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
A Preliminary Result: Joint Asymptotics
Theorem 3. Given proper s→∞2 and λ→ 0, we have3
( √n(β
(j)n − β
(j)0 )√
Nh(fN (z0)− f0(z0)
)) N
(0, σ2
(Ω−1 00 Σ22
)),
where Ω = E(X− E(X|Z))⊗2.
2The asymptotic independence between β(j)n and fN (z0) is mainly due
to the fact that n/N = s−1 → 0.3The asymptotic variance Σ22 of fN is the same as that of for.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Efficiency Boosting
Theorem 4 implies that β(j)n is semiparametric efficient:
√n(β(j)
n − β0) N(0, σ2(E(X− E(X|Z))⊗2)−1).
We next illustrate an important feature of massive data:strength-borrowing. That is, the aggregation ofcommonality in turn boosts the estimation efficiency of
β(j)n from semiparametric level to parametric level.
By imposing a lower bound on s (such that strength areborrowed from a sufficient number of sub-populations), weshow that4
√n(β(j)
n − β(j)0 ) N(0, σ2(E[XXT ])−1)
as if the commonality information were available.
4Recall that β(j)n = argminβ∈Rp
1n
∑ni=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Efficiency Boosting
Theorem 4 implies that β(j)n is semiparametric efficient:
√n(β(j)
n − β0) N(0, σ2(E(X− E(X|Z))⊗2)−1).
We next illustrate an important feature of massive data:strength-borrowing. That is, the aggregation ofcommonality in turn boosts the estimation efficiency of
β(j)n from semiparametric level to parametric level.
By imposing a lower bound on s (such that strength areborrowed from a sufficient number of sub-populations), weshow that4
√n(β(j)
n − β(j)0 ) N(0, σ2(E[XXT ])−1)
as if the commonality information were available.
4Recall that β(j)n = argminβ∈Rp
1n
∑ni=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Efficiency Boosting
Theorem 4 implies that β(j)n is semiparametric efficient:
√n(β(j)
n − β0) N(0, σ2(E(X− E(X|Z))⊗2)−1).
We next illustrate an important feature of massive data:strength-borrowing. That is, the aggregation ofcommonality in turn boosts the estimation efficiency of
β(j)n from semiparametric level to parametric level.
By imposing a lower bound on s (such that strength areborrowed from a sufficient number of sub-populations), weshow that4
√n(β(j)
n − β(j)0 ) N(0, σ2(E[XXT ])−1)
as if the commonality information were available.
4Recall that β(j)n = argminβ∈Rp
1n
∑ni=1
(Y
(j)i − βTX
(j)i − fN (Z
(j)i ))2
.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
0.0 0.2 0.4 0.6 0.8
0.6
0.7
0.8
0.9
1.0
log(s)/log(N)
CP
Coverage Probability of betacheck
N=256N=528N=1024N=2048N=4096
ŽǀĞƌĂŐĞ WƌŽďĂďŝůŝƚLJ ŽĨ /Ϯ
Figure: Coverage probability of 95% confidence interval based on β(j)n
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Large Scale Heterogeneity Testing
Consider a high dimensional simultaneous testing:
H0 : β(j) = β(j) for all j ∈ J, (2)
where J ⊂ 1, 2, . . . , s and |J | → ∞, versus
H1 : β(j) 6= β(j) for some j ∈ J ; (3)
Test statistic:
T0 = supj∈J
supk∈[p]
√n|β(j)k − βk|;
We can consistently approximate the quantile of the nulldistribution via bootstrap even when |J | diverges at anexponential rate of n by a nontrivial application of a recentGaussian approximation theory.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Large Scale Heterogeneity Testing
Consider a high dimensional simultaneous testing:
H0 : β(j) = β(j) for all j ∈ J, (2)
where J ⊂ 1, 2, . . . , s and |J | → ∞, versus
H1 : β(j) 6= β(j) for some j ∈ J ; (3)
Test statistic:
T0 = supj∈J
supk∈[p]
√n|β(j)k − βk|;
We can consistently approximate the quantile of the nulldistribution via bootstrap even when |J | diverges at anexponential rate of n by a nontrivial application of a recentGaussian approximation theory.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing
Large Scale Heterogeneity Testing
Consider a high dimensional simultaneous testing:
H0 : β(j) = β(j) for all j ∈ J, (2)
where J ⊂ 1, 2, . . . , s and |J | → ∞, versus
H1 : β(j) 6= β(j) for some j ∈ J ; (3)
Test statistic:
T0 = supj∈J
supk∈[p]
√n|β(j)k − βk|;
We can consistently approximate the quantile of the nulldistribution via bootstrap even when |J | diverges at anexponential rate of n by a nontrivial application of a recentGaussian approximation theory.
A Partially Linear Modelling Efficiency Boosting Heterogeneity Testing