Doubly robust inference when combining probability and non-probability samples with high dimensional data
We consider integrating a non-probability sample with a probability sample which provides high dimensional representative covariate information of the target population. We propose a two-step approach for variable selection and finite population inference. In the first step, we use penalized estimating equations with folded concave penalties to select important variables and show selection consistency for general samples. In the second step, we focus on a doubly robust estimator of the finite population mean and re-estimate the nuisance model parameters by minimizing the asymptotic squared bias of the doubly robust estimator. This estimating strategy mitigates the possible first-step selection error and renders the doubly robust estimator root n consistent if either the sampling probability or the outcome model is correctly specified.
Medienart: |
E-Artikel |
---|
Erscheinungsjahr: |
2020 |
---|---|
Erschienen: |
2020 |
Enthalten in: |
Zur Gesamtaufnahme - volume:82 |
---|---|
Enthalten in: |
Journal of the Royal Statistical Society. Series B, Statistical methodology - 82(2020), 2 vom: 01. Apr., Seite 445-465 |
Sprache: |
Englisch |
---|
Beteiligte Personen: |
Yang, Shu [VerfasserIn] |
---|
Links: |
---|
Themen: |
Data integration |
---|
Anmerkungen: |
Date Revised 13.11.2020 published: Print-Electronic Citation Status PubMed-not-MEDLINE |
---|
doi: |
10.1111/rssb.12354 |
---|
funding: |
|
---|---|
Förderinstitution / Projekttitel: |
|
PPN (Katalog-ID): |
NLM317305522 |
---|
LEADER | 01000naa a22002652 4500 | ||
---|---|---|---|
001 | NLM317305522 | ||
003 | DE-627 | ||
005 | 20231225163005.0 | ||
007 | cr uuu---uuuuu | ||
008 | 231225s2020 xx |||||o 00| ||eng c | ||
024 | 7 | |a 10.1111/rssb.12354 |2 doi | |
028 | 5 | 2 | |a pubmed24n1057.xml |
035 | |a (DE-627)NLM317305522 | ||
035 | |a (NLM)33162780 | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
100 | 1 | |a Yang, Shu |e verfasserin |4 aut | |
245 | 1 | 0 | |a Doubly robust inference when combining probability and non-probability samples with high dimensional data |
264 | 1 | |c 2020 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a ƒaComputermedien |b c |2 rdamedia | ||
338 | |a ƒa Online-Ressource |b cr |2 rdacarrier | ||
500 | |a Date Revised 13.11.2020 | ||
500 | |a published: Print-Electronic | ||
500 | |a Citation Status PubMed-not-MEDLINE | ||
520 | |a We consider integrating a non-probability sample with a probability sample which provides high dimensional representative covariate information of the target population. We propose a two-step approach for variable selection and finite population inference. In the first step, we use penalized estimating equations with folded concave penalties to select important variables and show selection consistency for general samples. In the second step, we focus on a doubly robust estimator of the finite population mean and re-estimate the nuisance model parameters by minimizing the asymptotic squared bias of the doubly robust estimator. This estimating strategy mitigates the possible first-step selection error and renders the doubly robust estimator root n consistent if either the sampling probability or the outcome model is correctly specified | ||
650 | 4 | |a Journal Article | |
650 | 4 | |a Data integration | |
650 | 4 | |a Double robustness | |
650 | 4 | |a Generalizability | |
650 | 4 | |a Penalized estimating equation | |
650 | 4 | |a Variable selection | |
700 | 1 | |a Kim, Jae Kwang |e verfasserin |4 aut | |
700 | 1 | |a Song, Rui |e verfasserin |4 aut | |
773 | 0 | 8 | |i Enthalten in |t Journal of the Royal Statistical Society. Series B, Statistical methodology |d 2003 |g 82(2020), 2 vom: 01. Apr., Seite 445-465 |w (DE-627)NLM19000388X |x 1467-9868 |7 nnns |
773 | 1 | 8 | |g volume:82 |g year:2020 |g number:2 |g day:01 |g month:04 |g pages:445-465 |
856 | 4 | 0 | |u http://dx.doi.org/10.1111/rssb.12354 |3 Volltext |
912 | |a GBV_USEFLAG_A | ||
912 | |a GBV_NLM | ||
951 | |a AR | ||
952 | |d 82 |j 2020 |e 2 |b 01 |c 04 |h 445-465 |