Experimental design issues in big data: the question of bias
From MaRDI portal
Publication:3296456
DOI10.1007/978-3-030-21140-0_20zbMATH Open1436.62385arXiv1712.06916OpenAlexW2972118069MaRDI QIDQ3296456FDOQ3296456
Authors: Eva Riccomagno, Henry P. Wynn
Publication date: 7 July 2020
Published in: Statistical Learning of Complex Data (Search for Journal in Brave)
Abstract: Data can be collected in scientific studies via a controlled experiment or passive observation. Big data is often collected in a passive way, e.g. from social media. In studies of causation great efforts are made to guard against bias and hidden confounders or feedback which can destroy the identification of causation by corrupting or omitting counterfactuals (controls). Various solutions of these problems are discussed, including randomization.
Full work available at URL: https://arxiv.org/abs/1712.06916
Recommendations
Cites Work
- The central role of the propensity score in observational studies for causal effects
- Causality. Models, reasoning, and inference
- Bayesian inference for causal effects: The role of randomization
- Learning functions and approximate Bayesian computation design: ABCD
- Maximum Entropy Sampling and Optimal Bayesian Experimental Design
- A Basis for the Selection of a Response Surface Design
- A minimax approach to randomization and estimation in survey sampling
- A minimax approach to sample surveys
- Information-Based Optimal Subdata Selection for Big Data Linear Regression
- I-robust and D-robust designs on a finite design space
- Minimum bias designs with constraints
- Principles of experimental design for big data analysis
- Generic identifiability of linear structural equation models by ancestor decomposition
Cited In (2)
This page was built for publication: Experimental design issues in big data: the question of bias
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q3296456)