Image credit: Unsplash

Statistical Inference for Private Data

This project studies how to perform reliable statistical inference when data access is constrained by privacy requirements. The central challenge is that privacy-preserving mechanisms deliberately perturb the data or the inferential procedure, while scientific users still need valid uncertainty quantification, hypothesis tests, confidence sets, and predictive guarantees.

The work connects robust statistics, asymptotic theory, and modern privacy frameworks. Early work on perturbed M-estimation investigates how robust estimators behave under privacy-motivated perturbations. More recent work studies differentially private inference for categorical data, equivalence testing under privacy constraints, conformal prediction with private quantile search, and Bayesian approximations for privatized data.

The broader aim is to build tools that make privacy constraints part of the statistical model rather than an afterthought. This includes understanding the inferential cost of privacy, designing procedures that retain useful operating characteristics after privatization, and developing methods that are suitable for sensitive data settings in public health, social science, and administrative data.

Selected related publications include:

  • Perturbed M-Estimation: A Further Investigation of Robust Statistics for Differential Privacy
  • Differentially Private Conformal Prediction via Quantile Binary Search
  • Fiducial Matching: Differentially Private Inference for Categorical Data
  • Equivalence Testing Under Privacy Constraints
  • Large-Sample Bayesian Approximations for Privatized Data
  • Data of the Defense and the Defense of Data
Roberto Molinari
Assistant Professor in Statistics

My research interests include robust statistics, signal processing, model selection and differential privacy.

Publications

This publication contributes to work in data privacy, robust statistics, with a focus reflected in its title: Equivalence Testing Under Privacy Constraints.

This publication contributes to work in data privacy, applied statistics, with a focus reflected in its title: Large-Sample Bayesian Approximations for Privatized Data.

This publication contributes to work in data privacy, applied statistics, with a focus reflected in its title: Differentially Private Conformal Prediction via Quantile Binary Search.

This publication contributes to work in data privacy, applied statistics, with a focus reflected in its title: Fiducial Matching: Differentially Private Inference for Categorical Data.

This work highlights the links between statistical robustness and differential privacy, investigating a new method to protect data privacy using robust statistical tools

An easy-to-read article highlighting the need and availability of solutions allowing the DoD to share more unclassified data with low disclosure risk