Posts by Collection

portfolio

publications

Fair Data Representation for Machine Learning at the Pareto Frontier

Published in Journal of Machine Learning Research (JMLR), 24 (2023), 1–63, 2023

We develop a principled framework for constructing fair data representations that achieve explicit, controllable trade-offs between utility and fairness—characterizing and computing solutions along a Pareto frontier.

Recommended citation: Shizhou Xu, Thomas Strohmer. (2023). “Fair Data Representation for Machine Learning at the Pareto Frontier.” Journal of Machine Learning Research, 24, 1–63.
Download Paper | Download Slides

On the (In)Compatibility between Individual and Group Fairness

Published in Under review (SIAM Journal on Mathematics of Data Science — SIMODS), 2024

We analyze fundamental tensions between individual and group fairness notions, clarifying when they can or cannot be simultaneously satisfied and what trade-offs are unavoidable.

Recommended citation: Shizhou Xu, Thomas Strohmer. (2024). “On the (In)Compatibility between Individual and Group Fairness.” Under review at SIAM Journal on Mathematics of Data Science.
Download Paper | Download Slides

WHOMP: Improving Upon Randomized Controlled Trials via Wasserstein Homogeneity

Published in Under review (Journal of the American Statistical Association), 2024

We introduce WHOMP, a Wasserstein-homogeneity optimality principle for subgroup splitting in comparative experiments (clinical trials, social experiments, and A/B tests). The method yields interpretable criteria, efficient estimators, and strong empirical gains over random partitioning, covariate-adaptive randomization, rerandomization, and anti-clustering baselines.

Recommended citation: Shizhou Xu, Thomas Strohmer. (2024). “WHOMP: Improving Upon Randomized Controlled Trials via Wasserstein Homogeneity.” Under review at Journal of the American Statistical Association.
Download Paper | Download Slides

Machine Unlearning via Information-Theoretic Regularization

Published in Manuscript, 2025

We develop information-theoretic regularization principles for machine unlearning, aiming to remove targeted information while maintaining general utility and enabling principled evaluation.

Recommended citation: Shizhou Xu, Thomas Strohmer. (2025). “Machine Unlearning via Information-Theoretic Regularization.” under review at Mathematical Foundations of Machine Learning.
Download Paper | Download Slides

Multi-resolution Enhancement for Full Spectrum Neural Representations

Published in Under review at Nature Machine Intelligence, 2025

We develop multi-resolution enhancement strategies for full-spectrum neural representations, improving fidelity across scales with an emphasis on robust learning and generalization.

Recommended citation: Yuan Ni, Z. Chen, Shizhou Xu, C. Peng, R. Plumley, C. H. Yoon, J. Thayer, J. Turner. (2025). “Multi-resolution Enhancement for Full Spectrum Neural Representations.” Under review at Nature Machine Intelligence.
Download Paper | Download Slides

talks

Fairness in Machine Learning

Published:

Invited talk introducing core fairness notions, practical pitfalls, and research directions in trustworthy ML.

teaching

MATH 127C — Real Analysis (Summer 2025)

, , 2025

This page mirrors announcements, policies, and a living schedule for the Summer 2025 offering. Lecture notes and problem sets reflect the topics we covered this term: metric spaces, compactness/connectedness, multivariable differentiability (Jacobian, chain/implicit/inverse theorems), $k$–volume and Gram determinants, change of variables, Fubini/Tonelli, and Green/Stokes/Divergence.