When Sample Selection Bias Precipitates Model Collapseedit
ICML 2026 paper on local sample-selection bias, model collapse, and collaborative Wasserstein-geometry proxies.
When Sample Selection Bias Precipitates Model Collapse is an ICML 2026 conference paper by Xinbao Qiao, Xianglong Du, Wei Liu, Jingqi Zhang, Peihua Mai, Meng Zhang, and Yan Pang. It examines how local verification in data silos can turn recursive synthetic-data training into a diversity-loss process, and how collaborative distributional proxies can reduce that failure mode.

Overviewedit
The paper studies model collapse in recursive synthetic-data training. Prior work often treats data selection as a stabilizing tool: a verifier filters generated samples so that only high-quality synthetic data are reused for training. This paper makes the verifier the object of analysis. When the verifier sees only a biased local slice of the target distribution, selection can repeatedly reward samples near that local slice and remove tail modes that future generators need.
The motivating setting is a data-silo environment. A hospital, bank, or proprietary institution may evaluate synthetic samples against its own limited reference data. Selection then becomes a confirmation-bias mechanism: samples close to the local view are retained, while distributional tails needed for generalization are pruned away.
Methodedit
The paper first formalizes biased top-alpha selection under Gaussian modeling and connects it to variance collapse across recursive generations. It then proposes collaborative evaluation methods that replace a single local verifier with distributional proxies computed across parties without raw-data exchange. The methodological shift is from sample quality as judged by one silo to distributional fit against a proxy for the global target.
Two schemes are described:
- Scheme I, collaborative geodesic interpolation, constructs proxy measures along Wasserstein geodesics between synthetic and local real distributions;
- Scheme II, collaborative Wasserstein barycenter estimation, computes a reusable barycenter proxy for the collective reference distribution.
Both schemes use Wasserstein-gradient-based sample scoring, so the synthetic pool is evaluated against a multi-party distributional reference rather than one biased silo.

Key formulaedit
The paper's theory links local selection, diversity decay, and Wasserstein cost. In the following summary, is the selected top- region, is the filtered synthetic distribution at generation , and is the target distribution.
Local verifier selection is summarized as truncated sampling:
The resulting diversity decay can be expressed through the covariance trace:
A Wasserstein generalization bound then relates model risk under the target distribution to the filtered distribution:
The collaborative scoring rule can be viewed through a dual potential :
The formulas explain the paper's main mechanism: biased selection can make the retained distribution increasingly narrow, while collaborative Wasserstein proxies try to reduce the discrepancy between filtered synthetic data and the global target distribution.
Resultsedit
The manuscript reports DDPM-style recursive image-generation experiments on CIFAR-10, STL-10, and CelebA. Baselines include Random selection, K-means, CenterMatch, and CovMatch. Under non-IID or locally skewed references, local-selection baselines can fall behind random selection, while the collaborative schemes better preserve both sample quality and mode coverage.


Placementedit
This work belongs to Synthetic Data, Synthetic Data, Recursive Synthetic Data Training, Data Selection, Sample Selection Bias, Data Silos, Collaborative Evaluation, and Wasserstein Geometry. It is the synthetic-data counterpart to Qiao's unlearning papers: instead of asking how to remove data after training, it asks how selection and verification shape the data stream before future training.