pith. machine review for the scientific record. sign in

arxiv: 1511.00830 · v6 · submitted 2015-11-03 · 📊 stat.ML · cs.LG

Recognition: unknown

The Variational Fair Autoencoder

Authors on Pith no claims yet
classification 📊 stat.ML cs.LG
keywords latentvariationdatafactorsremainingrepresentationssensitivevariational
0
0 comments X
read the original abstract

We investigate the problem of learning representations that are invariant to certain nuisance or sensitive factors of variation in the data while retaining as much of the remaining information as possible. Our model is based on a variational autoencoding architecture with priors that encourage independence between sensitive and latent factors of variation. Any subsequent processing, such as classification, can then be performed on this purged latent representation. To remove any remaining dependencies we incorporate an additional penalty term based on the "Maximum Mean Discrepancy" (MMD) measure. We discuss how these architectures can be efficiently trained on data and show in experiments that this method is more effective than previous work in removing unwanted sources of variation while maintaining informative latent representations.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 3 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Fair Dataset Distillation via Cross-Group Barycenter Alignment

    cs.LG 2026-04 unverdicted novelty 6.0

    Dataset distillation introduces fairness gaps from subgroup pattern mismatches rather than just imbalance; distilling to a group-agnostic barycenter of predictive information reduces these gaps.

  2. Mitigating Shortcut Learning via Feature Disentanglement in Medical Imaging: A Benchmark Study

    cs.CV 2026-02 unverdicted novelty 6.0

    Benchmark shows that combining data rebalancing with feature disentanglement mitigates shortcut learning more effectively than rebalancing alone in medical imaging models.

  3. Distributed Deep Variational Approach for Privacy-preserving Data Release

    cs.CR 2026-05 unverdicted novelty 5.0

    GPP trains local variational encoders in federated settings to release representations that keep utility within 1% of an autoencoder baseline while driving adversary AUC on sensitive attributes to near-random levels o...