To Pool or Not To Pool: Analyzing the Regularizing Effects of Group-Fair Training on Shared Models
CoRR(2024)
摘要
In fair machine learning, one source of performance disparities between
groups is over-fitting to groups with relatively few training samples. We
derive group-specific bounds on the generalization error of welfare-centric
fair machine learning that benefit from the larger sample size of the majority
group. We do this by considering group-specific Rademacher averages over a
restricted hypothesis class, which contains the family of models likely to
perform well with respect to a fair learning objective (e.g., a power-mean).
Our simulations demonstrate these bounds improve over a naive method, as
expected by theory, with particularly significant improvement for smaller group
sizes.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要