Federated Learning (FL) focuses on collaborative model training without sharing private data but faces fairness concerns due to biases in clients' datasets.
Heterogeneous data distributions across clients can lead to unfair models impacting different clients differently.
FeDa4Fair introduces a library to generate datasets and benchmarks for evaluating fairness in FL methods at global and client levels.
The paper aims to support more robust fairness research by facilitating consistent benchmarking and evaluating fairness outcomes for diverse clients.