Fair AI/ML 1 Core-Stability Federated Learning
I will begin a new project work 24 Fall: FAIRNESS IN AI/ML THROUGH THE LENS OF SOCIAL CHOICE
The basic outline is following Tutorial: Fairness in AI/ML via Social Choice by Evi Micha & Nisarg Shah in EC 2024.
The first paper I have read in last semester, Fairness in Federated Learning via Core-Stability, which was one of the possible topics.
Brief notes is here.
⚠️We found that the basis of this paper may make no sense!!!
I will try to understand some mathematical proof in this semester.
Core-Stability
Revisit the definition of Core-Stability: A predictor is called core stable if there exist no other and no subset of agents, s.t. for all , with at least one strict inequality.
⚠️⚠️⚠️What is WRONG with the claim
Setting in definition, there is no predictor where .
The problem is, according to the definition, when we plug in , there exist no s.t.
with at least one strict inequality.
But can it be directly sum up? ???
A counter example is here:
Assume is core-stable and ,
According to the definition, there exist no (We plug in ) s.t.
with at least one strict inequality.
But according to definition, there may exist s.t.
summing them up, we have
Scaled factor suggests some kind of proportionality here. Basically core-stability means there is no big benefit for agent to form a coalition.
Core means “no deviating subgroup”
Core stability is very strong, only exists in special cases.
Advantage
More fair than FedAvg. is a very large integer.
Agent 1 | Agent 2 | Agent 3 | |
---|---|---|---|
In FedAvg, is preferred, but not fair to Agent 1 and 2. (They have incentive to collide.)
Robustness to poor local data quality of agents.
❗Scaling the utility of any single agents does not change the core-stable allocation.
Core-Stability in FL
CoreFed exists in Linear Regression and Logistic Regression (Classification)
Approximate core stable predictor exists in Deep Neural Network.
Existence
🤔Core-Stability exist in FL if:
Utility function is continuous.
The conical combination of utility functions is convex.
, the set is convex.
🤔Important theorem: Kakutani’s Fixed Point Theorem: A set valued function admits a fixed point, i.e. , s.t. , if
- is non-empty, compact, and convex
Compact: The set is closed (It must contains all its boundary points), and bounded (It can be covered by a “sphere” with certain radius)
Convex: A set is convex if, for any two points and in the set, the line segment connecting and also lies entirely within the set. Mathematically, for any and any , the point: .
- is non-empty, compact, and convex.
- has a closed graph. i.e., sequence converging to and converging to , s.t. and , we have .
Define Core-Stable Feasible Predictors
, .
Lemma 4.1: Let be s.t. (It is a fixed point). Then is a core-stable predictor.
Proof by contradiction.
By Takutani’s Theorem, it suffices to show that has a closed graph, to ensure that it admits a fixed point.
Lemma 4.2: has a closed graph.
Hard proof.
🤔Theorem 1. In FL, where utilities are continuous and any conical combination of utilities is convex, a core-stable predictor exists.
⚠️If we change to , if the two condition still holds within a radius of of every point. The conditions tend to be true in DNN of small . Then the predictor is locally core-stable.
Computation of Core-Stable Predictor
There is efficient way to compute core-stable predictor.
It is optima of convex program:
The object reminds me of Nash Social Welfare😎
Utility of each agent is concave, thus above program is convex. Thus this program is concave maximization subject to convex constraint.
🤔Theorem 2. If is concave for all , then any predictor maximizes the convex program is core stable.
Proof by contradiction
If is not core-stable, there must exist and subset , s.t.
for all with at least one strict inequality.
Thus for any ,
admitting contradiction.
must be Pareto Optimal.
CoreFed
finite samples drawn from distribution .
Empirical distribution .
Stochastic Gradient Descent.
…
To be continued~