0:00
/
0:00
Transcript

"Improving LLM Group Fairness on Tabular Data via In-Context Learning"

The podcast on this paper is generated with Google's Illuminate.

LLMs can make fair predictions on tabular data through optimized prompting and example selection

This paper introduces four effective approaches to improve fairness in LLM predictions on tabular data while maintaining high performance. The methods include fair prompt optimization, soft prompt tuning, strategic few-shot example selection, and self-refining predictions through chain-of-thought reasoning.

-----

https://arxiv.org/abs/2412.04642

🤔 Original Problem:

LLMs show bias when making predictions on tabular data, producing unfair outcomes across demographic groups. Traditional debiasing methods for natural language tasks don't work well for tabular data fairness.

-----

🛠️ Solution in this Paper:

→ The paper implements fair prompt optimization using meta-LLMs to generate fairness-specific instructions

→ It develops soft prompt tuning by optimizing tokens in embedding space with fairness penalties

→ It creates strategic few-shot examples by selecting nearest neighbors and manipulating positive/negative example ratios

→ It introduces self-refinement where LLMs analyze and adjust their own predictions to achieve demographic parity

-----

💡 Key Insights:

→ Fairness in tabular tasks differs from NLP fairness - it focuses on class label distributions across groups

→ Simple prompt engineering can significantly improve fairness without sacrificing accuracy

→ Strategic selection of in-context examples outperforms random selection for fairness

-----

📊 Results:

→ Fair prompt optimization achieved 0.94-0.99 demographic parity ratio across datasets

→ Strategic few-shot examples maintained 90%+ accuracy while improving fairness by 30-40%

→ Self-refinement improved demographic parity by 20-30% with minimal accuracy loss

Discussion about this video