0:00
/
0:00
Transcript

"Fine Tuning Large Language Models to Deliver CBT for Depression"

The podcast on this paper is generated with Google's Illuminate.

This paper demonstrates how fine-tuning small 7-8B parameter LLMs can effectively deliver Cognitive Behavioral Therapy (CBT) for depression, achieving significant improvements in therapeutic competencies compared to base models.

-----

https://arxiv.org/abs/2412.00251

🤖 Original Problem:

Depression affects 20.6% of Americans, but many can't access CBT therapy due to cost, stigma, and therapist scarcity. AI could help, but current solutions using prompt engineering are limited and brittle.

-----

🔧 Solution in this Paper:

→ Researchers fine-tuned three small LLMs (Mistral 7B, Qwen 7B, Llama 8B) on 58 sets of synthetic CBT transcripts.

→ Each transcript set contained 20 therapy sessions, designed with proper phase progression and CBT techniques.

→ Models were evaluated through simulated patient interactions using DeepSeek-V2.5 as the patient.

→ Performance was measured using a modified Cognitive Therapy Rating Scale (CTRS).

-----

💡 Key Insights:

→ Small LLMs can effectively learn therapeutic competencies through targeted fine-tuning

→ CBT-tuned models maintained better session structure and technique implementation

→ Models showed strong empathy but struggled with agenda adherence and deep exploration

→ Performance degraded as context length approached 4000 tokens

-----

📊 Results:

→ CBT-tuned models outperformed base versions by 11.33 points on CTRS (p<0.001)

→ Llama 8B performed best with mean CTRS score of 67.86 ±7.24

→ All models showed significant improvements in therapeutic competencies

→ Strongest gains in Agenda Setting, Guided Discovery, and CBT technique application

Discussion about this video