0:00
/
0:00
Transcript

"ICLR: In-Context Learning of Representations"

Generated below podcast on this paper with Google's Illuminate.

LLMs don't just memorize word meanings, they can relearn them based on new examples

LLMs can learn new meanings for words through in-context examples, reshaping their internal representations to match the new context-specific semantics.

-----

https://arxiv.org/abs/2501.00070

🤔 Original Problem:

→ LLMs have fixed semantic representations learned during pretraining, but need flexibility to handle novel meanings in different contexts

-----

🔧 Solution in this Paper:

→ Designed a "graph tracing" task where familiar words (like apple, bird) are arranged in specific graph structures (grid, ring, hexagonal)

→ Fed the model with examples of random walks on these graphs to see if it can learn new relationships between words

→ Analyzed how internal representations reorganize using Principal Component Analysis and Dirichlet energy measurements

→ Used Llama-3.1-8B as primary model, with additional experiments on other variants

-----

💡 Key Insights:

→ LLMs can dynamically alter their pretrained word representations based purely on in-context examples

→ Representation reorganization happens suddenly at a critical context length, following power-law scaling

→ When words have strong semantic links (like days of week), new structure emerges in higher dimensions without fully overriding original meanings

→ Process follows energy minimization patterns similar to physical systems

-----

📊 Results:

→ Model achieves high accuracy (>80%) in predicting valid next nodes after seeing sufficient examples

→ Dirichlet energy of representations decreases before accuracy improvements, indicating structure learning

→ Performance scales with graph size following power-law relationship

------

Are you into AI and LLMs❓ Join me on X/Twitter with 52K+ others, to remain on the bleeding-edge of AI every day.

𝕏/🐦 https://x.com/rohanpaul_ai

Discussion about this video

User's avatar