Fine-tuning often leads to catastrophic forgetting. Where the model loses its ability to generalize across other tasks.
Share this post
LINES: POST-TRAINING LAYER SCALING PREVENTS…
Share this post
Fine-tuning often leads to catastrophic forgetting. Where the model loses its ability to generalize across other tasks.