Trans-LoRA enables data-free, lossless transfer of LoRA modules between different LLM architectures.
- Works for LoRA, DoRA, and Prompt Tuning PEFT methods
- Achieves up to 10% performance improvement in some cases
π https://arxiv.org/pdf/2405.17258
Original Problem π:
Transferring LoRA modules between base models without access to original training data, crucial for cloud applications where providers can't retain client data.
-----
Solution in this Paper π οΈ:
β’ Trans-LoRA: Uses synthetic data for LoRA transfer
β’ Components:
- LLM-based synthetic data generator
- Discriminator for filtering synthetic data
- Knowledge distillation from source to target LoRA
β’ Operates with minimal seed examples (5 samples)
β’ Supports transfer across model families and PEFT types
-----
Key Insights from this Paper π‘:
β’ Synthetic data can effectively replace original training data for LoRA transfer
β’ Discriminator filtering crucial for matching source LoRA training distribution
β’ Lossless or improved transfer possible across different base models and PEFT methods
β’ Scalable performance with increased synthetic data generation
-----
Results π:
β’ Outperforms source LoRAs and target base models consistently
β’ Effective across Llama2 and Gemma model familiesβ’
β’ Supports continuous transfer through multiple models without degradation