Creates explainable RL agents by evolving programs instead of training neural networks.
This paper propses to make black-box RL agents transparent by evolving readable programs guided by TD3 (Twin Delayed Deep Deterministic Policy Gradient) critics
📚 https://arxiv.org/abs/2410.21940
🎯 Original Problem:
Deep Reinforcement Learning (DRL) agents use black-box neural networks, making them hard to understand and trust. This lack of transparency hinders their adoption in real-world control systems where engineers need guarantees of stability and robustness.
-----
🔧 Solution in this Paper:
→ Combines TD3 (Twin Delayed Deep Deterministic Policy Gradient) with Genetic Programming to create human-readable programs
→ Programs are represented as sequences of real values (genome) encoding operators and literals
→ Uses stack-based execution approach with pre-populated input states
→ TD3 critics guide the genetic algorithm by providing gradients for program optimization
→ Introduces stochasticity in operator mapping to smooth the optimization landscape
-----
💡 Key Insights:
→ Direct optimization through TD3 critics instead of environment interactions improves sample efficiency
→ Programs can influence exploration during training rather than being distilled after
→ Stochastic aspects in program representation create smoother optimization landscape
→ Simple operator set allows for readable yet effective policies
-----
📊 Results:
→ Achieves comparable performance to vanilla TD3
→ Several orders of magnitude more sample-efficient than pure genetic programming
→ Successfully generates interpretable navigation programs for SimpleGoal environment
→ Maintains policy quality while providing explainability
Share this post