Parameter Stress Analysis in Reinforcement Learning: Applying Synaptic Filtering to Policy Networks
Zain ul Abdeen, Ming Jin
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
This paper explores reinforcement learning (RL) policy robustness by systematically analyzing network parameters under internal and external stresses. blackWe apply synaptic filtering methods using high-pass, low-pass, and pulse-wave filters from pravin2024fragility, as an internal stress by selectively perturbing parameters, while adversarial attacks apply external stress through modified agent observations. This dual approach enables the classification of parameters as fragile, robust, or antifragile, based on their influence on policy performance in clean and adversarial settings. Parameter scores are defined to quantify these characteristics, and the framework is validated on proximal policy optimization (PPO)-trained agents in Mujoco continuous control environments. The results highlight the presence of antifragile parameters that enhance policy performance under stress, demonstrating the potential of targeted filtering techniques to improve RL policy adaptability. These insights provide a foundation for future advancements in the design of robust and antifragile RL systems.