SOTAVerified

Parameter Stress Analysis in Reinforcement Learning: Applying Synaptic Filtering to Policy Networks

2026-03-05Unverified0· sign in to hype

Zain ul Abdeen, Ming Jin

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper explores reinforcement learning (RL) policy robustness by systematically analyzing network parameters under internal and external stresses. blackWe apply synaptic filtering methods using high-pass, low-pass, and pulse-wave filters from pravin2024fragility, as an internal stress by selectively perturbing parameters, while adversarial attacks apply external stress through modified agent observations. This dual approach enables the classification of parameters as fragile, robust, or antifragile, based on their influence on policy performance in clean and adversarial settings. Parameter scores are defined to quantify these characteristics, and the framework is validated on proximal policy optimization (PPO)-trained agents in Mujoco continuous control environments. The results highlight the presence of antifragile parameters that enhance policy performance under stress, demonstrating the potential of targeted filtering techniques to improve RL policy adaptability. These insights provide a foundation for future advancements in the design of robust and antifragile RL systems.

Reproductions