SOTAVerified

Optimizing Attention and Cognitive Control Costs Using Temporally-Layered Architectures

2023-05-30Code Available0· sign in to hype

Devdhar Patel, Terrence Sejnowski, Hava Siegelmann

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The current reinforcement learning framework focuses exclusively on performance, often at the expense of efficiency. In contrast, biological control achieves remarkable performance while also optimizing computational energy expenditure and decision frequency. We propose a Decision Bounded Markov Decision Process (DB-MDP), that constrains the number of decisions and computational energy available to agents in reinforcement learning environments. Our experiments demonstrate that existing reinforcement learning algorithms struggle within this framework, leading to either failure or suboptimal performance. To address this, we introduce a biologically-inspired, Temporally Layered Architecture (TLA), enabling agents to manage computational costs through two layers with distinct time scales and energy requirements. TLA achieves optimal performance in decision-bounded environments and in continuous control environments, it matches state-of-the-art performance while utilizing a fraction of the compute cost. Compared to current reinforcement learning algorithms that solely prioritize performance, our approach significantly lowers computational energy expenditure while maintaining performance. These findings establish a benchmark and pave the way for future research on energy and time-aware control.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Ant-v2TLAMean Reward5,163.54Unverified
HalfCheetah-v2TLAMean Reward9,571.99Unverified
Hopper-v2TLAMean Reward3,458.22Unverified
InvertedDoublePendulum-v2TLAMean Reward9,356.67Unverified
InvertedPendulum-v2TLAMean Reward1,000Unverified
MountainCarContinuous-v0TLAMean Reward93.88Unverified
Pendulum-v1TLAMean Reward-154.92Unverified
Walker2d-v2TLAMean Reward3,878.41Unverified

Reproductions