SOTAVerified

Soft Actor-Critic for Discrete Action Settings

2019-10-16Code Available0· sign in to hype

Petros Christodoulou

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Soft Actor-Critic is a state-of-the-art reinforcement learning algorithm for continuous action settings that is not applicable to discrete action settings. Many important settings involve discrete actions, however, and so here we derive an alternative version of the Soft Actor-Critic algorithm that is applicable to discrete action settings. We then show that, even without any hyperparameter tuning, it is competitive with the tuned model-free state-of-the-art on a selection of games from the Atari suite.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Atari 2600 AlienSACScore216.9Unverified
Atari 2600 AmidarSACScore7.9Unverified
Atari 2600 AssaultSACScore350Unverified
Atari 2600 AsterixSACScore272Unverified
Atari 2600 Battle ZoneSACScore4,386.7Unverified
Atari 2600 Beam RiderSACScore432.1Unverified
Atari 2600 BreakoutSACScore0.7Unverified
Atari 2600 Crazy ClimberSACScore3,668.7Unverified
Atari 2600 EnduroSACScore0.8Unverified
Atari 2600 FreewaySACScore4.4Unverified
Atari 2600 FrostbiteSACScore59.4Unverified
Atari 2600 James BondSACScore68.3Unverified
Atari 2600 KangarooSACScore29.3Unverified
Atari 2600 Ms. PacmanSACScore690.9Unverified
Atari 2600 PongSACScore-20.98Unverified
Atari 2600 Q*BertSACScore280.5Unverified
Atari 2600 Road RunnerSACScore305.3Unverified
Atari 2600 SeaquestSACScore211.6Unverified
Atari 2600 Space InvadersSACScore160.8Unverified
Atari 2600 Up and DownSACScore250.7Unverified

Reproductions