SOTAVerified

Learning Chess Blindfolded

2021-01-01Unverified0· sign in to hype

Shubham Toshniwal, Sam Wiseman, Karen Livescu, Kevin Gimpel

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Transformer language models have made tremendous strides in natural language understanding. However, the complexity of natural language makes it challenging to ascertain how accurately these models are tracking the world state underlying the text. Motivated by this issue, we consider the task of language modeling for the game of chess. Unlike natural language, chess notations describe a simple, constrained, and deterministic domain. Probing for the world (board) state can be done simply via language modeling prompts. Additionally, we have access to a vast number of chess games coupled with the exact state at every move, allowing us to measure the impact of various ways of including grounding during language model training. Overall, we find that with enough training data, transformers can learn to track pieces and predict legal moves when trained solely from move sequences. However, in adverse circumstances (small training sets or prediction following long move histories), providing access to board state information during training can yield consistent improvements.

Tasks

Reproductions