Chess Transformer is a language model trained to play chess. The neural network predicts the next move based on the history of the moves in the game. The Transformer model was trained on 2.8 million chess games in Portable Game Notation. The developers have published a Colab laptop in which you can play a game of chess with a model.
The GPT-2 with 774 million parameters was used as the transformer architecture. GPT-2 is a state-of-the-art generative model for natural language processing tasks. The model was developed by researchers from OpenAI.
The model was retrained in 30 thousand steps. The trained model correctly filters out invalid moves and demonstrates defense strategies such as Slav Exchange.
Chess Transformer is an example of how transformer models, initially adapted to work with NLP problems, can solve more general strategic modeling problems.
Interested in Deep Learning?
If you found this article helpful, click the💚 or 👏 button below or share the article on Facebook so your friends can benefit from it too.