Discussion about this post

User's avatar
Hdberb's avatar

I was also interested in that headline since viewing chess similar to an LLM (using the position as the context vector, then using a transformer to predict the next move). After reading your skepticism, here are my thoughts:

1. Why does it matter if it was trained on stockfish moves rather than high level games? LLM tend to imitate human text, but I've seen papers on them being trained on multiplication (which can be done by a calculator), wouldn't stockfish be analogous to that since it far exceeds humans in chess.

I agree, it would be quite interesting to see a transformer that predicts which moves a human would make (https://maiachess.com/). Perhaps this paper is not that groundbreaking since LeelaChess is 2700 on Lichess using depth 1 (https://lichess.org/@/LazyBot).

2. I agree with the part about FEN, a way to mitigate that would be to include bins for mate in <3, <6, <12, <24. This will find mate is most simple endgames (perhaps not Bishop + Knight).

I think the most exciting possible use of LLM research on chess would be a way to store chess positions (and games) in a vector database. Chess2Vec already exists, but for instance an expanded version of Lichess' opening explorer could be used to find similar positions/games in the middle/end game. A chess win probability model that accounts for rating would also be interesting, perhaps by using NN on similar positions or MCTS or Markov Chains if possible.

Expand full comment
Anian Ruoss's avatar

Thank you for your interest in our paper, Arjun!

We actually beat a GM 6-0 in blitz.

Moreover, our bot played against humans of all levels, from beginners all the way to NMs, IMs, and GMs. So far, our bot only lost 2 games (out of 365).

Does that alleviate your concerns?

Expand full comment

No posts