Othello (Reversi) is one of the most popular board games in the world. It is a two-player abstract strategy game, played using 64 round disks, each colored dark on one side and light on the other, and a square, 8 by 8, uncheckered board. The players are traditionally referred to as dark and light, with dark making the first move by convention.
The starting position occurs with 4 discs on the 4 central squares of the grid, 2 with dark sides facing up and 2 with light sides up, such that both discs of each color rest on intersecting diagonals. The dark player now makes a move by placing a dark disc on the board so that there exists at least one straight horizontal, vertical, or diagonal occupied line between the new disc and another dark piece, with one or more contiguous light pieces between them, a theme known as outflanking. After the move, the dark player captures all light pieces on the line between the new piece and an anchoring dark piece, by turning over those discs so that they now face dark sides up. Light makes the next move and play alternates until a player on move has no legal moves or the board is completely filled. The player with the greater number of discs facing up at the end wins. More information about the game can be found here, or alternatively, in this video tutorial.
The game logic was implemented using object-oriented programming in Python. This implementation allows gameplay on all square board sizes ranging from 4 by 4 to 11 by 11 inclusive, as chosen by the user. The user can also choose one of 4 options to play each side: Human, Randy, Minimax or Alphabeta. Hence, the application allows a human player to play against another human, as well as to visualize games between two engine instances in real time. The application has been programmed to show all possible moves in a given position, using red circles, so that it becomes user-friendly, even for human players who may be complete beginners in the game. The interface also shows the move history using standard Othello board notation so players can keep track of their previous moves and moves by engines.
Randy is a random engine that is based on the Python Random library, and selects a move at random in any given position.
Minimax is a perfectly calculating artificially intelligent engine that uses the recursive algorithm, minimax, for turn-based two player games, under the assumptions of the deterministic, adversarial, zero sum and fully observable game model. This engine calculates the full game tree from every possible position, considering every possible move by both sides, at every step, in its search for the best possible move in each position. It is impossible for human players to beat this engine unless one plays the objectively perfect moves in every position. Unfortunately, since game trees become exponentially more complex with increasing board size, the engine tends to approach infeasible runtimes in finding best moves on larger boards. For this reason, a timeout feature was incorporated in this implementation, that causes any engine to lose by timeout if it is unable to make a move within 30 seconds in any given position. In this scenario, a minimax engine will time out on any board larger than 4 by 4, because it becomes infeasible to brute-force search the objectively strongest move in all positions within 30 seconds in such cases, by expanding the full game tree. Interestingly, Minimax can also occasionally lose to Randy on a 4 by 4 board, when playing the dark pieces, because the player moving second has an inherent advantage in such a case and Randy may make the best moves out of pure coincidence since there are a very small number of moves to choose from. More information about the minimax algorithm can be found here.
The timeout limitation of Minimax is solved by introducing the Alphabeta engine. Alphabeta is a Minimax engine equipped with Alpha-beta pruning, a modern, state-of-the-art artificial intelligence algorithm that allows the pruning of game trees by eliminating and scrapping complete branches and sub-branches of a game tree that are not relevant to the best move, thereby bypassing massive potential wastages in calculation time and resources. This is somewhat analogous to a human player automatically discarding obviously bad moves at the start of each new position, by intuition, and focusing deeply on calculating variations arising from better candidate moves instead. Furthermore, move ordering was accomplished in the Alphabeta engine by referencing the Heapq Python module, as this helps ordering possible moves by perceived move utility, greatly optimizing the pruning process, and fine-tuning the effectiveness of the algorithm so that pruning becomes much more likely, thereby reducing wasted time and resources. Unfortunately, even with these modifications, a Minimax engine equipped with Alpha-beta pruning and heap move ordering is not always able to find the best move in every position within 30 seconds for large boards and complex positions. To address this, depth-limited Alpha-beta pruning was further effectuated in the Alphabeta engine. This feature enables the algorithm to recursively expand the game tree up to a certain depth and stop the search there by treating the resulting positions as though they were terminal positions. For instance, if the Alphabeta engine runs on a depth of 10 on a given position, it will calculate all possible variations arising out of the position up to 10 plies (moves) after the current position, and pick the move that leads to the best possible position for a side after 10 moves. That is, the engine picks the move to maximize the number of discs of the side on move, 10 moves down the line from the given position, without knowledge or calculation further down the game tree. Thus, in this scenario, the engine no longer makes the optimal moves but makes the moves that lead to the best position after 10 moves instead, for example. From analysis, such moves quite frequently tend to be the objectively strongest moves in every position (even though the game tree was not fully traversed) or reasonably very strong moves otherwise, except in rare cases, owing to the calculating power of the engine. This creates an interesting dynamic between the human player and the engine, which makes it difficult for all but the strongest human players to occasionally beat the Alphabeta engine. Thus, it poses an interesting challenge for players, while also making the engine practical to run, in terms of time taken to calculate move variations. From trial-and-error experience, a maximum depth of 5 seemed suitable to run Alphabeta for all board sizes up to and including the upper limit, while posing enough of a challenge to experienced human players, and so, that depth has been used to run the Alphabeta engine for all board sizes, in this application. More information about the Alpha-beta pruning algorithm can be found here.
This implementation uses Flask on the backend, deployed online using Heroku, and React on the frontend, which was installed on Netlify.