In AlphaGo, the value network takes a board position and estimates the likelihood of a win. The value network is one of two neural networks that act as a heuristic functions to guide a Monte Carlo tree search of game positions. It is used in conjunction with a policy network. While introduced in AlphaGo, the same principle can be applied generally in game playing.
Used in Chap. 11: page 157