Aja Huang

is this you? claim profile


  • Bayesian Optimization in AlphaGo

    During the development of AlphaGo, its many hyper-parameters were tuned with Bayesian optimization multiple times. This automatic tuning process resulted in substantial improvements in playing strength. For example, prior to the match with Lee Sedol, we tuned the latest AlphaGo agent and this improved its win-rate from 50 in the final match. Of course, since we tuned AlphaGo many times during its development cycle, the compounded contribution was even higher than this percentage. It is our hope that this brief case study will be of interest to Go fans, and also provide Bayesian optimization practitioners with some insights and inspiration.

    12/17/2018 ∙ by Yutian Chen, et al. ∙ 128 share

    read it

  • Move Evaluation in Go Using Deep Convolutional Neural Networks

    The game of Go is more challenging than other board games, due to the difficulty of constructing a position or move evaluation function. In this paper we investigate whether deep convolutional networks can be used to directly represent and learn this knowledge. We train a large 12-layer convolutional neural network by supervised learning from a database of human professional games. The network correctly predicts the expert move in 55 positions, equalling the accuracy of a 6 dan human player. When the trained convolutional network was used directly to play games of Go, without any search, it beat the traditional search program GnuGo in 97 matched the performance of a state-of-the-art Monte-Carlo tree search that simulates a million positions per move.

    12/20/2014 ∙ by Chris J. Maddison, et al. ∙ 0 share

    read it