DeepAI AI Chat
Log In Sign Up

Feature selection as Monte-Carlo Search in Growing Single Rooted Directed Acyclic Graph by Best Leaf Identification

by   Aurelien Pelissier, et al.

Monte Carlo tree search (MCTS) has received considerable interest due to its spectacular success in the difficult problem of computer Go and also proved beneficial in a range of other domains. A major issue that has received little attention in the MCTS literature is the fact that, in most games, different actions can lead to the same state, that may lead to a high degree of redundancy in tree representation and unnecessary additional computational cost. We extend MCTS to single rooted directed acyclic graph (SR-DAG), and consider the Best Arm Identification (BAI) and the Best Leaf Identification (BLI) problem of an expanding SR-DAG of arbitrary depth. We propose algorithms that are (epsilon, delta)-correct in the fixed confidence setting, and prove an asymptotic upper bounds of sample complexity for our BAI algorithm. As a major application for our BLI algorithm, a novel approach for Feature Selection is proposed by representing the feature set space as a SR-DAG and repeatedly evaluating feature subsets until a candidate for the best leaf is returned, a proof of concept is shown on benchmark data sets.


page 1

page 2

page 3

page 4


Monte-Carlo Graph Search for AlphaZero

The AlphaZero algorithm has been successfully applied in a range of disc...

Monte-Carlo Tree Search by Best Arm Identification

Recent advances in bandit tools and techniques for sequential learning a...

P-MCGS: Parallel Monte Carlo Acyclic Graph Search

Recently, there have been great interests in Monte Carlo Tree Search (MC...

Maximin Action Identification: A New Bandit Framework for Games

We study an original problem of pure exploration in a strategic bandit m...

Monte Carlo Search Algorithm Discovery for One Player Games

Much current research in AI and games is being devoted to Monte Carlo se...

Towards Understanding the Effects of Evolving the MCTS UCT Selection Policy

Monte Carlo Tree Search (MCTS) is a sampling best-first method to search...

Improving multivariate Horner schemes with Monte Carlo tree search

Optimizing the cost of evaluating a polynomial is a classic problem in c...

Code Repositories


Bandit based Reinforcement Learning applied on Feature selection, a Monte Carlo search tree algorithm is trained to find the best feature subset on various training set.

view repo