A quasi-Monte Carlo data compression algorithm for machine learning

04/06/2020 ∙ by Josef Dick, et al. ∙ 0

We introduce an algorithm to reduce large data sets using so-called digital nets, which are well distributed point sets in the unit cube. These point sets together with weights, which depend on the data set, are used to represent the data. We show that this can be used to reduce the computational effort needed in finding good parameters in machine learning algorithms. To illustrate our method we provide some numerical examples for neural networks.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 23

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.