OpenML Benchmarking Suites and the OpenML100

08/11/2017
by   Bernd Bischl, et al.
0

We advocate the use of curated, comprehensive benchmark suites of machine learning datasets, backed by standardized OpenML-based interfaces and complementary software toolkits written in Python, Java and R. Major distinguishing features of OpenML benchmark suites are (a) ease of use through standardized data formats, APIs, and existing client libraries; (b) machine-readable meta-information regarding the contents of the suite; and (c) online sharing of results, enabling large scale comparisons. As a first such suite, we propose the OpenML100, a machine learning benchmark suite of 100 classification datasets carefully curated from the thousands of datasets available on OpenML.org.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2023

A Suite of Fairness Datasets for Tabular Classification

There have been many papers with algorithms for improving fairness of ma...
research
06/25/2019

Mirovia: A Benchmarking Suite for Modern Heterogeneous Computing

This paper presents Mirovia, a benchmark suite developed for modern day ...
research
05/10/2023

RiverBench: an Open RDF Streaming Benchmark Suite

RDF data streaming has been explored by the Semantic Web community from ...
research
01/11/2019

Machine Learning Automation Toolbox (MLaut)

In this paper we present MLaut (Machine Learning AUtomation Toolbox) for...
research
12/14/2017

Rasa: Open Source Language Understanding and Dialogue Management

We introduce a pair of tools, Rasa NLU and Rasa Core, which are open sou...
research
08/21/2012

Minerva and minepy: a C engine for the MINE suite and its R, Python and MATLAB wrappers

We introduce a novel implementation in ANSI C of the MINE family of algo...
research
10/05/2019

A Benchmark to Evaluate InfiniBand Solutions for Java Applications

Low-latency network interconnects, such as InfiniBand, are commonly used...

Please sign up or login with your details

Forgot password? Click here to reset