Language models are weak learners

06/25/2023
by   Hariharan Manikandan, et al.
0

A central notion in practical and theoretical machine learning is that of a weak learner, classifiers that achieve better-than-random performance (on any given distribution over data), even by a small margin. Such weak learners form the practical basis for canonical machine learning methods such as boosting. In this work, we illustrate that prompt-based large language models can operate effectively as said weak learners. Specifically, we illustrate the use of a large language model (LLM) as a weak learner in a boosting algorithm applied to tabular data. We show that by providing (properly sampled according to the distribution of interest) text descriptions of tabular data samples, LLMs can produce a summary of the samples that serves as a template for classification and achieves the aim of acting as a weak learner on this task. We incorporate these models into a boosting approach, which in some settings can leverage the knowledge within the LLM to outperform traditional tree-based boosting. The model outperforms both few-shot learning and occasionally even more involved fine-tuning procedures, particularly for tasks involving small numbers of data points. The results illustrate the potential for prompt-based LLMs to function not just as few-shot learners themselves, but as components of larger machine learning pipelines.

READ FULL TEXT
research
08/08/2016

Boosting as a kernel-based method

Boosting combines weak (biased) learners to obtain effective learning al...
research
09/04/2022

ProBoost: a Boosting Method for Probabilistic Classifiers

ProBoost, a new boosting algorithm for probabilistic classifiers, is pro...
research
08/23/2023

PREFER: Prompt Ensemble Learning via Feedback-Reflect-Refine

As an effective tool for eliciting the power of Large Language Models (L...
research
06/27/2012

An Online Boosting Algorithm with Theoretical Justifications

We study the task of online boosting--combining online weak learners int...
research
01/31/2020

Boosting Simple Learners

We consider boosting algorithms under the restriction that the weak lear...
research
01/23/2023

The Impossibility of Parallelizing Boosting

The aim of boosting is to convert a sequence of weak learners into a str...
research
04/29/2021

Entailment as Few-Shot Learner

Large pre-trained language models (LMs) have demonstrated remarkable abi...

Please sign up or login with your details

Forgot password? Click here to reset