MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning

07/16/2023
by   Zhenwen Liang, et al.
0

Reasoning in mathematical domains remains a significant challenge for relatively small language models (LMs). Many current methods focus on specializing LMs in mathematical reasoning and rely heavily on knowledge distillation from powerful but inefficient large LMs (LLMs). In this work, we explore a new direction that avoids over-reliance on LLM teachers, introducing a multi-view fine-tuning method that efficiently exploits existing mathematical problem datasets with diverse annotation styles. Our approach uniquely considers the various annotation formats as different "views" and leverages them in training the model. By postpending distinct instructions to input questions, models can learn to generate solutions in diverse formats in a flexible manner. Experimental results show that our strategy enables a LLaMA-7B model to outperform prior approaches that utilize knowledge distillation, as well as carefully established baselines. Additionally, the proposed method grants the models promising generalization ability across various views and datasets, and the capability to learn from inaccurate or incomplete noisy data. We hope our multi-view training paradigm could inspire future studies in other machine reasoning domains.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/14/2023

CoRF : Colorizing Radiance Fields using Knowledge Distillation

Neural radiance field (NeRF) based methods enable high-quality novel-vie...
research
06/07/2020

Multi-view Contrastive Learning for Online Knowledge Distillation

Existing Online Knowledge Distillation (OKD) aims to perform collaborati...
research
03/25/2023

Multi-view knowledge distillation transformer for human action recognition

Recently, Transformer-based methods have been utilized to improve the pe...
research
05/20/2023

DisCo: Distilled Student Models Co-training for Semi-supervised Text Mining

Many text mining models are constructed by fine-tuning a large deep pre-...
research
10/21/2022

Multi-View Reasoning: Consistent Contrastive Learning for Math Word Problem

Math word problem solver requires both precise relation reasoning about ...
research
08/03/2023

Scaling Relationship on Learning Mathematical Reasoning with Large Language Models

Mathematical reasoning is a challenging task for large language models (...
research
05/27/2023

FERMAT: An Alternative to Accuracy for Numerical Reasoning

While pre-trained language models achieve impressive performance on vari...

Please sign up or login with your details

Forgot password? Click here to reset