Inductive-bias Learning: Generating Code Models with Large Language Model

08/19/2023
by   Toma Tanaka, et al.
0

Large Language Models(LLMs) have been attracting attention due to a ability called in-context learning(ICL). ICL, without updating the parameters of a LLM, it is possible to achieve highly accurate inference based on rules “in the context” by merely inputting a training data into the prompt. Although ICL is a developing field with many unanswered questions, LLMs themselves serves as a inference model, seemingly realizing inference without explicitly indicate “inductive bias”. On the other hand, a code generation is also a highlighted application of LLMs. The accuracy of code generation has dramatically improved, enabling even non-engineers to generate code to perform the desired tasks by crafting appropriate prompts. In this paper, we propose a novel “learning” method called an “Inductive-Bias Learning (IBL)”, which combines the techniques of ICL and code generation. An idea of IBL is straightforward. Like ICL, IBL inputs a training data into the prompt and outputs a code with a necessary structure for inference (we referred to as “Code Model”) from a “contextual understanding”. Despite being a seemingly simple approach, IBL encompasses both a “property of inference without explicit inductive bias” inherent in ICL and a “readability and explainability” of the code generation. Surprisingly, generated Code Models have been found to achieve predictive accuracy comparable to, and in some cases surpassing, ICL and representative machine learning models. Our IBL code is open source: https://github.com/fuyu-quant/IBLM

READ FULL TEXT

page 11

page 12

page 13

page 14

page 15

page 16

page 17

research
05/23/2023

Generating Data for Symbolic Language with Large Language Models

While large language models (LLMs) bring not only performance but also c...
research
03/14/2022

EIT: Efficiently Lead Inductive Biases to ViT

Vision Transformer (ViT) depends on properties similar to the inductive ...
research
07/07/2023

QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models

We present ongoing work on a new automatic code generation approach for ...
research
08/25/2023

Does Asking Clarifying Questions Increases Confidence in Generated Code? On the Communication Skills of Large Language Models

Large language models (LLMs) have significantly improved the ability to ...
research
09/08/2022

Data Feedback Loops: Model-driven Amplification of Dataset Biases

Datasets scraped from the internet have been critical to the successes o...
research
11/15/2022

Evaluating How Fine-tuning on Bimodal Data Effects Code Generation

Despite the increase in popularity of language models for code generatio...
research
10/10/2022

SimSCOOD: Systematic Analysis of Out-of-Distribution Behavior of Source Code Models

While large code datasets have become available in recent years, acquiri...

Please sign up or login with your details

Forgot password? Click here to reset