JaCoText: A Pretrained Model for Java Code-Text Generation

03/22/2023
by   Jessica López Espejel, et al.
0

Pretrained transformer-based models have shown high performance in natural language generation task. However, a new wave of interest has surged: automatic programming language generation. This task consists of translating natural language instructions to a programming code. Despite the fact that well-known pretrained models on language generation have achieved good performance in learning programming languages, effort is still needed in automatic code generation. In this paper, we introduce JaCoText, a model based on Transformers neural network. It aims to generate java source code from natural language text. JaCoText leverages advantages of both natural language and code generation models. More specifically, we study some findings from the state of the art and use them to (1) initialize our model from powerful pretrained models, (2) explore additional pretraining on our java dataset, (3) carry out experiments combining the unimodal and bimodal data in the training, and (4) scale the input and output length during the fine-tuning of the model. Conducted experiments on CONCODE dataset show that JaCoText achieves new state-of-the-art results.

READ FULL TEXT
research
04/23/2020

QURIOUS: Question Generation Pretraining for Text Generation

Recent trends in natural language processing using pretraining have shif...
research
06/10/2023

A Comprehensive Review of State-of-The-Art Methods for Java Code Generation from Natural Language Text

Java Code Generation consists in generating automatically Java code from...
research
02/16/2022

Probing Pretrained Models of Source Code

Deep learning models are widely used for solving challenging code proces...
research
12/02/2019

GANCoder: An Automatic Natural Language-to-Programming Language Translation Approach based on GAN

We propose GANCoder, an automatic programming approach based on Generati...
research
04/16/2021

Text2App: A Framework for Creating Android Apps from Text Descriptions

We present Text2App – a framework that allows users to create functional...
research
03/10/2023

Contrastive Language-Image Pretrained (CLIP) Models are Powerful Out-of-Distribution Detectors

We present a comprehensive experimental study on pretrained feature extr...
research
06/22/2021

On Adversarial Robustness of Synthetic Code Generation

Automatic code synthesis from natural language descriptions is a challen...

Please sign up or login with your details

Forgot password? Click here to reset