ChatGPT and Software Testing Education: Promises Perils

02/07/2023
by   Sajed Jalil, et al.
0

Over the past decade, predictive language modeling for code has proven to be a valuable tool for enabling new forms of automation for developers. More recently, we have seen the advent of general purpose "large language models", based on neural transformer architectures, that have been trained on massive datasets of human written text spanning code and natural language. However, despite the demonstrated representational power of such models, interacting with them has historically been constrained to specific task settings, limiting their general applicability. Many of these limitations were recently overcome with the introduction of ChatGPT, a language model created by OpenAI and trained to operate as a conversational agent, enabling it to answer questions and respond to a wide variety of commands from end users. The introduction of models, such as ChatGPT, has already spurred fervent discussion from educators, ranging from fear that students could use these AI tools to circumvent learning, to excitement about the new types of learning opportunities that they might unlock. However, given the nascent nature of these tools, we currently lack fundamental knowledge related to how well they perform in different educational settings, and the potential promise (or danger) that they might pose to traditional forms of instruction. As such, in this paper, we examine how well ChatGPT performs when tasked with answering common questions in a popular software testing curriculum. Our findings indicate that ChatGPT can provide correct or partially correct answers in 55.6 or partially correct explanations of answers in 53.0 prompting the tool in a shared question context leads to a marginally higher rate of correct responses. Based on these findings, we discuss the potential promises and perils related to the use of ChatGPT by students and instructors.

READ FULL TEXT

page 1

page 4

page 5

page 6

research
03/09/2023

Large Language Models (GPT) Struggle to Answer Multiple-Choice Questions about Code

We analyzed effectiveness of three generative pre-trained transformer (G...
research
05/30/2022

Automatic Short Math Answer Grading via In-context Meta-learning

Automatic short answer grading is an important research direction in the...
research
09/19/2023

Exploring Self-Reinforcement for Improving Learnersourced Multiple-Choice Question Explanations with Large Language Models

Learnersourcing involves students generating and sharing learning resour...
research
03/17/2021

Towards a question answering assistant for software development using a transformer-based language model

Question answering platforms, such as Stack Overflow, have impacted subs...
research
08/08/2023

Generative AI in Computing Education: Perspectives of Students and Instructors

Generative models are now capable of producing natural language text tha...
research
03/02/2023

AI and the FCI: Can ChatGPT Project an Understanding of Introductory Physics?

ChatGPT is a groundbreaking “chatbot"–an AI interface built on a large l...
research
04/21/2023

Generative AI Perceptions: A Survey to Measure the Perceptions of Faculty, Staff, and Students on Generative AI Tools in Academia

ChatGPT is a natural language processing tool that can engage in human-l...

Please sign up or login with your details

Forgot password? Click here to reset