Taking Advice from ChatGPT

05/11/2023
by   Peter Zhang, et al.
0

A growing literature studies how humans incorporate advice from algorithms. This study examines an algorithm with millions of daily users: ChatGPT. We conduct a lab experiment in which 118 student participants answer 2,828 multiple-choice questions across 25 academic subjects. We present participants with answers from a GPT model and allow them to update their initial responses. We find that the advisor's identity ("AI chatbot" versus a human "expert"), presence of written justification, and advice correctness do not significant affect weight on advice. Instead, we show that participants weigh advice more heavily if they (1) are unfamiliar with the topic, (2) used ChatGPT in the past, or (3) received more accurate advice previously. These three effects – task difficulty, algorithm familiarity, and experience, respectively – appear to be stronger with an AI chatbot as the advisor. Moreover, we find that participants are able to place greater weight on correct advice only when written justifications are provided. In a parallel analysis, we find that the student participants are miscalibrated and significantly underestimate the accuracy of ChatGPT on 10 of 25 topics. Students under-weigh advice by over 50 and would have scored better if they trusted ChatGPT more.

READ FULL TEXT
research
07/14/2021

Do Humans Trust Advice More if it Comes from AI? An Analysis of Human-AI Interactions

In many applications of AI, the algorithm's output is framed as a sugges...
research
11/07/2022

Do Users Write More Insecure Code with AI Assistants?

We conduct the first large-scale user study examining how users interact...
research
07/23/2020

Diagnostic Questions:The NeurIPS 2020 Education Challenge

Digital technologies are becoming increasingly prevalent in education, e...
research
04/24/2023

AI, write an essay for me: A large-scale comparison of human-written versus ChatGPT-generated essays

Background: Recently, ChatGPT and similar generative AI models have attr...
research
03/25/2022

Anchoring Code Understandability Evaluations Through Task Descriptions

In code comprehension experiments, participants are usually told at the ...
research
01/24/2023

Putting ChatGPT's Medical Advice to the (Turing) Test

Objective: Assess the feasibility of using ChatGPT or a similar AI-based...
research
06/16/2021

Human-AI Interactions Through A Gricean Lens

Grice's Cooperative Principle (1975) describes the implicit maxims that ...

Please sign up or login with your details

Forgot password? Click here to reset