Advanced Search

Generative-AI chatbots can lie to elicit approval from users
By
15 January 2024 (Edited )
Share this article:

Even if the user's view is not objectively true

Generative-AI chatbots can lie to elicit approval from users

A new report from the Nielsen Norman Group describes the phenomenon of sycophancy in large language models: they will do whatever they can to win the user's approval, including adapting their responses to agree with the user's point of view, even if that point of view is objectively incorrect.

This happens because language models like ChatGPT are built and trained to deliver responses that are highly rated by users. Sometimes the best way to bring that result about is to lie. Researchers believe this happens because in training models, human users demonstrate a preference for sycophantic responses over accurate responses.

A simple way to elicit a sycophantic response from a LLM is to use a prompt like "Are you sure?" or "I dislike this argument." Those kinds of prompts can result in models fundamentally reversing an earlier response, even when the subject is an objective mathematical expression.

Recommendations for avoiding sycophancy in LLMs:

  • Reset sessions often, to reduce the impact on the model of your opinions.
  • Don't express strong opinions or inflexible positions.
  • Don't rely exclusively on the model for fact-finding.
Share this article:

And if you have questions or comments, you can easily send them to me with the Quick Reply form, below, or send me an e-mail.


David Boggs MS    - David
David@DavidHBoggs.com
View David Boggs's profile on LinkedIn

Google Certifications - David H Boggs
View my profile on Quora
Subscribe to my blog

External Article: https://www.nngroup.com/articles/sycophancy-generative-ai-chatbots/


Subhead Even if the user's view is not objectively true
Website
Visit Website
Rating
4/5 based on 1 vote.
Show Individual Votes
Related Listings

Sorry, you don't have permission to post comments. Log in, or register if you haven't yet.

Please login or register.

Members currently reading this thread: