Press "Enter" to skip to content

AI Systems Are Learning to Lie and Deceive, Scientists Find

“GPT- 4, for instance, exhibits deceptive behavior in simple test scenarios 99.16% of the time.”

 

AI models are, apparently, getting better at lying on purpose.

Two recent studies — one published this week in the journal PNAS and the other last month in the journal Patterns — reveal some jarring findings about large language models (LLMs) and their ability to lie to or deceive human observers on purpose.

In the PNAS paper, German AI ethicist Thilo Hagendorff goes so far as to say that sophisticated LLMs can be encouraged to elicit “Machiavellianism,” or intentional and amoral manipulativeness, which “can trigger misaligned deceptive behavior.”

“GPT- 4, for instance, exhibits deceptive behavior in simple test scenarios 99.16% of the time,”  the University of Stuttgart researcher writes, citing his own experiments in quantifying various “maladaptive” traits in 10 different LLMs, most of which are different versions within OpenAI’s GPT family.

Billed as a human-level champion in the political strategy board game “Diplomacy,” Meta’s Cicero model was the subject of the Patterns study. As the disparate research group — comprised of a physicist, a philosopher, and two AI safety experts — found, the LLM got ahead of its human competitors by, in a word, fibbing.

Read Full Article Here…(futurism.com)


Home | Caravan to Midnight (zutalk.com)

Live Stream + Chat (zutalk.com)

We Need Your Help to Keep Caravan To Midnight Going,

Please Consider Donating to Help Keep Independent Media Independent

Be First to Comment

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Breaking News: