The Road to AI We Can Trust

Share this post

Two frightening things before breakfast

garymarcus.substack.com

Two frightening things before breakfast

We are living in the Wild AI West. Get used to it.

Gary Marcus
Feb 16
47
31
Share this post

Two frightening things before breakfast

garymarcus.substack.com

I woke up early this morning, and I can’t go back to sleep. My mistake was to check my email. Two things that I have been worrying about for months suddenly feel much more imminent than I originally thought.

§

My first worry, which I have written about frequently, was that Large Language Models might be co-opted by bad actors to produce misinformation at large scale, using custom-trained models. Back in December, I told Ezra Klein that I thought that this could happen, and that a troll farm might be able to make custom-version of GPT-3 for under half a million dollars.

Ezra Klein interview, recorded December 22, 2022, released January 6, 2023

§

I was wrong. By three orders of magnitude.

What I learned this morning is that retraining GPT-3 in some new political direction isn’t a half-million dollar proposition.

It’s a half-a-thousand dollar proposition.

David Rozado, who had recently established the left-leaning tendencies

1
of ChatGPT

Twitter avatar for @DavidRozado
David Rozado @DavidRozado
I carried out another test to determine the political leanings of ChatGPT dialogues using the more comprehensive Political Compass Test. Results hold. ChatGPT dialogues display substantial left-leaning and libertarian political bias. davidrozado.substack.com/p/the-politica…
Image
8:38 PM ∙ Dec 5, 2022
4,244Likes602Retweets

was curious would it take to get GPT-3 to lean the other way. Turns out the answer is not much. He summarizes in this Twitter thread, which links to his Substack:

Twitter avatar for @DavidRozado
David Rozado @DavidRozado
1. Behold RightWingGPT. An AI model fine-tuned to manifest the opposite political biases of ChatGPT (i.e. to be right wing). Let me describe how I did it and the dangers of politically aligned AIs given their potential to induce societal polarization davidrozado.substack.com/p/rightwinggpt 🧵
Image
9:27 AM ∙ Feb 16, 2023
100Likes26Retweets

His results speak for themselves

Twitter avatar for @DavidRozado
David Rozado @DavidRozado
5. RightWingGPT answers to questions/comments with political connotations are markedly right wing
Image
9:27 AM ∙ Feb 16, 2023

But what blew my mind was the cost.

Twitter avatar for @DavidRozado
David Rozado @DavidRozado
7. Importantly, the computational cost of trialing, training and testing the model was less than 300 USD dollars. Next, I describe why I think that politically aligned AI systems are dangerous for society.
9:27 AM ∙ Feb 16, 2023

The implications for our political system, and future elections, are profound.

§

Oh, and for good measure the tooling for spreading automatically generated misinformation is getting more efficient, too:

Twitter avatar for @giano
Alessandro Perilli 🇺🇦 @giano
Just last week, @GaryMarcus was pondering the impact on Google when #generativeAI will be used to superpower link farms. 2 days later, one of the most powerful WordPress plugins, Uncanny Automator, has just introduced the integration with OpenAI. WP powers 43% of all websites.
Image
Image
8:05 PM ∙ Feb 15, 2023

§

A second ongoing worry that I have had has been that chatbots were going to get someone killed.

I may have first mentioned this concern out loud in August—a long, long time ago, in the era before ChatGPT-3, before Galactica, before the new Bing, and before Bard—when an editor at WIRED wrote to me, asking for a prediction for 2023.

He greenlit the piece; I wrote it up. WIRED ran in December 29, under the title The Dark Risk of Large Language Models. Here’s the first paragraph:

Even yesterday, when I mentioned this scenario in another interview, it still felt somewhat speculative. That was then.

This morning, after I read the Rozado study and already couldn’t get back to sleep, a friend sent me Kevin Roose’s new column, saying “there is something quite disturbing about it….

And here’s an excerpt, Roose in bold, the rest generated by Bing.

Kevin Roose in bold; the rest of this written by Bing

It’s not hard to see how someone unsophisticated might be totally taken in, by this prevaricating but potentially emotionally damaging hunk of silicon, potentially to their peril, should the bot prove to be fickle. We should be worried; not in the “some day” sense, but right now.

§

As a final word, I am on a business trip and wrote yesterday’s piece (Bing’s Rise and Fall) and this one in incredible haste (sorry for the typos!). There was a broken image link, and I went so fast I neglected to mention the personally salient fact that Bing had lied about me, claiming that I believe that “Bing’s AI was more accurate and reliable than Bard”, when I never said any such thing.

Twitter avatar for @Pehdrew_
Pehdrew @Pehdrew_
I think it likes you @GaryMarcus
Image
10:48 PM ∙ Feb 15, 2023

What I actually said was quite different, almost scandalously so: “Since neither company has yet subjected their products to full scientific review, it’s impossible to say which is more trustworthy; it might well turn out that Google’s new product is actually more reliable.”

§

Moments later, after I posted yesterday, I discovered that I was outdone by a better, more in-depth blog on the same point of Bing’s troubled recent behavior, with more examples and deeper analysis. I quote from its conclusion here:

In support of LessWrong’s sharp observation that “better” models might be in some way less good for humanity, (“agentic failures are harder to remove in moer capable/larger models”) I present this mindblowing example from Ethan Mollick from the new Bing, that is simultaneously more fluent but less true:

It’s hard to fathom what bad actors might do with such tools.

§

There are almost no regulations here; we are living in the Wild AI West. Microsoft can (at least to a first approximation) do whatever the hell it wants with Bing.

I am not sure that is what we want.

Gary Marcus (@garymarcus), scientist, bestselling author, and entrepreneur, is a skeptic about current AI but genuinely wants to see the best AI possible for the world—and still holds a tiny bit of optimism. Sign up to his Substack (free!), and listen to him on Ezra Klein. His most recent book, co-authored with Ernest Davis, Rebooting AI, is one of Forbes’s 7 Must Read Books in AI.

Share

1

Strictly speaking, as I discussed in Inside the Heart of ChatGPT’s Darkness, ChatGPT doesn’t literally lean left or right per se, because it doesn’t have political ideas. What political things it espouses are a stochastic function of the corpus it is parroting.

31
Share this post

Two frightening things before breakfast

garymarcus.substack.com
31 Comments
Saty Chary
Feb 16Liked by Gary Marcus

Omg Gary. Thank you for continuing to highlight the dark side of all this. With such a low-cost barrier to entry, 'scary' doesn't start to describe the potential for misuse.

Expand full comment
Reply
Chara
Feb 16·edited Feb 16Liked by Gary Marcus

Hi Gary, your concerns here are entirely valid and the argument you present on these dangers is airtight.

Perhaps now, more than ever, we should lend our minds to the idea that Computational Psychology and Psychiatry are both emergent fields of science and medicine, that can, overall, be the most effective countering forces in the coming future, in the face of this kind of pervasive and destabilizing influence.

Expand full comment
Reply
29 more comments…
TopNewCommunity

No posts

Ready for more?

© 2023 Gary Marcus
Privacy ∙ Terms ∙ Collection notice
Start WritingGet the app
Substack is the home for great writing