23 Comments
User's avatar
Amy A's avatar

I will never understand how (outside of outstanding hubris) anyone ever thought the crap on the internet would produce AGI. It’s a never ending game of whack-a-mole, where it seems like the leaders are fooling themselves and (trying) to fool everyone else. And damaging the environment.

The Information’s most recent podcast blames the internet as entertainment, which is an interesting addition to the conversation.

Expand full comment
Robert McTague's avatar

I think you could add "deluded by the insatiable allure of limitless profits" to that.

Expand full comment
Eric Cort Platt's avatar

Even if they'd used the finest and highest literature and philosophy in the world it could not produce AGI.

Expand full comment
AKcidentalwriter's avatar

Folks have to realize that A.I is human sourced and anything human source is not infallible! Failure is human..... and so is making it right. However, A.I is selling a utopian vision of perfection. All of these A.I 's are based on human output! So this illusion that has become a delusion needs to told to the masses but that won't happen because it is bathed in innovation which is now a word that has been bastardized.

Expand full comment
Tane Piper's avatar

> He of all people should know better.

A ketamine addiction will do that to the brain.

Expand full comment
LV's avatar

What amazes me is that none of these guys - probably because they’ve never taken a humanities class - has any response addressing the basic problem of inference faced by deep learning that Hume outlined for us 300 years ago. These algorithms try to learn from data by brute statistical force, finding patterns without starting with strong postulates. They have not somehow escaped the basic problem of inference that way, especially as they’ve exhausted all the available training data.

Expand full comment
David Sterry's avatar

I get the same feeling. I'm anxious to explore how a model can be trained on less data but with more behavioral and environmental feedback. From what I see, RLHF uses canned feedback instead of anything closer to like a group of human parents chiding the model. I suppose it comes partly down to cost but also that there's no guarantee that humans brought in to correct the model, won't sabotage it in various ways.

Expand full comment
Roumen Popov's avatar

Musk might be trying to bring down OpenAI by stealing their hype. He already stole $6 billion worth of investment money that otherwise could have gone to OpenAI. I don't know what happened between him and Altman but it's very likely personal now :)

Expand full comment
brutalist's avatar

“should” is irrelevant. Musk doesn’t reliably know anything because he forms his judgements purely on vibes, social positioning, and whether or not he thinks something (e.g. slowing down AI development) is “woke.”

Expand full comment
Tam Hunt's avatar

I don't think Musk flipped on his views so much as his strategy. He still believes, as he stated recently, that AGI has a 10-20% chance of existential risk. But since he apparently believes now that the pause has no chance (and indeed it got promptly forgotten even as a movement almost as soon as the ink dried), he is now placing his efforts into developing truth-seeking AGI as the antidote to woke/dangerous AGI. I'm not endorsing his methods, just trying to shed light on his evolution.

Expand full comment
Notorious P.A.T.'s avatar

"[Musk] of all people should know."

Eh. A person who thinks it's viable to build a 300-mile vacuum tube across California, whose cars can not be exited when they break down, whose space rockets explode after launch, who thinks a rocket to Mars will have room for concert halls, and who built a truck that falls apart after delivery probably does not have great judgment in technological matters.

Expand full comment
Simon Au-Yong's avatar

Who watches the watchmen?

The insurance damage on the city caused by that robot must amount to billions.

Expand full comment
Steve Berman's avatar

Elon does know better. But his personal animus toward people wrong him is severely overdeveloped. He will pound away on big bad LLM AI until he kills whatever OpenAI and Altman are doing. Then he will decide to be reasonable. Maybe.

Expand full comment
DinoNerd's avatar

There's a bit of a fundamental question here. Nothing is 100% reliable. It's become common in software engineering to lean in to that. There's no goal to do as well as possible; just to do well enough that people will buy the product. Customers don't expect reliability; after long years of experience they just reboot the computer when it crashes, and get on with whatever they were trying to do.

From where I sit, the proponents of LLMs figure that they can get their products reliable enough that customers will accept them, and that's all they need to do. Their bet on the one hand is that this is consistent with let's say one egregious mistake per 100 or maybe 1000 queries - maybe more. And on the other, that they can bandaid their systems well enough that people who aren't trying to demonstrate breakage can get that level of reliability routinely, once they've trained themselves to phrase their prompts appropriately.

At that point, they are done. Their investment pays off; maybe someone gives selected lead researchers the Nobel prize.

Meanwhile, their product gets used in area where morally speaking, 99% or even 99.9% reliability is not enough. That's unfortunate, but Not Their Problem (TM). And in any case future research (read "even more bandaids") can be applied to hopefully raise that reliability still farther.

I'm not a fan of this approach, but I can see how it would seem utterly reasonable to many in software engineering, and your criticisms ridiculous and petty. Why "make the perfect the enemy of the good", when you can ship new products faster if you use a chatbot to produce them? Surely QA will catch any mistakes that are really important (sic). And as we all know, humans also make errors. Why single out chatbot errors as somehow worse than the errors that would have been made by the humans they replace?

Expand full comment
Gerald's avatar

Mockery from a bunch of delusional tech bros is something to be proud of. Congratulations.

Expand full comment
Bob B's avatar

Having read your various posts and your book, all I see you do is complain about the state of AI without having any actual suggestions (other than stop working on NNs). As far as I can tell, LLMs do a great job of replicating human behavior complete with hallucinations (see Fox News) and represent a huge step forward. Maybe AGI is not some Spock-like entity but is more like Dr McCoy.

Expand full comment
Joy in HK fiFP's avatar

And we need this, why?

Expand full comment
Shon Pan's avatar

Last I heard Musk still said he favors Pause but has changed teams because the world is refusing to Pause.

#PauseAI here of course and working to at least get oversight

Expand full comment
User's avatar
Comment removed
Jun 1, 2024Edited
Comment removed
Expand full comment
Simon Au-Yong's avatar

Andy. We’ve had language models in development for over four decades. How much more time is needed? 😊

Expand full comment
Amy A's avatar

It’s not like Musk and Altman ever make things about themselves, eh 🤫

Expand full comment
AdamHD's avatar

The meme is loud and not really serious.

It isn't like Gary is responding to a strong, well-reasoned couterargument. He doesn't have a whole lot to work with.

Having said that, I agree with you, science takes time (a sentiment that I'm sure Gary agrees with). But, there is an awful lot of hype to the contrary, with Hinton making claims like a probability of about half that we will be dealing with an AI attempt to take control within the next 5-20 years.

Expand full comment
User's avatar
Comment removed
Jun 3, 2024Edited
Comment removed
Expand full comment
AdamHD's avatar

I think you may have misunderstood the main points of my comment (or I have misunderstood your points and need you to clarify them). I wasn't arguing that deep learning is hitting a wall. I know that Gary is, but that wasn't what I was saying.

In his blog post, Gary was responding to a meme that exemplifies the glib dismissal and hype often dominating these debates. My first point was that this meme, and Musk's amplification of it, was 'loud and not really serious,' highlighting the double standard in your original comment.

The prevalence of exaggerated claims and overheated rhetoric in the AI community is a problem. My second point was that even leading deep learning scientists, like Geoffrey Hinton, have made hyperbolic claims about the potential threats of AI. While I agree that current systems do not pose some kind of immediate existential threat, my concern is that such claims from prominent figures in the field contribute to a climate of hype and unrealistic expectations. Science takes time.

This hype makes it difficult to have grounded, evidence-based discussions about the real-world impacts and challenges of AI with the general public. The general public need to be more aware and educated on AI, and these claims and memes do not help.

Expand full comment
User's avatar
Comment removed
Jun 1, 2024
Comment removed
Expand full comment
Gary Marcus's avatar

QED “high-level reasoning is not one of them .. we need a lot mreo [tools]” sounds like a restatement of my 2018 article and characterization of the wall

Expand full comment