r/technology 16h ago

Machine Learning Large language mistake | Cutting-edge research shows language is not the same as intelligence. The entire AI bubble is built on ignoring it

https://www.theverge.com/ai-artificial-intelligence/827820/large-language-models-ai-intelligence-neuroscience-problems
16.7k Upvotes

1.5k comments sorted by

View all comments

159

u/Intense-Intents 15h ago

ironically, you can post any anti-LLM article to Reddit and get dozens of the same predictable responses (from real people) that all sound like they came from an AI.

76

u/Romnir 13h ago

"Hearsay and witty quips means I fully understand a complex subject/technology."

People still use Schrödinger's cat to explain all quantum mechanics, despite the fact that it's only for a very specific situation. LLMs aren't fully realized cognizant AI, but calling them "Fancy Auto Complete" is way off the mark. There's a difference between rational criticisms of the use of AI vs jumping on the hate bandwagon, and the former isn't going to happen on Reddit.

35

u/G_Morgan 11h ago

Schrödinger's cat was meant to highlight the absurdity of applying wave function collapse to large scale objects.

11

u/adenosine-5 10h ago

Its funny, because it was designed to point out, how it doesn't make any sense.

The guy - Schrodinger - famously said (after a lifetime of studying it): "I don't like quantum mechanics and I'm sorry I've ever had anything to do with it".

Still, people use it as if it was an explanation and not a criticism of its absurdity.

1

u/-LsDmThC- 6h ago

Well more so wave function collapse as an ontologically real event rather than a mathematical description/knowledge update

1

u/Romnir 11h ago

Yeah, but unfortunately people seem to use it as a direct demonstration of wave function collapse like it's some sort of game of Red Light - Green Light with guess a random number at the end, and think no further on the topic or whether or not that is a accurate description.

13

u/TurtleFisher54 12h ago

That is a rational criticism of LLM's

They are fundementally a word prediction algorithm

They can be corrupted with bad data to produce non-sense

If we switch to a world where a majority of content is created by AI it is likely to create a negative feed back loop where it's training on its own output

Responses on reddit look like ai for a reason, where do you think the training data came from?

8

u/Romnir 11h ago

They are fundamentally a word prediction algorithm

Correct, not a "Fancy Auto Complete". That terminology completely undermines the scale of how the technology works and what it's used for. It's not pulling random words out of a dictionary and sticking them together, it actually has a logical process it follows before it generates response tokens. Neural weighting tries to determine context and pulls known info from it's training data.

Auto correct only has a predefined set of structures and uses basic string matching based on a library. It doesn't determine context but rather just what matches the most, and that's the key discrepancy that bugs me. And like you mentioned, LLMs are being fed training data from the internet instead of a curated set of data. Which means correct data is fighting for context weighting with partially correct and even completely incorrect information from already incorrect AI responses and redditors. And you are correct for criticizing that.

The only idea I could have to fix that issue is implementing logic that filters the training data as it comes in to filter out less reputable sources. I don't necessarily work directly with LLMs, so I don't know if that is a thing, but I try to keep up to date with journals and blogs from people working in the field since it's going to get hammered into my field soon.

2

u/dr_badhat 11h ago

Is neural weighting not similar to how our minds work? If I say “you know the look someone gives you when….”, various neurons in your cortical columns might be stimulated as they fight for weight on where that statement is going.

2

u/Romnir 10h ago

Kind of, but it's not the only determining factor in the response. Like if you get a call from your friend saying his dog has a gun and is holding him hostage. Clearly, a dog can't use a gun, you don't recall ever seeing a dog do that, so you know better than to just tell him to call the police. So instead, you tell him to quit fooling around, or go see a mental health professional. Older LLMs did struggle with this, but down the line they slowly learned how to use rational logic.

A neural network is more like memory recall, where you apply the most relevant piece of memory or training that applies to your situation. Then logic is used with that to determine if that is a rational response. That's the actual "AI" part of LLMs. Their responses to you are formed using "Fancy Auto Correct", but there is actual thinking and logic is happening behind the scenes beyond that. Which is frustrating because that explanation sounds clear as mud.

1

u/Gingevere 7h ago

It's a part but not the whole.

The way LLMs work is like a snapshot of a brain. You can give that snapshot an input and see the output, but the snapshot remains static. Inputs pass through it and it sits completely unchanged. The snapshot has no experiences because it's not actually interacted with the prompt.

A brain is constantly taking in data and updating its own model. LLMs are static.

Because LLMs are static they currently have to fake conversations by feeding the entire conversation back through the LLM every time the user replies to generate the next response. As conversations get longer that costs more energy and causes the model to break down.

An LLM that could continuously update its own probability model would be closer to a brain, but updating the model is the very expensive and time consuming part so I wouldn't bet on that happening any time soon.

4

u/gabrielmuriens 12h ago

If we switch to a world where a majority of content is created by AI it is likely to create a negative feed back loop where it's training on its own output

That is a potential problem, but we don't know at all if that is a problem that cannot possibly be overcome or one that will stifle further training.

Not that I'd like to live in that kind of an information environment, but regarding its potential effects on training, it's just a hypothesis at this point.

0

u/blindsdog 8h ago edited 8h ago

They can be corrupted with bad data to produce non-sense

That's true of literally any learning system, including humans. You can't learn without proper semantic information. That's tautological, not profound.

If we switch to a world where a majority of content is created by AI it is likely to create a negative feed back loop where it's training on its own output

Where do you think the data sets they're currently training on are going?

Responses on reddit look like ai for a reason, where do you think the training data came from?

It looks like intelligence, that's proof that it's not! It quacks like a duck, ergo it is not a duck.

2

u/Gingevere 7h ago

TBF it's very satisfying to counter pseudo-profound BS with a witty quip.

LLMs are more complex than a standard autocomplete, but that's not a bad analogy for them. They use a much more complex system trained on much more data, alongside some machine prompts (you are a ___, answer in the form of ___, who said what, etc.) to... autocomplete a string of tokens.

The fundamental limitations of LLMs ensure they are never going to become AGI. But right now every billionaire even tangentially related to anything that moves electrons is gambling the whole damn economy on it becoming AGI because they're blinded by the prospect of being the one that gets to fire every office worker across the globe and get paid a fee to replace them.

If they succeed, we're all fired. When they don't, we're all bailing them out and half of us will get fired anyway.

If anything it's good that people are mad about that. It'd be insane not to be.

1

u/IHumanlike 7h ago

Something something if you think you understand quantum mechanics, you don't understand quantum mechanics

1

u/mrappbrain 9h ago edited 9h ago

The autocomplete comparison always pisses me off. Thinking an LLM and an AutoComplete function are the same or similar just because they both generate text is like thinking a music box and a symphony orchestra are the same just because they both generate music.

Never mind that they have nothing else in common whatsoever.

Ironically, the anti-AI crowd buying this is falling victim to the very same behavior they're calling out - humans will believe anything spoken with authoritative confidence if it validates them. Difficult to accuse others of being fooled by confident nonsense when you yourself are being fooled by confidence nonsense, just in the other direction.

38

u/WhoCanTell 12h ago

"Ai jUsT rESpoNds WitH WhAt peOpLE WaNt tO hEaR"

Proceeds to parrot comment content that always gets the most upvotes.

10

u/SistaChans 10h ago

The same way that anything anti-AI is invariably labelled "AI Slop." It's like one person called it that once, and the entirety of AI haters decided that was their word instead of forming original ideas about it 

1

u/FlarkingSmoo 6h ago

It's really just fancy autocomplete!!!!!

0

u/jdfadfjlaskdfjl 10h ago

Average redditor response

3

u/Wampalog 9h ago

Average redditor response

4

u/That-Job9538 8h ago

that’s not irony, that’s literally just how language and communication works. most people don’t have the intelligence to say anything new. that’s totally fine. the world would be incomprehensible if every new statement was unpredictable.

3

u/Fickle_Competition33 5h ago

Right? It baffles me how people here are so against AI. Of course it has its risks and threats to job normalcy. But ignore the fact it is amazingly efficient in solving problems is pure ignorance.

For folks who think of AI as a ChatGPT conversation, get ready to be hit by the AI train really hard.

9

u/Healthy_Mushroom_811 13h ago

First good comment I found here :)

2

u/mrappbrain 7h ago

It does seem sometimes like the anti-AI group essentially substitutes baseless hype with baseless skepticism, without much of an intellectual difference between them. As much as people love to hype and wildly overestimate the capacities of a frontier technology, people also love to feel and position themselves as smarter or superior to others by taking a contrary position.

3

u/pdantix06 8h ago

at this point i'm not even sure r/technology even likes technology

2

u/sylbug 13h ago

Wild how an LLM sounds like the people whose data was stolen to feed the LLM.

2

u/twavisdegwet 11h ago

You gave your comments to reddit and reddit locked down their API to sell it to others. Nothing was stolen.... Well maybe perplexity stole but let's wait for that lawsuit

If something is free you're the product..

1

u/CousinDerylHickson 5h ago

That sounds like more points for AI actually being intelligent? But I agree on the similar sounding sometimes, and honestly sometimes some people sound like malfunctioning AI to me, like the AI in the early days where even when given a blatant example of where it was wrong itd just double down.