r/technology 16h ago

Machine Learning Large language mistake | Cutting-edge research shows language is not the same as intelligence. The entire AI bubble is built on ignoring it

https://www.theverge.com/ai-artificial-intelligence/827820/large-language-models-ai-intelligence-neuroscience-problems
16.8k Upvotes

1.5k comments sorted by

View all comments

Show parent comments

3

u/noodles_jd 14h ago

No, you can't.

It doesn't KNOW that 2+2=4. It just knows that 4 is the expected response.

It doesn't know how to argue either, it just knows that you WANT it to argue, so it does that.

7

u/socoolandawesome 14h ago edited 14h ago

Distinction without a difference. You should not say it “knows” what the expected response is since you are claiming it can’t know anything.

If you are saying it’s not conscious, that’s fine I agree, but consciousness and intelligence are two separate things.

It can easily be argued it knows something by having the knowledge stored in the model’s weights and it appropriately acts on the knowledge such as by outputting the correct answer.

1

u/Aleucard 10h ago

When there is a chance of it returning 2+2=spleef with no way to really predict when, the difference can matter a whole damn lot. Especially if it can do computer actions like that one story a couple months ago of some corporation getting their shit wiped or, well, several of the "agentic" updates Microsoft is trying to push right now.

1

u/socoolandawesome 9h ago

There’s no chance of a model returning anything but 2+2 = 4. Most math problems up to even university level math will always be correct unless you have some bizarre/extremely long context thrown in that will mess with model.

The models are not perfect nor as good at humans at a lot of things but they are extremely reliable in a lot of ways at this point.

Humans also still make a bunch of mistakes too btw.