this post was submitted on 16 Oct 2023
23 points (61.9% liked)

Technology

34862 readers
101 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 1 year ago (17 children)

They are saying the internal vector space that LLMs use is too complicated and too unrelated to the output to be understandable to humans.

Yes, that's exactly what I'm saying.

That doesn't mean they're having thoughts in there

I mean. Not in the way we do, and not with any agency, but I hadn't argued either way on thoughts because I don't know the answer to that.

we know exactly what they're doing inside that vector space -- performing very difficult math that seems totally meaningless to us.

Huh? We know what they are doing but we don't? Yes, we know the math, people wrote it. I coded my first neural network 35 years ago. I understand the math. We don't understand how the math is able to do what LLMs do. If that's what you're saying then we agree on this.

The vectors do not represent concepts. The vectors are math. When the vectors are sent through language decomposition they become words, but they were never concepts at any point.

"The neurons are cells. When neurotransmitters are sent through the synapses, they become words, but they were never concepts at any point."

What do you mean by "they were never concepts"? Concepts of things are abstract. Nothing physical can "be" an abstract concept. If you think about a chair, there isn't suddenly a physical chair in your head. There's some sort of abstract representation. That's what word vectors are. Different from how it works in a human brain, but performing a similar function.

A word vector is an attempt to mathematically represent the meaning of a word.

From this page. Or better still, this article explaining how they are used to represent concepts. Like this is the whole reason vector embeddings were invented.

[–] [email protected] 0 points 1 year ago* (last edited 1 year ago) (16 children)

We do understand how the math results in LLMs. Reread what I said. The neural network vectors and weights are too complicated to follow for an individual, and do not relate on a 1:1 mapping with the words or sentences the LLM was trained on or will output, so individuals cannot deduce the output of an LLM easily by studying its trained state. But we know exactly what they’re doing conceptually, and individually, and in aggregate. Read your own sources from your previous post, that’s what they’re telling you.

Concepts are indeed abstract but LLMs have no concepts in them, simply vectors. The vectors do not represent concepts in anything close to the same way that your thoughts do. They are not 1:1 with objects, they are not a “thought,” and anyway there is nothing to “think” them. They are literally only word weights, transformed to text at the end of the generation process.

Your concept of a chair is an abstract thought representation of a chair. An LLM has vectors that combine or decompose in some way to turn into the word “chair,” but are not a concept of a chair or an abstract representation of a chair. It is simply vectors and weights, unrelated to anything that actually exists.

That is obviously totally different in kind to human thought and abstract concepts. It is just not that, and not even remotely similar.

You say you are familiar with neural networks and AI but these are really basic underpinnings of those concepts that you are misunderstanding. Maybe you need to do more research here before asserting your experience?

Edit: And in relation to your links -- the vectors do not represent single words, but tokens, which indeed might be a whole word, but could just as well be part of a word or an entire phrase. Tokens do not represent the meaning of a word/partial word/phrase, just the statistical use of that word given the data the word was found in. Equating these vectors with human thoughts oversimplifies the complexities inherent in human cognition and misunderstands the limitations of LLMs.

[–] [email protected] 0 points 1 year ago (12 children)

Your concept of a chair is an abstract thought representation of a chair. An LLM has vectors that combine or decompose in some way to turn into the word “chair,” but are not a concept of a chair or an abstract representation of a chair. It is simply vectors and weights, unrelated to anything that actually exists.

Just so incredibly wrong. Fortunately, I'll have save myself time arguing with such a misunderstanding. GPT-4 is here to help:

This reads like a misunderstanding of how LLMs (like GPT) work. Saying an LLM's understanding is "simply vectors and weights" is like saying our brain's understanding is just "neurons and synapses". Both systems are trying to capture patterns in data. The LLM does have a representation of a chair, but it's in its own encoded form, much like our neurons have encoded representations of concepts. Oversimplifying and saying it's unrelated to anything that actually exists misses the point of how pattern recognition and information encoding works in both machines and humans.

[–] [email protected] 0 points 1 year ago (1 children)

Are you kidding me? I sourced GPT4 itself disagreeing with you that it is intelligent and you told me it's lying. And here you are, using it to try to reinforce your point? Are you for real or is this some kind of complicated game?

[–] [email protected] 1 points 1 year ago (1 children)
[–] [email protected] -1 points 1 year ago* (last edited 1 year ago) (2 children)

Here, let's ask GPT4 itself since you've decided it's suddenly an okay source:

Your statement is correct in asserting that the vector representation in a language model is not an abstract representation. It's purely a mathematical construct. However, saying it's "unrelated to anything that actually exists" might be an overstatement. These vectors do capture statistical patterns in human language, which are reflections of human thought and culture. They're just not capable of the deep, nuanced understanding that comes from human experience.

I accept it's an overstatement. But it is neither "incredibly wrong," nor is it thought. (Or intelligence.)

[–] [email protected] 1 points 1 year ago (1 children)

So you admit that you were wrong?

[–] [email protected] -1 points 1 year ago

I was in this case -- but the overall point I made is still correct. If winning this minor battle is what you were seeking, congratulations. You are no closer to understanding the truth of this or what we were actually talking about. Not that that was either your point or within your capabilities.

[–] [email protected] 1 points 1 year ago (1 children)

I'd just like to step in here and mention that asking an LLM is probably not a good proof (and this is directed at both of you). Its understanding of AI is from before it was trained, so it is wildly out of date at this point given how much has happened in the space since.

[–] [email protected] 1 points 1 year ago (1 children)

GPT4 has knowledge of its own training since it was trained in 2022.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

Care to provide some proof of that? They did update their system prompt to include a few things like it is now GPT4 (it used to always say GPT3). Other than that, I don't think it knows anything. But in general, I was more talking about developments in AI since it was trained which it certainly does not know.

Edit: hmm I just reviewed our discussion and I note you only provided one link which was to the psychological definition of intelligence. You otherwise are providing no sources to back up your claims while my responses are full of them. Please start backing up your assertions, or provide some evidence you are an expert in the field.

[–] [email protected] 1 points 1 year ago (1 children)
[–] [email protected] 1 points 1 year ago (1 children)

I'm aware of that date.

The OpenAI GPT-4 video literally states that GPT-4 finished training in August 2022.

Either way, to clarify / reiterate, you're refuting a different point than I've made. I said:

Its understanding of AI is from before it was trained, so it is wildly out of date at this point given how much has happened in the space since.

I'm not talking about whether it knows about its own training (I doubt that it does). I'm talking about it knowing about what's happened in the broader AI landscape since.

[–] [email protected] 1 points 1 year ago (1 children)

I mean, your argument is still basically that it's thinking inside there; everything I've said is germane to that point, including what GPT4 itself has said.

[–] [email protected] 1 points 1 year ago

I mean, your argument is still basically that it’s thinking inside there; everything I’ve said is germane to that point, including what GPT4 itself has said.

My argument?

That doesn’t mean they’re having thoughts in there I mean. Not in the way we do, and not with any agency, but I hadn’t argued either way on thoughts because I don’t know the answer to that.

Are you assuming I’m saying that LLMs are sentient, conscious, have thoughts or similar? I’m not. Jury’s out on the thought thing, but I certainly don’t believe the other two things.

I'm not saying it's thinking or has thoughts. I'm saying I don't know the answer to that, but if it is it definitely isn't anything like human thoughts.

load more comments (10 replies)
load more comments (13 replies)
load more comments (13 replies)