this post was submitted on 22 Feb 2024
488 points (96.2% liked)

Technology

59440 readers
5037 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Google apologizes for ‘missing the mark’ after Gemini generated racially diverse Nazis::Google says it’s aware of historically inaccurate results for its Gemini AI image generator, following criticism that it depicted historically white groups as people of color.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 130 points 9 months ago (88 children)

I don't know how you'd solve the problem of making a generative AI accurately create a slate of images that both a) inclusively produces people with diverse characteristics and b) understands the context of what characteristics could feasibly be generated.

But that's because the AI doesn't know how to solve the problem.

Because the AI doesn't know anything.

Real intelligence simply doesn't work like this, and every time you point it out someone shouts "but it'll get better". It still won't understand anything unless you teach it exactly what the solution to a prompt is. It won't, for example, interpolate its knowledge of what US senators look like with the knowledge that all of them were white men for a long period of American history.

[–] [email protected] 4 points 9 months ago (1 children)

Real intelligence simply doesn't work like this

There's a certain point where this just feels like the Chinese room. And, yeah, it's hard to argue that a room can speak Chinese, or that the weird prediction rules that an LLM is built on can constitute intelligence, but that doesn't mean it can't be. Essentially boiled down, every brain we know of is just following weird rules that happen to produce intelligent results.

Obviously we're nowhere near that with models like this now, and it isn't something we have the ability to work directly toward with these tools, but I would still contend that intelligence is emergent, and arguing whether something "knows" the answer to a question is infinitely less valuable than asking whether it can produce the right answer when asked.

[–] [email protected] 4 points 8 months ago (2 children)

I really don't think that LLMs can be constituted as intelligent any more than a book can be intelligent. LLMs are basically search engines at the word level of granularity, it has no world model or world simulation, it's just using a shit ton of relations to pick highly relevant words based on the probability of the text they were trained on. That doesn't mean that LLMs can't produce intelligent results. A book contains intelligent language because it was written by a human who transcribed their intelligence into an encoded artifact. LLMs produce intelligent results because it was trained on a ton of text that has intelligence encoded into it because they were written by intelligent humans. If you break down a book to its sentences, those sentences will have intelligent content, and if you start to measure the relationship between the order of words in that book you can produce new sentences that still have intelligent content. That doesn't make the book intelligent.

[–] [email protected] 3 points 8 months ago (1 children)

But you don't really "know" anything either. You just have a network of relations stored in the fatty juice inside your skull that gets excited just the right way when I ask it a question, and it wasn't set up that way by any "intelligence", the links were just randomly assembled based on weighted reactions to the training data (i.e. all the stimuli you've received over your life).

Thinking about how a thing works is, imo, the wrong way to think about if something is "intelligent" or "knows stuff". The mechanism is neat to learn about, but it's not what ultimately decides if you know something. It's much more useful to think about whether it can produce answers, especially given novel inquiries, which is where an LLM distinguishes itself from a book or even a typical search engine.

And again, I'm not trying to argue that an LLM is intelligent, just that whether it is or not won't be decided by talking about the mechanism of its "thinking"

[–] [email protected] 6 points 8 months ago (1 children)

We can’t determine whether something is intelligent by looking at its mechanism, because we don’t know anything about the mechanism of intelligence.

I agree, and I formalize it like this:

Those who claim LLMs and AGI are distinct categories should present a text processing task, ie text input and text output, that an AGI can do but an LLM cannot.

So far I have not seen any reason not to consider these LLMs to be generally intelligent.

[–] [email protected] 0 points 8 months ago

Literally anything based on opinion or creating new info. An AI cannot produce a new argument. A human can.

It took me 2 seconds to think of something LLMs can't do that AGI could.

[–] [email protected] 1 points 8 months ago (2 children)

What do you mean it has no world model? Of course it has a world model, composed of the relationships between words in language that describes that world.

If I ask it what happens when I drop a glass onto concrete, it tells me. That’s evidence of a world model.

[–] [email protected] 7 points 8 months ago* (last edited 8 months ago)

A simulation of the world that it runs to do reasoning. It doesn't simulate anything, it just takes a list of words and then produces the next word in that list. When you're trying to solve a problem, do you just think, well I saw these words so this word comes next? No, you imagine the problem and simulate it in both physical and abstract terms to come up with an answer.

[–] [email protected] 3 points 8 months ago* (last edited 8 months ago) (1 children)

I can see the argument that it has a sort of world model, but one that is purely word relationships is a very shallow sort of model. When I am asked what happens when a glass is dropped onto concrete, I don't just think about what I've heard about those words and come up with a correlation, I can also think about my experiences with those materials and with falling things and reach a conclusion about how they will interact. That's the kind of world model it's missing. Material properties and interactions are well enough written about that it ~~simulates ~~ emulates doing this, but if you add a few details it can really throw it off. I asked Bing Copilot "What happens if you drop a glass of water on concrete?" and it went into excruciating detail about how the water will splash, mentions how it can absorb into it or affect uncured concrete, and now completely fails to notice that the glass itself will strike the concrete, instead describing the chemistry of how using "glass (such as from the glass of water)" as aggregate could affect the curing process. Having a purely statistical/linguistic world model leaves some pretty big holes in its "reasoning" process.

[–] [email protected] 2 points 8 months ago (1 children)

I believe you meant to say emulates instead of simulates

[–] [email protected] 1 points 8 months ago

Thanks, that is a better word there.

load more comments (86 replies)