this post was submitted on 05 Feb 2024
195 points (89.5% liked)

Technology

59207 readers
3158 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Did nobody really question the usability of language models in designing war strategies?

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 3 points 9 months ago (1 children)

You see the irony right? I genuinely can’t fathom your intent when telling this story, but it is an absolutely stellar example.

Yes, I did mean for it to be an example.

And yes, I do think that correctly framing a question is crucial whether you're dealing with a person or an LLM. But I was elaborating on whether a person's process of answering a question is fundamentally similar to an LLM's process. And this is one way that it's noticeably different. A person will size up who is asking, what they're asking, and how they're asking it... and consider whether they should actually answer the exact question that was asked or suggest a better question instead.

You can certainly work around it, as the asker, but it does require deliberate disambiguation. I think programmers are used to doing that, so it may feel like not that big of a deal, but if you start paying attention to how often people are tossing around half-formed questions or statements and just expecting the recipient to fill in the gaps... It's basically 100% of the time.

We're fundamentally social creatures first, and intelligent creatures second. (Or third, or not at all, depending.) We think better as groups. If you give 10 individuals a set of difficult questions, they'll bomb almost all of them. If you give the questions to a group of 10, they'll get almost all of them right. (There's several You Are Not So Smart episodes on this, but the main one is 111.)

Asking a question to an LLM is just completely different from asking a person. We're not optimized for correctly filling out scantron sheets as individuals, we're optimized for brainstorming ideas and pruning them as a group.

[–] [email protected] 1 points 9 months ago* (last edited 9 months ago)

If you fed that information into one I bet you would get different answers.

That is information that isn't available to it generally.