this post was submitted on 25 Aug 2024
328 points (92.5% liked)

Technology

60033 readers
2895 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 33 points 3 months ago (2 children)

AI as a general concept probably will at some point. But LLMs have all but reached the end of the line and they're not nearly smart enough.

[–] [email protected] 8 points 3 months ago (3 children)

LLMs have already reached the end of the line 🤔

I don’t believe that. At least from an implementation perspective we’re extremely early on, and I don’t see why the tech itself can’t be improved either.

Maybe it’s current iteration has hit a wall, but I don’t think anyone can really say what the future holds for it.

[–] [email protected] 24 points 3 months ago* (last edited 3 months ago) (1 children)

LLMs have been around since roughly ~~2016~~ 2017 (comment below corrected me that Attention paper was 2017). While scaling the up has improved their performance/capabilities, there are fundamental limitations on the actual approach. Behind the scenes, LLMs (even multimodal ones like gpt4) are trying to predict what is most expected, while that can be powerful it means they can never innovate or be truth systems.

For years we used things like tf-idf to vectorize words, then embeddings, now transformers (supped up embeddings). Each approach has it limits, LLMs are no different. The results we see now are surprisingly good, but don't overcome the baseline limitations in the underlying model.

[–] [email protected] 6 points 3 months ago

I'm not trained in formal computer science, so I'm unable to evaluate the quality of this paper's argument, but there's a preprint out that claims to prove that current computing architectures will never be able to advance to AGI, and that rather than accelerating, improvements are only going to slow down due to the exponential increase in resources necessary for any incremental advancements (because it's an NP-hard problem). That doesn't prove LLMs are end of the line, but it does suggest that additional improvements are likely to be marginal.

Reclaiming AI as a theoretical tool for cognitive science

[–] [email protected] -5 points 3 months ago* (last edited 3 months ago) (1 children)

we're extremely early on

Oh really! The analysis has been established since the 80's. Its so far from early on that statement is comical

[–] [email protected] 0 points 3 months ago

"at some point" being like 400 years in the future? Sure.

Ok that's probably a little bit of an exaggeration. 250 years.