this post was submitted on 24 Jan 2024
7 points (54.1% liked)
Technology
59174 readers
3103 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
No, but I have learned over the years that when you see multiple papers discovering similar things at odds with the held consensus and see some even independently replicated that there's usually more than just smoke.
The paper was titled "Skill-Mix: a Flexible and Expandable Family of Evaluations for AI models." Quanta, while a Pulizer winner in 2022 for explanatory reporting, is after all a publisher not a research institution. Though I dispute your issues with the headline as it's in line with similar article headlines such as "Bees understand the concept of zero".
You wouldn't be the only person looking at it through that lens. It was more popular a few years ago I think, and hasn't really caught on for LLMs vs other ML approaches and here it strikes me a bit like those with hammers looking for nails - the degree to which there's functional overlaps in network introspection such as the linked Anthropic work suggests to me that the internalized delineations are a bit fuzzier than would cleanly map onto a category theory view - but it's possible that as time goes on that it gets some research wins assuming it can come up with testable predictions that are successful. But it's more of a 'how' than a 'what' question - whether a network understands abstract concepts tangental to language it is trained on and develops world models (an idea that would have been laughed out of the room just three years ago by any serious researchers despite your impression) using something that can be explained through category theory or through another interpretation, the result is arguably the more important finding than the interpretation of the means.
It seems like you may be more committed to arguing the semantics and nuances of the tree in front of you than discussing the forest - that's fine, it's just not that interesting to me in turn.
To hijack your analogy its more akin to me stating a tree is a plant and you saying "So are these" pointing at a forest of plastic Christmas trees.
I'm pretty curious why you imagine you have so many downvotes?
Because laypeople are very committed to a certain perspective of LLMs right now.
You should see the downvotes I got a year or two ago explaining immunology research to antivaxxers.
Have you ever considered you might be the laypeople?
Equating a debate about the origin of understanding to antivaxxers...
You argue like a Trump supporter.