this post was submitted on 08 Mar 2024
72 points (86.7% liked)
Technology
59148 readers
2721 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I really wish all of these companies racing to replace their existing software features and employees with LLMs understood this. So many applications are dependent on a response being 100% accurate for a very specific request as opposed to being 80% accurate for a wide variety of requests. "Based on training data, here's what a response to your input might look like" is pretty good for conversational language and image generation, but it sucks for anything requiring computation or expertise. Worst of all, it's so confidently wrong about things I might as well be back on Reddit.
They totally understand it. And OpenAI has solved it. For example while researching The Ultimate Answer to Life the Universe and Everything, I asked it to calculate 6 by 9 in base 13 and got the correct answer - 42.
ChatGPT didn't use the LLM to calculate that. It only used the LLM understand an obscure and deliberately confusing chapter of the Hitchhiker's Guide book, to write and execute this python script.