this post was submitted on 16 Feb 2024
265 points (100.0% liked)
Technology
59148 readers
2721 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
If it was a human agent, surely they would still liable?
They're an agent of the company. They're acting on behalf of the company, in accordance to their policy and procedures. It then becomes a training issue if they were providing incorrect information?
Yes, if it was a human agent they would certainly be liable for the mistake, and the law very much already recognises that.
That's my whole point here; the company should be equally liable for the behaviour of an AI agent as they are for the behaviour of a human agent when it gives plausible but wrong information.