this post was submitted on 29 Nov 2023
128 points (89.5% liked)
Technology
59374 readers
3169 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Seems like the article is trying to combine two issues into one, the lack of representation of woman on OpenAI's Board, and the concerns of some prominent AI researchers (who happen to be women) about OpenAI's ambition and profitability above safety.
On the representation side, this seems like a chicken and egg problem where there won't be any change in diversity if no one wants to make a move because the board isn't already diverse enough.
And on the AI safety side, there won't be any change unless someone sits on the board and pushes for safety proactively, instead of reactively through legislation.
There won't be any change because the board that pushed back just got replaced with people who won't.
And they're getting an opportunity to apply and bring back some balance, but decided not to.
It also elides "AI safety" (Toner's thing) and "AI ethics" (Gebru's thing). They're two different things. Jammed together here because both are women (FFS).
"AI safety" is the sci-fi, paperclip maximisation, fantasies about the potential future of AI.
"AI ethics" is the real actual harms done in the here and now, by embedding existing biases into decision-making, and consuming enormous amounts of resource.
Meredith Whittaker sums up the difference nicely in this interview: