this post was submitted on 27 Apr 2024
884 points (95.8% liked)

Technology

59174 readers
3700 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 13 points 6 months ago (3 children)

Tesla's self driving appears to be less safe and causes more accidents than their competitors.

"NHTSA’s Office of Defects Investigation said in documents released Friday that it completed “an extensive body of work” which turned up evidence that “Tesla’s weak driver engagement system was not appropriate for Autopilot’s permissive operating capabilities."

Tesla bad.

[–] [email protected] 5 points 6 months ago (1 children)

Can you link me the data that says Tesla's competitors self-driving is more safe and causes less accidents and WHICH ONES? I would really like to know who else has this level of self-driving while also having less accidents.

[–] [email protected] 1 points 6 months ago (1 children)

The data doesn't exist, no other company has a level of "autonomy" that will let your car plow through shit without you paying attention.

[–] [email protected] 2 points 6 months ago (1 children)

I don't quite understand what they mean by this. It tracks drivers with a camera and the steering wheel sensor and literally turns itself off if you stop paying attention. What more can they do?

[–] [email protected] 2 points 6 months ago* (last edited 6 months ago)

The NHSTA hasn't issued rules for these things either.

the U.S. gov has issued general guidelines for the technology/industry here:

https://www.transportation.gov/av/4

They have an article on it discussing levels of automation here:

https://www.nhtsa.gov/vehicle-safety/automated-vehicles-safety

By all definitions layed out in that article:

BlueCruise, Super Cruise, Mercedes' thing is a lvl3 system ( you must be alert to reengage when the conditions for their operation no longer apply )

Tesla's FSD is a lvl 3 system (the system will warn you when you must reengage for any reason)

Waymo and Cruise are a lvl 4 system (geolocked)

Lvl 5 systems don't exist.

What we don't have is any kind of federal laws:

https://www.ncsl.org/transportation/autonomous-vehicles

Separated into two sections – voluntary guidance and technical assistance to states – the new guidance focuses on SAE international levels of automation 3-5, clarifies that entities do not need to wait to test or deploy their ADS, revises design elements from the safety self-assessment, aligns federal guidance with the latest developments and terminology, and clarifies the role of federal and state governments.

The guidance reinforces the voluntary nature of the guidelines and does not come with a compliance requirement or enforcement mechanism.

(emphasis mine)

The U.S. has operated on a "states are laboratories for laws" principal since its founding. The current situation is in line with that principle.

These are not my opinions, these are all facts.

[–] [email protected] -4 points 6 months ago (1 children)

No one else has the same capability in as wide a geographic range. Waymo, Cruise, Blue Cruise, Mercedes, etc are all geolocked to certain areas or certain stretches of road.

[–] [email protected] 3 points 6 months ago (1 children)

Ok? Nobody else is being as wildly irresponsible, therefore tesla should be... rewarded?

[–] [email protected] 1 points 6 months ago (1 children)

I'm saying larger sample size == larger numbers.

Tesla announced 300 million miles on FSD v12 in just the last month.

https://www.notateslaapp.com/news/2001/tesla-on-fsd-close-to-license-deal-with-major-automaker-announces-miles-driven-on-fsd-v12

Geographically, that's all over the U.S, not just in hyper specific metro areas or stretches of road.

The sample size is orders of magnitude bigger than everyone else, by almost every metric.

If you include the most basic autopilot, Tesla surpassed 1 billion miles in 2018.

These are not opinions, just facts. Take them into account when you decide to interpret the opinion of others.

[–] [email protected] 0 points 6 months ago* (last edited 6 months ago) (1 children)

That's not how rates work tho. Larger sample size doesn't correlate with a higher rate of accidents, which is what any such study implies, not just raw numbers. Your bullshit rationalization is funny. In fact, a larger sample size tends to correspond with lower rates of flaws, as there is less chance that an error/fault makes an outsized impact on the data.

[–] [email protected] 1 points 6 months ago* (last edited 6 months ago) (1 children)

No one's talking about rates. The article itself, all the articles linked in these comments are talking about counts. Numbers of incidents. I'm not justifying anything because I'm not injecting my opinion here. I'm only pointing out that without context, counts don't give you enough information to draw a conclusion, that's just math. You can't even derive a rate without that context!

[–] [email protected] 1 points 6 months ago (1 children)

That's not my point though. We both know that the government agency doing this work is primarily interested in the rates, whether or not reports from the media are talking about the total numbers or not. The only reason they started the process of investigation was because of individual incidents, yes, but they're not looking for a few cases, but a pattern.

(Like this one:https://www.ranzlaw.com/why-are-tesla-car-accident-rates-so-high/)

[–] [email protected] 1 points 6 months ago* (last edited 6 months ago)

Once more, I'm literally not injecting an opinion here or arguing for or against anyone's point. All the articles here talked about counts of individual accidents with zero context about sample size, something that is absolutely crucial to establishing exactly what you're talking about, rates. You can shit all over that, and then pretend you didn't, but Im only pointing out that the math doesn't work unless that context is there.

(I find it funny that the article you just posted is literally an ad for a traffic accident lawyer: here's the study the ad is citing. The ad did some creative interpretation on those numbers, ignoring things like DUI's for example: https://www.lendingtree.com/insurance/brand-incidents-study/#:~:text=Tesla%20drivers%20have%20the%20highest%20accident%20rate%20compared%20with%20all,over%2020.00%20per%201%2C000%20drivers.)