The reason I ask, is I’ve been seeing alot of news and cases of Tesla’s self driving acting up and being a point of contention. But back in 2016-17 my ex’s uncle and aunt got a Model X when they first dropped and they “auto-drove” us like 50 miles without any noticeable issue.

Was i just gambling my life or has the tech somehow gotten worse?

  • Sanctus@lemmy.world
    link
    fedilink
    English
    arrow-up
    37
    arrow-down
    1
    ·
    2 months ago

    The tech never “evolved” in the first place. It was a deformity from the start. LIDAR was developed for this reason. But Tesla uses cheap ass cameras that try to interpret what it is seeing through the visual data. I’m guessing with my layman’s knowledge this is why they veer at semi trucks. Because the technology itself is based on a shitty premise.

    • BertramDitore@lemm.ee
      link
      fedilink
      English
      arrow-up
      24
      ·
      2 months ago

      In general I think you’re right about the tech just being shitty, but a slight correction: LiDAR was not developed for self-driving, it’s just a relevant application of the technology. LiDAR has been around for quite a while, and was initially best known as a remote sensing technology. It is effective at remote sensing because it can penetrate certain solid materials, most importantly foliage. So when an aerial LiDAR dataset is collected for a forested area, since the light can penetrate through most of the foliage, one can essentially ‘delete’ the vegetation from the resulting point cloud, leaving a bare earth model, which is a very close approximation of the landscape’s actual topography if there had been no trees. This can be especially valuable for archaeological research, as foliage is often a significant obstacle for accurately mapping large sites, or even finding them in the first place.

      All of that to say, yeah, self-driving buzz made LiDAR well known as tech, but it wasn’t developed for that purpose.

      • Sanctus@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        2 months ago

        Thanks, I did not know that and it is honestly a way cooler application. Foliage X Ray vision.

    • RaoulDook@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 months ago

      In a non-Tesla car I’ve driven, there was an autopilot cruise control mode that just used cameras. In practice it only works out well if you’re driving long distance on a highway with low traffic. It’s still nice to have (much better than having no autopilot cruise mode) but I don’t trust it around multiple lanes of other cars doing unpredictable shit. Also quits working in the rain when the cameras are obscured.

  • FuglyDuck@lemmy.world
    link
    fedilink
    English
    arrow-up
    35
    ·
    2 months ago

    Tesla’s marketing has consistently lied to you about the state of their “full self driving”.

    It’s not. It never was. This is a perfect example of “fake it ‘till you make it”

  • fine_sandy_bottom@lemmy.federate.cc
    link
    fedilink
    arrow-up
    21
    ·
    2 months ago

    The tech hasn’t regressed, it just hasn’t progressed while the marketing has.

    Look up the automation levels: https://au.pcmag.com/cars-auto/94559/is-your-car-autonomous-the-6-levels-of-self-driving-explained

    My wife’s car is 6 years old, and is level 2. Nothing amazing now, but kinda cool in 2018.

    Since then expectations have increased dramatically, and the problems you’re hearing about are cars expected to have the higher levels of automation but failing to achieve that.

    It seems like this is one of those technical problems that gets exponentially more difficult to solve, the closer we get to solving it. What I mean is, suppose a human averages 100,000km per “incident”. It was easy to make a car do 90,000km per incident, less so to have it do 95,000km per incident, but we’re finding it very very difficult to get that last 5% performance.

    • FuglyDuck@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      2 months ago

      It seems like this is one of those technical problems that gets exponentially more difficult to solve, the closer we get to solving it. What I mean is, suppose a human averages 100,000km per “incident”. It was easy to make a car do 90,000km per incident, less so to have it do 95,000km per incident, but we’re finding it very very difficult to get that last 5% performance.

      to this, most cars with relatively high degrees of automation bounce out of it when there’s something that it feels it can’t solve adequately. Most of the time that works, but it would be like if a human were somehow able to only drive when everything was routine. (90% of our driving is routine. it’s that last ten percent that isn’t- and that’s when we crash. well. barring things like drunk driving or distracted driving. Humans are dumb. autonomous driving was designed by humans and isn’t any smarter)

      • fine_sandy_bottom@lemmy.federate.cc
        link
        fedilink
        arrow-up
        3
        ·
        2 months ago

        You’re not wrong, but that’s not really what I meant although perhaps I didn’t explain it very well.

        Another way to say the same thing, if you group together all the various components or aspects of “driving”, 95% of them might be solved relatively easily, but getting the last 5% right is extraordinarily difficult.

        It’s deceiving because the first time you saw a Level 2 car in 2018 it’s natural to think that if they’ve made so much progress seemingly overnight, then surely in the next few years we will have Level 6 cars.

        I do take your point that humans are also good drivers 95% of the time and mistakes only occur within 5% of situations. The issue there is the imperative that autonomous cars must be better than a human in all circumstances. If a human makes, on average, 5 serious mistakes every 500,000km, but an autonomous car makes 6, you’d probably not want to put your family in that autonomous car.

        • FuglyDuck@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 months ago

          well, to be fair, I’m pretty ardently apposed to self driving cars- i wouldn’t put my family in one even if it was better than humans.

          The reason being is that all it takes is one bug in one code and your entire family is fucked; and with the way corporations are now handling updates, I frankly don’t trust them to maintain it properly at all. (AKA forced updates, with shit-for-testing cloudstrike, for example; could have been prevented, but they just had to push that update globally, all at the same time, to everyone. imagine a malicious admin doing a terror attack by making all your cars crash. or some intern pushing the wrong code and suddenly your car is bricked. I’ll keep my dumb car, thank you very much.)

          • fine_sandy_bottom@lemmy.federate.cc
            link
            fedilink
            arrow-up
            1
            ·
            2 months ago

            Yeah. I tend to agree.

            Being able to drive without killing someone is only one aspect of an autonomous vehicle, and security is one that I’m not confident about in the least.

            I’ve noticed that my wife’s Level 2 car is just hopeless outside of the city. Sure that’s where most people live and it’s fine for most people.

            Driving on country roads it spends more time having self-disabled it’s autonomous features than not, simply because it can’t see the road or what have you.

  • NeoNachtwaechter@lemmy.world
    link
    fedilink
    arrow-up
    11
    arrow-down
    1
    ·
    2 months ago

    Was i just gambling my life or has the tech somehow gotten worse?

    We can safely assume that tech has evolved to the better during this time.

    But it is still too dangerous. It should not be allowed on public roads yet.

    • BlameThePeacock@lemmy.ca
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      3
      ·
      2 months ago

      The public honestly shouldn’t be allowed on roads yet either… they’re fucking terrible at driving.

    • Drivebyhaiku@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      2 months ago

      That the tech has evolved to be better actually is an assumption. The novel data problem hasn’t been meaningfully addressed really at all so mostly we assume that progress has been made… but it’s not meaningful progress. The promises being made for future capability is mostly pretty stale hype that hasn’t changed year to year with a lot of the targets remaining unchanged. We are getting more data on where specifically and how it’s failing, which is something, but overall it appears to be a plateau of non-linear progress with different updates being sometimes less safe than newer ones.

      That actually safe self driving cars might be decades away however is antithetical to the hype run marketing campaigns that are working overtime to put up smoke and mirrors around the issue.

  • gusgalarnyk@lemmy.world
    link
    fedilink
    arrow-up
    9
    ·
    2 months ago

    I appreciate these comments saying the tech hasn’t degraded and it’s been standstill, or that it was never great in the first place, all of which is true but I would like to interject my own Model 3 experience. When we first bought the Tesla in 2019 the self driving functionality on the highway felt safe and functional in nominal conditions. When we sold the Tesla 2 years ago (2022) the self driving felt noticably more finicky. It struggled to switch lanes, recognize when lanes started and ended, and had noticably more issues with maintaining proper speed and distance with other cars.

    It probably wasn’t significantly more dangerous, but it felt like it was. What was a feature we used for the first year or two without much complaint turned into something we never used and our driving time when down in that third year not up so it wasn’t exposure time I don’t think.

    • ContrarianTrail@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 months ago

      That’s Autopilot though, not FSD. Untill not too long ago it was still switching to Autopilot on highways where as in cities it was using an entirely different system. Nowdays the neural nets based system is used on highways too assuming you have FSD subscribtion. If not, then it’ll keep using Autopilot which is less advanced system.

  • hendrik@palaver.p3x.de
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    2 months ago

    Probably the gambling thing. And it’s not like they wreck themselves every 50 miles, so the anecdotal evidence doesn’t really apply. I mean unles they can’t do that anymore. But I think the negative press is about some more rare incidents.

  • ch00f@lemmy.world
    link
    fedilink
    arrow-up
    7
    ·
    2 months ago

    The first Model X has Autopilot 1 which was a system designed by Mobileye. Tesla’s relationship with Mobileye fell apart and they replaced it with an Nvidia based system in 2017(?). It was really really bad at the start as they were essentially starting from scratch. This system also used 8 cameras instead of the original 1.

    Then Tesla released AP hardware 3 which was a custom-built silicon chip designed specifically for self-driving which also enabled proper navigation of surface streets in addition to the just highway lanekeeping offered in AP1. This broadened scope of actually dealing with turns and traffic from multiple angles is probably where the reputation of it being dangerous has come from.

    My HW3 enabled Model 3 does make mistakes, though it’s rarely anything like hitting a pedestrian or running off the road. Most of my issues are with navigational errors. If the GPS gets messed up in the tunnel, it’ll suddenly decide to take an exit that it isn’t supposed to, or it’ll get in the left lane to pass someone 1/4 mile from a right-exit.

      • ch00f@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        2 months ago

        Heh, I guess I should have phrased that differently.

        But yeah, it’s actually really courteous. Sometimes a little too much. It’ll move over to the left side of the lane if it sees a cyclist or pedestrian on the shoulder to the right. Unfortunately, it doesn’t understand when there’s a 3 ft concrete barrier between me and the pedestrian and will do it anyway. Makes some narrow bridge crossings a little scarier than necessary.

  • thesohoriots@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    ·
    2 months ago

    There are also many more Teslas on the road, and the “full self driving” incidents are more widely reported on since the new ownership likes to overpromise and vastly underdeliver. Other commenters have already addressed the tech side, but a few years ago, the Tesla-specific FSD was found to be active right up until a split second before some prolific collisions with emergency vehicles, leading to speculation on liability. Tesla aside, I think it’s just laziness on the part of drivers used to FSD doing the menial tasks of driving.

  • NevermindNoMind@lemmy.world
    link
    fedilink
    arrow-up
    6
    ·
    2 months ago

    Just a guess, but it’s probably a combination of two things. First, if we say a self driving car is going to hit an edge case it can’t resolve once in every, say, 100,000 miles, the number of Tesla’s and other self driving cars on the roads now means more miles driven more frequently which means those edge cases are going to occur more frequently. Second, people are becoming over reliant on self driving - they are (incorrectly ) trusting it more and paying less attention, meaning less chance of human intervention when those edge cases occur. So probably the self driving is overall better, but the number of accidents overall is increasing.

  • Phoenixz@lemmy.ca
    link
    fedilink
    arrow-up
    5
    arrow-down
    2
    ·
    2 months ago

    No. Elon nusk simply is a liar on the level of trump and always claimed Tesla’s could self-drive then from one end of the country to another whereas in reality it wouldn’t be able to get itself out of a parking lot.

    Any claim and or idea from musl can basically be considered a lie

  • WastedJobe@feddit.org
    link
    fedilink
    arrow-up
    3
    ·
    2 months ago

    A guy on YT used to (or still does idk) upload entire FSD trips after every update. I haven’t watched any lately but the trend was that the software was improving with some regressions every now and then, which were fixed in the next updates.

  • bitchkat@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    2 months ago

    That car had AP1 which used hardware developed by MobileEye. Tesla doest like licensing software so they ditched that for a homegrown solution.

  • ContrarianTrail@lemm.ee
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    9
    ·
    edit-2
    2 months ago

    Tesla recently switched from human-written code to full neural networks with their FSD system, which resulted in a significant leap in how reliable and human-like it became. Nowadays, it’s safe to assume it will perform better than the average human driver on most trips, but it’s still not reliable enough to be blindly trusted. You still need to sit there, paying attention to make sure it doesn’t do anything stupid. It’s, by far the most advanced ‘self-driving’ system available on a vehicle you can buy.

    The reason it’s constantly in the news is that anything related to Elon, AI, or self-driving is guaranteed to get clicks in today’s mediascape. It’s not a flawless system by any means but it’s also not as bad as people make it to be.