Craig Doty II, a Tesla owner, narrowly avoided a collision after his vehicle, in Full Self-Driving (FSD) mode, allegedly steered towards an oncoming train.

Nighttime dashcam footage from earlier this month in Ohio captured the harrowing scene: Doty’s Tesla rapidly approaching a train with no apparent deceleration. He insisted his Tesla was in Full Self-Driving mode when it barreled towards the train crossing without slowing down.

  • magnetosphere@fedia.io
    link
    fedilink
    arrow-up
    242
    arrow-down
    2
    ·
    7 months ago

    Man, am I glad that I couldn’t afford a Tesla when I thought they were cool and didn’t find Musk repulsive.

        • Logh@lemmy.ml
          link
          fedilink
          English
          arrow-up
          16
          ·
          7 months ago

          The hypersub sounds difficult, but it’s really easy. I know you are hungry now, but it’s worth it to wait a few years. We’ll make enough hypersubs for everyone by 2028.

        • Evotech@lemmy.world
          link
          fedilink
          English
          arrow-up
          8
          ·
          7 months ago

          And when you get it it’s clearly missing some of the ingredients but he tells you to trust him

        • Aceticon@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          7 months ago

          And when you get it the bread is cardboard, the cheese is cheap factory stuff but there is little flag pinned on top and sauces with quit daring appearences (some kind of gel with little stars and a bright fluerscent yellow one) have been applied with a clear decorative intent.

      • sp3ctr4l@lemmy.zip
        link
        fedilink
        English
        arrow-up
        9
        ·
        edit-2
        7 months ago

        At this point, I believe I know more about sandwich making than any human alive on Earth.

        Scene: a gigantic pop up tent with diesel generators in a desert, featuring many granite counters as well as top of the line kitchen appliances, and gamer lighting, all being set up in by workers who are immediately laid off once the jobsite is completed

        Elon enters the tent with 4 Tesla bots slowly shambling behind him. One lags out when its remote link to a human controller is severed. Minutes later the remnants of a starlink satellite crash through the far end of the tent kitchen.

        3 remaining Tesla bots proceed to bumble around like idiots, unable to open packets of deli ham, entirely ripping off the tops of deli mustard containers

        Elon is awkwardly smiling and doing jazz hands the whole time

        A neuralink mind controlled pig walks in as one Tesla bot wields a knife. Elon raises his hand to his ear, nods, then pushes a button on some phone app

        The pig screams, then a popping noise is heard, and the pig collapses to the ground with smoke and blood coming out of its ears and nose

        knife wielding tesla bot attempts to cut the pig’s flank, falls, cannot recover

        the two remaining tesla bots continue in vain to open a loaf of bread without ripping the entire loaf apart. One slips and falls backwards, the other one runs out of battery and is frozen in place, holding a single piece of wonder bread

        Elon curses, reaches into a refrigerator and hands you a crustable

        • Furbag@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          7 months ago

          Cut him a break, he’s pioneering sandwich making with this innovative tech. A few hiccups are to be expected.

          • sp3ctr4l@lemmy.zip
            link
            fedilink
            English
            arrow-up
            2
            ·
            7 months ago

            Don’t say “cut him a break” around a Tesla product, god knows what command it’ll interpret that as.

    • Scotty_Trees@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      7 months ago

      Just watched the video. Maybe it’s just me, but if you’re the driver in a Tesla, and it’s foggy as fuck outside, maybe, just maybe, don’t use the self driving aspect when visibility is that bad. The amount of people willing to trust Tesla with their lives (and others on the road…) is too damn high!

      • Scolding7300@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        7 months ago

        I’d put the blame on the branding, full self drive shluld mean full self drive, not most-conditions self drive without explicitly providing the limitations. Even irplane autopilot systems, which solve a simpler problem, have explicit limitations stated

        • assassin_aragorn@lemmy.world
          link
          fedilink
          English
          arrow-up
          9
          ·
          edit-2
          7 months ago

          There really needs to be legal pressure for them to change the name. I don’t see how it’s not false advertising.

    • EtherWhack@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      3
      ·
      7 months ago

      More and more they are starting to look like someone just vacuum molded the body over a chassis rather than putting any sort of real artistry to it. It also makes me wonder if Ellen is saying the look is “sexy,” does he have some sort of spandex fetish. Maybe he stole his mum’s pantyhose when alone and strutted around the house wearing them, thinking of himself as such. Maybe he still does…

    • AmidFuror@fedia.io
      link
      fedilink
      arrow-up
      32
      ·
      7 months ago

      So sick of people referring to “Do not ram train” mode. You see it all over social media, but especially Lemmy. It’s “Do not ram train (Supervised)” mode, and you’d have to be living under a rock for the last 5+ years to think you don’t have to actually take control of the wheel to stop it from ramming a train.

  • ShittyBeatlesFCPres@lemmy.world
    link
    fedilink
    English
    arrow-up
    79
    ·
    7 months ago

    If self-driving A.I. models need a workforce to help identify trains, we could probably assemble an army of toddlers willing to be paid in cookies. My friend’s kid gets HYPE and yells “TRAIN!” when he sees one. He can also reliably identify cows. He calls most construction equipment “big truck” but that might be good enough. If a Tesla thinks a backhoe is a big truck, it’ll avoid it.

      • assassin_aragorn@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        7 months ago

        No, newborns rapidly take in new information and learn. “AI” is just a sophisticated text probability model. It doesn’t know anything. It isn’t learning how things work. It just regurgitates.

        It’s like the difference between a student who understands the concepts versus memorizes the test answers.

      • Thorny_Insight@lemm.ee
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        23
        ·
        7 months ago

        It’s about as intelligent as a newborn.

        Newborns can’t even utter one cohesive word. I don’t get the point of making such an obviously false claims about anything.

  • itsonlygeorge@reddthat.com
    link
    fedilink
    English
    arrow-up
    78
    arrow-down
    2
    ·
    7 months ago

    Tesla opted not to use LIDAR as part of its sensor package and instead relies on cameras which are not enough to determine accurate location data for other cars/trains etc.

    This is what you get when billionaires cheap out on their products.

    • Imalostmerchant@lemmy.world
      link
      fedilink
      English
      arrow-up
      28
      ·
      7 months ago

      I never understood Musk’s reasoning for this decision. From my recollection it was basically “how do you decide who’s right when lidar and camera disagree?” And it felt so insane to say that the solution to conflicting data was not to figure out which is right but only to listen to one.

      • wirehead@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        7 months ago

        I mean, I think he’s a textbook example of why not to do drugs and why we need to eat the rich, but I can understand the logic here.

        When you navigate a car as a human, you are using vision, not LIDAR. Outside of a few edge cases, you aren’t even using parallax to judge distances. Ergo, a LIDAR is not going to see the text on a sign, the reflective stripes on a truck, etc. And it gets confused differently than the eye, absorbed by different wavelengths, etc. And you can jam LIDAR if you want. Thus, if we were content to wait until the self-driving-car is actually safe before throwing it out into the world, we’d probably want the standard to be that it navigates as well as a human in all situations using only visual sensors.

        Except, there’s some huge problems that the human visual cortex makes look real easy. Because “all situations” means “understanding that there’s a kid playing in the street from visual cues so I’m going to assume they are going to do something dumb” or “some guy put a warning sign on the road and it’s got really bad handwriting”

        Thus, the real problem is that he’s not using LIDAR as harm reduction for a patently unsafe product, where the various failure modes of the LIDAR-equipped self-driving cars show that those aren’t safe either.

      • Jakeroxs@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        4
        ·
        7 months ago

        Also that LIDAR is more expensive then cameras, which means higher end user price, as far as I remember.

      • smonkeysnilas@feddit.de
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 months ago

        I mean the decision was stupid from an engineering point of view, but the reasoning is not entirely off. Basically it follows the biological example: if humans can drive without Lidar and only using their eyes than this is proof that it is possible somehow. It’s only that the current computer vision and AI tech is way worse than humans. Elon chose to ignore this, basically arguing that it is merely a software problem for his developers to figure out. I guess in reality it is a bit more complex.

    • skyspydude1@lemmy.world
      link
      fedilink
      English
      arrow-up
      25
      ·
      7 months ago

      Not only that, but took out the radar, which while it has its own flaws, would have had no issue seeing the train through the fog. While they claimed it was because they had “solved vision” and didn’t need it anymore, it’s bullshit, and their engineering team knew it. They were in the middle of sourcing a new radar, but because of supply chain limitations (like everyone in 2021) with both their old and potential new supplier, they wouldn’t continue their “infinite growth” narrative and fElon wouldn’t get his insane pay package. They knew for a fact it would negatively affect performance significantly, but did it anyway so line could go up.

      While no automotive company’s hands are particularly clean, the sheer level of willful negligence at Tesla is absolutely astonishing and have seen and heard so many stories about their shitty engineering practices that the only impressive thing is how relatively few people have died as a direct result of their lax attitude towards basic safety practices.

    • Wrench@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      3
      ·
      7 months ago

      LIDAR would have similarly been degraded in the foggy conditions that this occurred in. Lasers are light too.

      While I do think Tesla holds plenty of responsibility for their intentionally misleading branding in FSD, as well as cost saving measures to not include lidar and/or radar, this particular instance boils down to yet another shitty and irresponsible driver.

      You should not be relying on FSD over train tracks. You should not be allowing FSD to be going faster than conditions allow. Dude was tearing down the road in thick fog, way faster than was safe for the conditions.

      • Rekorse@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        3
        ·
        7 months ago

        A Tesla drover might get the impression that the cars “opinion” is better than their own, which could cause them to hesitate before intervening or to allow the car to drive in a way they are uncomfortable with.

        The misinformation about the car reaches the level of negligence because even smart people are being duped by this.

        Honestly I think some people just dont believe someone could lie so publicly and loudly and often, that it must be something else besides a grift.

      • Pazuzu@midwest.social
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        7 months ago

        Maybe it shouldn’t be called full self driving if it’s not fully capable of self driving

      • FreddyDunningKruger@lemmy.ml
        link
        fedilink
        English
        arrow-up
        2
        ·
        7 months ago

        One of the first things you learn to get your driver’s license is the Basic Speed Law, you must not drive faster than the driving conditions would allow. If only Full Self Driving followed the law and reduced its max speed based on the same.

        • Rekorse@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          1
          ·
          7 months ago

          If you were to strictly take that rule seriously, you should not allow FSD to drive at all, as at any speed its more dangerous than the person driving it (given an average driver who’s not intoxicated).

  • karrbs@kbin.social
    link
    fedilink
    arrow-up
    60
    arrow-down
    17
    ·
    7 months ago

    If he had time to notice it not slowing down he had time to brake and take it out of full self driving. I understand as someone who is sceptical about the fsd mode that I am more proactive at taking over than those who trust it a little bit more. I just feel if a company tells you to supervise it you should supervise it.

    I still find fsd to be very finicky and vastly oversold

    • assembly@lemmy.world
      link
      fedilink
      English
      arrow-up
      96
      arrow-down
      7
      ·
      7 months ago

      If something is sold as fully self driving, I would like to think it should be capable of fully self driving and not a feature that will drive me face first into a train.

      • CrowAirbrush@lemmy.world
        link
        fedilink
        English
        arrow-up
        33
        ·
        7 months ago

        I too come from a time where company’s had to sell functional products or go bankrupt, but alas those days are long gone.

      • flyingjake@lemmy.one
        link
        fedilink
        English
        arrow-up
        19
        arrow-down
        1
        ·
        7 months ago

        To be fair, it could have fully driven itself into the train: “fully self driving” <> “fully safe driving” /s

      • halcyoncmdr@lemmy.world
        link
        fedilink
        English
        arrow-up
        29
        arrow-down
        24
        ·
        edit-2
        7 months ago

        Regardless of the naming, because everyone gets so stuck on fucking names and seems to ignore everything else because that makes for a quick comment with a ton of votes and feel good bullshit.

        It is sold as a work in progress piece of software that is constantly being updated and still needs to be supervised. It has a ton of warnings about it’s capabilities, and lack thereof when activating it. There is no question when actually setting up FSD in the vehicle that it is something still in testing and not to be treated as a full replacement for paying attention. It constantly watches you and will warn you if you aren’t paying active attention for too long. If you ignore those warnings enough it will deactivate itself, forcing you to drive, and with enough deactivation will remove the capability entirely.

        Image of the activation screen.

        Image of the free trial page for those that have not purchased it, thereby avoiding notes on the standard sales pages

        They’ve even updated the setting in the vehicle to be more specific, showing it as “Full Self-Driving (Supervised)”…
        https://www.tesla.com/ownersmanual/modely/en_us/GUID-2CB60804-9CEA-4F4B-8B04-09B991368DC5.html

        All of these reported situations are from people actively ignoring numerous safety and attention warnings, yet no one seems to ever put any blame on the driver in comments or articles. It’s always about blaming everything on Tesla when they’re actively telling every driver that it needs to be supervised because it will make mistakes.

        • Excrubulent@slrpnk.net
          link
          fedilink
          English
          arrow-up
          31
          arrow-down
          5
          ·
          7 months ago

          You’re right that it’s just a name, which means it’s within Tesla’s power to not call it “full self driving”. Like maybe keep the word “full” for when it’s better than “full self driving brackets not really”.

          The reason it’s called that is so when you’re buying the car, you can read “full self driving”, the salesman can call it “full self driving”, and then you can get excited and think you’re getting full self driving and pay stupid amounts of money for an iPhone on wheels.

          It’s also so Musk can get up on stage and lie for years about how you’ll be able to go coast to coast while you sleep by the end of the year or whatever it is. Having a bunch of warnings in the software setup is not enough for someone gargling Musk’s jizz to cough it up and see it for the bullshit that it is.

          You can give us all this extra info but you can’t change the core reality that the name is a lie.

          • ShepherdPie@midwest.social
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            23
            ·
            7 months ago

            The reason it’s called that is so when you’re buying the car, you can read “full self driving”, the salesman can call it “full self driving”, and then you can get excited and think you’re getting full self driving and pay stupid amounts of money for an iPhone on wheels.

            Who exactly are you describing here? Like someone who’s been living under a rock for the past 5+ years and thinks cars just drive themselves now, but who also has $50k-$100k to drop on a car and also doesn’t do any research beforehand? I think you’re blowing this way out of proportion.

            All these systems are flawed to some extent as nobody has cracked the code to L5 automation. There is some danger to it but there are many dangers to driving and this eliminates some of them. People have died in Teslas but many more have died in every other production vehicle that has ever existed. If this guy did the samr thing in a Toyota Camry, do you think we’d even be talking about it right now? These systems can only get better with a lot of real-world usage.

            I totally get where the other person is coming from. These arguments are so tired and meaningless at this point. If you want Musk to go away, then stop bringing him up at every turn because it just makes people sound like bizzaro-world Musk fans as they hate the guy but can’t stop talking about him and following his every move. Some of us want to discuss and debate the actual technology and Musk had nothing to do with developing it.

            • Excrubulent@slrpnk.net
              link
              fedilink
              English
              arrow-up
              12
              arrow-down
              1
              ·
              7 months ago

              Musk was involved in marketing and lying about it, though, and his extremely prolific public image is what gave it so much credibility. He’s lost a lot of that credibility now though, largely because people have spent a lot of time criticizing him. If you want him to disappear from the public eye that’s great, so do I, but he’s one of the most powerful men in the world, so that’s not going to happen.

              • ShepherdPie@midwest.social
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                7 months ago

                Well why are you lending him further credibility by continuing to make him prolific? You’re just feeding into his goal of garnering more attention at every turn. Media outlets will continue to focus on all his insane ramblings because people like you are guaranteed to click on it.

              • ShepherdPie@midwest.social
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                1
                ·
                7 months ago

                The systems with LIDAR aren’t L5 either. It’s impossible to claim what is and isn’t needed when nobody has actually come up with a solution.

                • 0x0@programming.dev
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  7 months ago

                  Never said they were, but LIDAR is a big improvement over “cameras and AI”. Most other manufactures use it, Tesla stopped using it because of cost sav… because their AI kicks ass and if eyes are enough for humans they’re enough for their cars too.

                  Hell any system where lives are involved should be triple-redundancy.

          • halcyoncmdr@lemmy.world
            link
            fedilink
            English
            arrow-up
            9
            arrow-down
            28
            ·
            7 months ago

            So tired of the same arguments. They don’t mean anything in the real world. Complain all you want about what the shit is called, it makes no real world difference.

            Legally, the driver is responsible for the fucking vehicle and these articles and comments like yours just give the impression you think they shouldn’t be responsible because of what it’s called. You’re giving shitty drivers a pass because they’re actively being stupid and you don’t like what Tesla names the software. That is the stupidest take in the world.

            Go ahead in a court of law and claim you are not responsible for an accident that happens while FSD is activated and let’s see whether the name matters for your liability.

            • Excrubulent@slrpnk.net
              link
              fedilink
              English
              arrow-up
              25
              arrow-down
              1
              ·
              7 months ago

              I notice you danced around the question of whether it was a lie to focus on bullshit legal stuff, which isn’t the arbiter of truth and reality. As I once heard a judge say, “You don’t come here for justice, you come here for a judgment.”

              Anyway since you think it matters, the lawsuit you’re talking about is happening and a judge has ruled that the case has merit to continue: https://www.reuters.com/legal/tesla-must-face-vehicle-owners-lawsuit-over-self-driving-claims-2024-05-15/

              • halcyoncmdr@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                14
                ·
                7 months ago

                Merit to continue just means it’s not clearly a bullshit lawsuit that should be thrown out to avoid wasting court time. Your linked lawsuit also is not about whether someone is legally responsible for a vehicle driving itself, it is again about the marketing.

                I don’t give a shit whether their marketing is a lie. Marketing is not the issue at hand, as much as you all want it to be for whatever reason instead of actually blaming the shitty drivers. It has no bearing on whether someone is responsible for the vehicle they are in the driver seat of hitting something or someone.

                Why do you not want to put any blame on these drivers? Drivers that ignore the warning when they turned on the function in the first place and that warns them every time they turn it on to still pay attention. Why are you so insistent that the blame should be on Tesla because of what they call it?

                Makes me start to think you’re that type of driver and trying to justify your belief that you shouldn’t be responsible for the actions of a 2 ton murder machine traveling at high speed that you are in control of.

                • Excrubulent@slrpnk.net
                  link
                  fedilink
                  English
                  arrow-up
                  13
                  ·
                  7 months ago

                  I don’t give a shit whether their marketing is a lie.

                  I mean clearly. “Full self-driving” is marketing, and it is a lie. That’s the point being made here. You don’t have to care about it, but that doesn’t mean it’s not a lie.

                  I’m not not blaming the drivers. They were foolish enough to buy a Tesla and trust it with their lives for a start. But I am also blaming the marketing. Two things can be true.

            • baru@lemmy.world
              link
              fedilink
              English
              arrow-up
              21
              arrow-down
              1
              ·
              7 months ago

              So tired of the same arguments. They don’t mean anything in the real world

              The second sentence is a fallacy.

              And it does matter that the company is calling it full self driving while it doesn’t fully self drive. That it would have that capability is something Musk has promised for many years. It’s also a reason that Tesla stock is worth so much.

              Go ahead in a court of law and claim you are not responsible for an accident

              That’s would be a very specific case. Tesla has been reminded multiple times that they need to take into account how people use their vehicles. The company is also under investigation for possible fraud because they are selling something that doesn’t do what people would respect it to do.

              You’re focusing on one thing, but there’s multiple ways that the company could be liable. There’s been multiple articles explaining that the company is either under investigation or that the company has been warned to change things or else.

              • mojofrododojo@lemmy.world
                link
                fedilink
                English
                arrow-up
                9
                ·
                7 months ago

                Well put. It’s a funny thing, words have meaning and if you advertise your products with those words, some portion of the population (gasp!) might believe you!

              • Thorny_Insight@lemm.ee
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                9
                ·
                7 months ago

                How does it not fully self drive? What’s your definition of full self driving then?

                Mercedes Drive Pilot is Level 3 and even it will prompt you to take over when necessary, does it not fully self drive then either? What about Waymo/Cruze? They have remote operators controlling the vehicles when they get stuck. Not fully self driving either? Is the standard that it needs to be absolutely flawless and never fail or what is it?

            • assassin_aragorn@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              7 months ago

              They don’t mean anything in the real world.

              Uh. They mean everything in the real world? You get sued for false advertising and fraud. Fox News got sued heavily for knowingly lying about voting machines. There’s a reason that companies have PR departments. Words matter a lot in the real world.

              Are the drivers stupid? Sure. They believed the FSD claim after all. But that doesn’t mean Tesla is off the hook. Deceiving stupid people is still deceit.

        • Shrank7242@lemmy.zip
          link
          fedilink
          English
          arrow-up
          11
          arrow-down
          14
          ·
          edit-2
          7 months ago

          Well said and thanks for posting the examples. It’s something that bothers me about any social media kind of site. Especially here on Lemmy. Nobody gives a damn about the incredible amount of negligence the drivers must have. It immediately becomes an anti Elon circlejerk every time.

          It’s similar with news articles, which this post doesn’t even link to, most of the articles name drop Tesla or Elon just because otherwise it’s not a story. “Somebody hit a car / person / train because they weren’t paying attention to the road” isn’t story worthy. But as soon as doubt can be cast on an Elon company, it become a must post thing. I can’t stand Musks antics either, but he gets too much free rent in peoples mind. It’s wild

          /rant

          • baru@lemmy.world
            link
            fedilink
            English
            arrow-up
            13
            ·
            7 months ago

            At the same time, there’s too many people who say that Full Self Driving obviously doesn’t mean that the vehicle still fully drive itself. Though for unknown reasons it is totally fine to keep using the name Full Self Driving.

            • dream_weasel@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              2
              ·
              7 months ago

              And that is between Tesla and the NTSB to sort even though I agree. The car itself doesn’t mince words describing it to you, and at the time you’re driving it, the required supervision is unambiguous.

              If it were called “Tesla Drive” or something else, everyone would still be here taking a shit on it nonetheless.

            • Thorny_Insight@lemm.ee
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              7
              ·
              7 months ago

              It’s called Full Self Driving (Supervised) nowdays. They changed the name.

              The vehicle is capable of driving you to the grocery store on the other side of the city and back, sometimes with zero interventions from the driver. If that’s not Full Self Driving then I don’t know what is.

        • Thorny_Insight@lemm.ee
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          9
          ·
          7 months ago

          Even this “article” is about nothing happening. The driver was paying attention and took over when the vehicle was about to do something it should. Just as they should.

          Also, even if FSD was 10x safer than a human driver and we replaced every single car on the roads with Teslas there would still be 8 people dying every single day in the US alone. Linking articles about these accidents does not prove it being unsafe. It only feeds the confirmation bias of the person posting it and the people upvoting it. People want it to be unsafe so that they can shit on Elon. The standards they apply to Tesla are ridiculous compared to that of other companies. The extremely limited Mercedes Drive Pilot is praised as revolutionary tech while FSD already checks most boxes for Level 4 self-driving.

      • karrbs@kbin.social
        link
        fedilink
        arrow-up
        7
        arrow-down
        5
        ·
        7 months ago

        Just some insight from my pov. Fsd is marketed as FSD (Supervised). I don’t agree with the jamming but it is what it is. I know it does janky stuff, it still forces you to pay attention. Do I believe this could happen, yes but do I doubt the driver always until proven otherwise.

        I have had my model y yell at me to take control when I was already out of any auto/fsd mode. I have many downs and many ups. I agree that the car should actively steer you into the train. I was curious if anyone had a link to the dash footage or even to an article with it.

        • mojofrododojo@lemmy.world
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          1
          ·
          7 months ago

          Fsd is marketed as FSD (Supervised)

          it is, now, it was not marketed with any kind of parenthetical qualifier until recently.

          • halcyoncmdr@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            5
            ·
            7 months ago

            Actually, it said Full Self Driving (BETA) until it was updated to (Supervised) recently.

            If anything, the beta qualifier is actually better than just saying supervised since that term means not complete and still being developed.

            • mojofrododojo@lemmy.world
              link
              fedilink
              English
              arrow-up
              6
              ·
              7 months ago

              never heard musk refer to it as anything but full self drive, no qualifiers.

              companies concerned with safety wouldn’t market the shit until it’s safe (see mercedes apparent lead).

              • Thorny_Insight@lemm.ee
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                3
                ·
                7 months ago

                Mercedes Drive Pilot is hilariously limited system. It for example needs a car in front of it that it can follow or else it wont work. It also only works on limited number of hand-picked highways in California and Nevada.

                There’s a video on YouTube comparing FSD to Mercedes’ equivalent driver assistant software (not the level 3 one) and it’s not even a competition. The Mercedes system is completely unusable.

      • Maddier1993@programming.dev
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        7 months ago

        Those should be your expectations when you are on the shop floor and that should allow you to reject the purchase if it’s s deal breaker for you. Not when you’re crossing railway tracks.

        • 5C5C5C@programming.dev
          link
          fedilink
          English
          arrow-up
          13
          ·
          7 months ago

          Weird how this notion of “personal responsibility” applies to every person except for those people who choose to intentionally misrepresenting the product by branding it in ways that are misleading. The people running this company aren’t responsible for their role in misleading the public, just because the fine print happens to indicate that the product isn’t actually what it’s marketed as?

          Now you’ll probably say something to the effect of “I never said that! You’re putting words in my mouth!” except what other motivation can you have to jump to the defense of the liar and blame people for being misled, except that you want to put all the responsibility on individuals for being misled and not on the company that is systematically and intentionally misleading them? Maybe you just manage to derive a smug sense of superiority thinking of yourself as someone who is invulnerable to this kind of tactic so blaming the victims lets you feel good about yourself.

          • Thorny_Insight@lemm.ee
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            6
            ·
            7 months ago

            You literally cannot buy FSD without being told that it needs driver supervision. It also tells you that every single time you enable it and it’s constantly nagging to you when you take your hands off the wheel aswell as if you’re looking at your phone etc. and given enough warnings the system locks you out of it.

            Has Musk been dishonest/misleading about it’s capabilities in the past? Yes. Is there a single Tesla owner with FSD who doesn’t know the truth? No.

    • mojofrododojo@lemmy.world
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      1
      ·
      7 months ago

      full self driving

      I just take serious issue with this label. It’s not fully self driving, it requires the user’s attention. the accidents and years of promises broken are just cherries on the shit sundae.

    • cedarmesa@lemmy.world
      link
      fedilink
      English
      arrow-up
      20
      arrow-down
      11
      ·
      7 months ago

      Elon here. Thanks for fighting the good fight. We will deposit 100 X bucks in your account.

    • NeoNachtwaechter@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      7 months ago

      If he had time to notice it not slowing down he had time to brake and take it out of full self driving.

      And that’s the way he survived. But wrecked that plastic box.

  • FangedWyvern42@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    arrow-down
    1
    ·
    7 months ago

    Every couple of months there’s a new story like this. And yet we’re supposed to believe this system is ready for use…

    • darki@lemmy.world
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      6
      ·
      7 months ago

      It is ready because Musk needs it to be ready. Watch out, this comment may bring the morale down, and Elron will be forced to … Cry like a baby 😆

      • Buffalox@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        7 months ago

        Didn’t he recently claim Tesla robotaxi is only months away?
        Well I suppose he didn’t say how many months, but the implication was less than a year, which has been his claim every year since 2016.

        • dustyData@lemmy.world
          link
          fedilink
          English
          arrow-up
          11
          ·
          7 months ago

          He said that Teslas were an investment worth hundreds of thousands of dollars because owners would be able to use them as robot taxis when they weren’t using their car and charge a small fee by next year…in 2019. Back then he promised 1 million robot taxis nationwide in under a year. Recently he gave the date august 8 to reveal a new model of robot taxi. So, by Cybertruck estimates, I would say a Tesla robot taxi is a possibility by late 2030.

          He is just spewing shit to keep the stock price afloat, as usual.

          • dual_sport_dork 🐧🗡️@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            7 months ago

            He also said they were ready to manufacture the 2nd generation Tesla Roaster “now,” which was back in 2014. No points for guessing that as of yet (despite taking in millions of dollars in preorders) they have not produced a single one.

            Given this very early and still quite relevant warning, I’m astounded that anyone is dumb enough to believe any promise Elon makes about anything.

    • Thorny_Insight@lemm.ee
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      20
      ·
      7 months ago

      In what way is it not ready to use? Does cars have some other driver assistant features that are fool proof? You’re not supposed to blindly trust any of those. Why would FSD be an exception? The standards people are aplying to it are quite unreasonable.

      • ammonium@lemmy.world
        link
        fedilink
        English
        arrow-up
        22
        arrow-down
        3
        ·
        7 months ago

        Because it’s called Full Self Drive and Musk has said it will be able to drive without user intervention?

        • dream_weasel@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          8
          arrow-down
          10
          ·
          7 months ago

          The naming is poor, but in no way does the car represent to you that no intervention is required. It also constantly asks you for input and even watches your eyes to make sure you pay attention.

            • dream_weasel@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              7 months ago

              Marketing besides the naming we have already established and Elon himself masturbating to it? Is there some other marketing that pushes this narrative, because I certainly have not seen it.

        • Thorny_Insight@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          12
          ·
          7 months ago

          It’s called Full Self Driving (Supervised)

          Yeah, it will be able to drive without driver intervention eventually. Atleast that’s their goal. Right now however, it’s level 2 and no-one is claiming otherwise.

          In what way is it not ready to use?

      • Piranha Phish@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        7 months ago

        It’s unreasonable for FSD to see a train? … that’s 20ft tall and a mile long? Am I understanding you correctly?

        Foolproof would be great, but I think most people would set the bar at least as high as not getting killed by a train.

        • Thorny_Insight@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          7 months ago

          Did you watch the video? It was insanely foggy there. It makes no difference how big the obstacle is if you can’t even see 50 meters ahead of you.

          Also, the car did see the train. It just clearly didn’t understand what it was and how to react to it. That’s why the car has a driver who does. I’m sure this exact edge case will be added to the training data so that this doesn’t happen again. Stuff like this takes ages to iron out. FSD is not a finished product. It’s under development and receives constant updates and keeps improving. That’s why it’s classified as level 2 and not level 5.

          Yes. It’s unreasonable to expect brand new technology to be able to deal with every possible scenario that a car can encounter on traffic. Just because the concept of train in a fog makes sense to you as a human doesn’t mean it’s obvious to the AI.

          • Piranha Phish@lemmy.world
            link
            fedilink
            English
            arrow-up
            11
            arrow-down
            1
            ·
            7 months ago

            In what way is it not ready to use?

            To me it seems you just spent three paragraphs answering your own question.

            can’t even see 50 meters ahead

            didn’t understand what it was and how to react to it

            FSD is not a finished product. It’s under development

            doesn’t mean it’s obvious to the AI

            If I couldn’t trust a system not to drive into a train, I don’t feel like I would trust it to do even the most common tasks. I would drive the car like a fully attentive human and not delude myself into thinking the car is driving me with “FSD.”

              • Piranha Phish@lemmy.world
                link
                fedilink
                English
                arrow-up
                6
                ·
                7 months ago

                Completely true. And I would dictate my driving characteristics based on that fact.

                I would drive at a speed and in a manner that would allow me to not almost crash into things. But especially trains.

                • Thorny_Insight@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  5
                  ·
                  7 months ago

                  I agree. In fact I’m surprised the vehicle even lets you enable FSD in that kind of poor visibility and based on the video it seemed to be going quite fast aswell.

                • Thorny_Insight@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  2
                  ·
                  7 months ago

                  Yeah there’s a wide range of ways to map the surroundings. Road infrastructure, however is designed for vision so I don’t see why just cameras wouldn’t be sufficient. The issue here is not that it’s didn’t see the train - it’s on video, after all - but that it didn’t know how to react to it.

      • Holyginz@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        2
        ·
        7 months ago

        No, the standards people are applying to it are the bare minimum for a full self driving system like what musk claims.

        • Thorny_Insight@lemm.ee
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          8
          ·
          7 months ago

          It’s a level 2 self driving system which by definition requires driver supervision. It’s even stated in the name. What are the standards it doesn’t meet?

      • assassin_aragorn@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        7 months ago

        You’re not supposed to blindly trust any of those. Why would FSD be an exception?

        Because that’s how Elon (and by extension Tesla) market it. Full self driving. If they’re saying I can blindly trust their product, then I expect it to be safe to blindly trust it.

        And if the fine print says I can’t blindly trust it, they need to be sued or put under legal pressure to change the term, because it’s incredibly misleading.

        • Thorny_Insight@lemm.ee
          link
          fedilink
          English
          arrow-up
          3
          ·
          7 months ago

          Full Self Driving (Beta), nowdays Full Self Driving (Supervised)

          Which of those names invokes trust to put your life in it’s hands?

          It’s not in fine print. It’s told to you when you purchase FSD and the vehicle reminds you of it every single time you enable the system. If you’re looking at your phone it starts nagging at you eventually locking you out of the feature. Why would they put driver monitoring system in place if you’re supposed to put blind faith into it?

          That is such an old, beat up strawman argument. Yes, Elon has said it would be fully autonomous in a year or so which turned out to be a lie but nobody today is claiming it can be blindly trusted. That simply just is not true.

          • Honytawk@lemmy.zip
            link
            fedilink
            English
            arrow-up
            3
            ·
            7 months ago

            It isn’t Full Self Driving if it is supervised.

            It’s especially not Full Self Driving if it asks you to intervene.

            It is false advertisement at best, deadly at worst.

            • Thorny_Insight@lemm.ee
              link
              fedilink
              English
              arrow-up
              1
              ·
              7 months ago

              It’s misleading advertising for sure. At no point have I claimed otherwise.

              The meaning of what qualifies as “full self driving” is still up for debate however. There are worse human drivers on the roads than what the current version of FSD is capable of. It’s by no means flawless but it’s much better than most people even realize. It’s a vehicle capable of self driving even if not fully.

          • assassin_aragorn@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            7 months ago

            Unfortunately, companies also have to make their products safe for idiots. If the system is in beta or must be supervised, there should be inherently safe design that prevents situations like this from happening even if an idiot is at the wheel.

            • Thorny_Insight@lemm.ee
              link
              fedilink
              English
              arrow-up
              3
              ·
              7 months ago

              ESP is not idiot proof either just to name one such feature that’s been available for decades. It assists the driver but doesn’t replace them.

              Hell, cars themselves are not idiot proof.

            • VirtualOdour@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              ·
              7 months ago

              Yeah and cars should have a system to stop idiots doing dumb things, best we have is a license so if it’s good enough for cars without added safety features is good enough for them with

    • dream_weasel@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      29
      ·
      7 months ago

      Ever couple of months you hear about every issue like this, just like you hear about every airline malfunction. It ignores the base rate of accurate performances which is very high.

      FSD is imperfect but still probably more ready for use than a substantial fraction of human drivers.

      • buddascrayon@lemmy.world
        link
        fedilink
        English
        arrow-up
        27
        arrow-down
        2
        ·
        7 months ago

        This isn’t actually true. The Tesla full self driving issues we hear about in the news are the ones that result in fatal and near fatal accidents, but the forums are chock full of reports from owners of the thing malfunctioning on a regular basis.

        • dream_weasel@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          9
          ·
          7 months ago

          It IS actually true. It does goofy stuff in some situations, but on the whole is a little better than your typical relatively inexperienced driver. It gets it wrong about when to be assertive and when to wait sometimes, it thinks there’s enough space for a courteous merge but there isn’t (it does some Chicago style merges sometimes), it follows the lines on the road like they are gospel, and doesn’t always properly estimate how to come to a smooth and comfortable stop. These are annoying things, but not outrageous provided you are paying attention like you’re obliged to do.

          I have it, I use it, and I make lots of reports to Tesla. It is way better than it used to be and still has plenty of room to improve, but a Tesla can’t reboot without having a disparaging article written about it.

          Also fuck elon, because I don’t think it gets said enough.

          • bane_killgrind@lemmy.ml
            link
            fedilink
            English
            arrow-up
            16
            arrow-down
            1
            ·
            7 months ago

            typical relatively inexperienced driver

            Look at the rates that teenagers crash, this is an indictment.

            provided you are paying attention

            It was advertised as fully autonomous dude. People wouldn’t have this much of a hard-on for trashing it if it wasn’t so oversold.

            • Thorny_Insight@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              4
              ·
              edit-2
              7 months ago

              This fully autonomous argument is beat to death already. Every single Tesla owner knows you’re supposed to pay attention and be ready to take over when necessary. That is such a strawman argument. Nobody blames the car when automatic braking fails to see the car infront of it. It might save your ass if you’re distracted but ultimately it’s always the driver whose responsible. FSD is no different.

              • Pazuzu@midwest.social
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 months ago

                If it’s not fully capable of self driving then maybe they shouldn’t call it full self driving

                • Thorny_Insight@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  edit-2
                  7 months ago

                  Sure. Make then change the name to something different. I’m fine with that.

                  Though I still don’t know what most people actually mean by full self driving and how it’s different from what FSD can do right now.

              • Rekorse@lemmy.dbzer0.com
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 months ago

                You realize it can be true that the driver is at fault when they crash and that the crash was more likely to happen because you have Elon contradicting his own marketing team constantly and confusing people.

                He literally would take reporters in his car and take his hands off the wheel. He just fundamentally doesn’t care about safety now. Probably doesn’t about safety later, just saw a way to make some money.

          • buddascrayon@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            ·
            7 months ago

            Seriously you sound like a Mac user in the '90s. “It only crashes 8 or 9 times a day, it’s so much better than it used to be. It’s got so many great features that I’m willing to deal with a little inconvenience…” Difference being that when a Mac crashes it just loses some data and has to reboot but when a Tesla crashes people die.

      • lolcatnip@reddthat.com
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        1
        ·
        7 months ago

        You hear so much about the people Jeffrey Dahmer murdered, but never anything about all the people he didn’t murder!

          • lolcatnip@reddthat.com
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            7 months ago

            I see you’ve decided to be condescending, and also made a falsifiable claim. This is the part where you bring some actual data or STFU.

            • dream_weasel@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              7 months ago

              Whatever you say Mr Dahmer joke instead of content. I see that was really all in good faith and maybe I unintentionally hurt your feelings by citing a source on base rate biases?

              What data would you like me to bring for discussion since you’ve been so open thus far? Do you want me to bring some data showing that teslas spend more time not having accidents than having accidents? I’m happy to go do some homework to enrich this interaction.

              It’s not as though you can just ask Tesla for every case of an FSD crash. The falsifiable claim is just me tossing a number, the point is that memorable bad press and bad stats are not the same.

              • Rekorse@lemmy.dbzer0.com
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 months ago

                Why would you think there isnt data on Tesla crashes? Are they hiding their broken cars from bystanders and police or something now?

  • buddascrayon@lemmy.world
    link
    fedilink
    English
    arrow-up
    32
    arrow-down
    2
    ·
    7 months ago

    When you look at the development notes on the self-driving at Tesla, anyone with a brain wouldn’t trust that shit, not even a little bit. Most of what they did is placate Musk’s petty whims and delusions. Any real R&D issues were basically glazed over it given quick software fixes.

    • VirtualOdour@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      5
      ·
      7 months ago

      Demonstrate what you mean because it really sounds like you’re describing what you feel should be true to justify your emotions about the bad Twitter man.

      And to be clear, I mean link the documents you claim to have read and the parts of them you claim demonstrate this.

            • WoahWoah@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              edit-2
              7 months ago

              Read the development notes from the first years of any technology you use. The research you’re “referencing” is six years old at this point.

              What’s next? You going to criticize an iPod Nano to make a point about the broken screen on your iPhone 8? Criticize Google assistant from 2019 to harangue OpenAI?

              Look at what six years of development means: https://youtu.be/qTDlRLeDxxM?si=dFZzLcO_a8wfy2QS

              • buddascrayon@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                7 months ago

                It’s not about how well or badly it worked when they were developing it, it’s about the developing process. It’s about the fact that they had to appease Elon Musk’s ego in all aspects of developing the self drive system. To a disastrous degree.

                And again there is a world of difference between the iPhone or Open AI or Google Assistant not working right and a car driving itself not working right because when those other things don’t work nobody dies or gets hurt. But a car can mame and kill people extremely easily.

        • VirtualOdour@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          7 months ago

          That’s a very old article about even older opinions, now totally outdated as shown by statements like;

          Almost five years on, Tesla still hasn’t completed its autonomous road trip — no car company has even come close.

          You’re using unsubstantiated statements from the start of development which is totally different to what you claimed before being asked for a source.

          Current development FSD has hit huge milestones which competitors have not.

  • riodoro1@lemmy.world
    link
    fedilink
    English
    arrow-up
    23
    ·
    7 months ago

    What a bunch of morons people were in 1912 to believe a ship could be unsinkable. Amirite guys?

    • Pazuzu@midwest.social
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 months ago

      The Titanic probably wouldn’t have sunk if it hit the iceberg head on. Clearly the Tesla simply mistook the train for an iceberg and itself for an ocean-liner and opted for a more ideal collision. The driver should have disabled ‘sea mode’ if they didn’t want that behavior, it’s all clearly spelled out in the owners manual.

  • cestvrai@lemm.ee
    link
    fedilink
    English
    arrow-up
    20
    ·
    7 months ago

    As a frequent train passenger, I’m not overly concerned.

    Seems a bit too weak to derail, probably only delay.

  • Noxy@yiffit.net
    link
    fedilink
    English
    arrow-up
    19
    ·
    7 months ago

    Feels like these things were more capable a decade ago when they had radar.

    Not that they should be called “full self driving” either then or now, but at least radar can deal fog better than regular ass cameras

  • MacStache@programming.dev
    link
    fedilink
    English
    arrow-up
    19
    ·
    7 months ago

    Those trains sure are weird and confusing, with their back and forth those tracks and all. Makes you wonder about train safety, it does!