Craig Doty II, a Tesla owner, narrowly avoided a collision after his vehicle, in Full Self-Driving (FSD) mode, allegedly steered towards an oncoming train.
Nighttime dashcam footage from earlier this month in Ohio captured the harrowing scene: Doty’s Tesla rapidly approaching a train with no apparent deceleration. He insisted his Tesla was in Full Self-Driving mode when it barreled towards the train crossing without slowing down.
Man, am I glad that I couldn’t afford a Tesla when I thought they were cool and didn’t find Musk repulsive.
Bro I wouldn’t trust Elon to make me a sandwich.
He’d keep telling you it’s nearly ready but you have already been waiting for hours.
The hypersub sounds difficult, but it’s really easy. I know you are hungry now, but it’s worth it to wait a few years. We’ll make enough hypersubs for everyone by 2028.
And when you get it it’s clearly missing some of the ingredients but he tells you to trust him
The lack of waiters is not his fault.
And when you get it the bread is cardboard, the cheese is cheap factory stuff but there is little flag pinned on top and sauces with quit daring appearences (some kind of gel with little stars and a bright fluerscent yellow one) have been applied with a clear decorative intent.
At this point, I believe I know more about sandwich making than any human alive on Earth.
Scene: a gigantic pop up tent with diesel generators in a desert, featuring many granite counters as well as top of the line kitchen appliances, and gamer lighting, all being set up in by workers who are immediately laid off once the jobsite is completed
Elon enters the tent with 4 Tesla bots slowly shambling behind him. One lags out when its remote link to a human controller is severed. Minutes later the remnants of a starlink satellite crash through the far end of the tent kitchen.
3 remaining Tesla bots proceed to bumble around like idiots, unable to open packets of deli ham, entirely ripping off the tops of deli mustard containers
Elon is awkwardly smiling and doing jazz hands the whole time
A neuralink mind controlled pig walks in as one Tesla bot wields a knife. Elon raises his hand to his ear, nods, then pushes a button on some phone app
The pig screams, then a popping noise is heard, and the pig collapses to the ground with smoke and blood coming out of its ears and nose
knife wielding tesla bot attempts to cut the pig’s flank, falls, cannot recover
the two remaining tesla bots continue in vain to open a loaf of bread without ripping the entire loaf apart. One slips and falls backwards, the other one runs out of battery and is frozen in place, holding a single piece of wonder bread
Elon curses, reaches into a refrigerator and hands you a crustable
Cut him a break, he’s pioneering sandwich making with this innovative tech. A few hiccups are to be expected.
Don’t say “cut him a break” around a Tesla product, god knows what command it’ll interpret that as.
Same
Just watched the video. Maybe it’s just me, but if you’re the driver in a Tesla, and it’s foggy as fuck outside, maybe, just maybe, don’t use the self driving aspect when visibility is that bad. The amount of people willing to trust Tesla with their lives (and others on the road…) is too damn high!
I’d put the blame on the branding, full self drive shluld mean full self drive, not most-conditions self drive without explicitly providing the limitations. Even irplane autopilot systems, which solve a simpler problem, have explicit limitations stated
There really needs to be legal pressure for them to change the name. I don’t see how it’s not false advertising.
More and more they are starting to look like someone just vacuum molded the body over a chassis rather than putting any sort of real artistry to it. It also makes me wonder if Ellen is saying the look is “sexy,” does he have some sort of spandex fetish. Maybe he stole his mum’s pantyhose when alone and strutted around the house wearing them, thinking of himself as such. Maybe he still does…
Did the owner not put their car in “Do not ram train” mode?
That’s an optional software upgrade. It’ll cost you $12.95 a month.
Oh come on, who wouldn’t pay for that? To not run into trains? That’s a bargain! Thanks Daddy Musk 🥰
And just think, the rest of us rubes have to manually not drive into trains, like barbarians
Why even live anymore.
drives into train
So sick of people referring to “Do not ram train” mode. You see it all over social media, but especially Lemmy. It’s “Do not ram train (Supervised)” mode, and you’d have to be living under a rock for the last 5+ years to think you don’t have to actually take control of the wheel to stop it from ramming a train.
deleted by creator
If self-driving A.I. models need a workforce to help identify trains, we could probably assemble an army of toddlers willing to be paid in cookies. My friend’s kid gets HYPE and yells “TRAIN!” when he sees one. He can also reliably identify cows. He calls most construction equipment “big truck” but that might be good enough. If a Tesla thinks a backhoe is a big truck, it’ll avoid it.
And that’s the biggest issue so far with “AI.” It’s about as intelligent as a newborn.
No, newborns rapidly take in new information and learn. “AI” is just a sophisticated text probability model. It doesn’t know anything. It isn’t learning how things work. It just regurgitates.
It’s like the difference between a student who understands the concepts versus memorizes the test answers.
Ohhh I like that one.
It’s about as intelligent as a newborn.
Newborns can’t even utter one cohesive word. I don’t get the point of making such an obviously false claims about anything.
A wise Jedi once said, “The ability to speak does not make you intelligent”
Intelligence is not binary, but a spectrum.
Please select all images containing TRAIN
Autistic kid dream job.
Tesla opted not to use LIDAR as part of its sensor package and instead relies on cameras which are not enough to determine accurate location data for other cars/trains etc.
This is what you get when billionaires cheap out on their products.
I never understood Musk’s reasoning for this decision. From my recollection it was basically “how do you decide who’s right when lidar and camera disagree?” And it felt so insane to say that the solution to conflicting data was not to figure out which is right but only to listen to one.
I mean, I think he’s a textbook example of why not to do drugs and why we need to eat the rich, but I can understand the logic here.
When you navigate a car as a human, you are using vision, not LIDAR. Outside of a few edge cases, you aren’t even using parallax to judge distances. Ergo, a LIDAR is not going to see the text on a sign, the reflective stripes on a truck, etc. And it gets confused differently than the eye, absorbed by different wavelengths, etc. And you can jam LIDAR if you want. Thus, if we were content to wait until the self-driving-car is actually safe before throwing it out into the world, we’d probably want the standard to be that it navigates as well as a human in all situations using only visual sensors.
Except, there’s some huge problems that the human visual cortex makes look real easy. Because “all situations” means “understanding that there’s a kid playing in the street from visual cues so I’m going to assume they are going to do something dumb” or “some guy put a warning sign on the road and it’s got really bad handwriting”
Thus, the real problem is that he’s not using LIDAR as harm reduction for a patently unsafe product, where the various failure modes of the LIDAR-equipped self-driving cars show that those aren’t safe either.
Also that LIDAR is more expensive then cameras, which means higher end user price, as far as I remember.
I wasn’t sure if he admitted that as being the reason (even though it obviously is)
I thought that was his main justification, idk tho, I don’t listen to the earnings calls or interviews myself lol
All about saving a buck.
I mean the decision was stupid from an engineering point of view, but the reasoning is not entirely off. Basically it follows the biological example: if humans can drive without Lidar and only using their eyes than this is proof that it is possible somehow. It’s only that the current computer vision and AI tech is way worse than humans. Elon chose to ignore this, basically arguing that it is merely a software problem for his developers to figure out. I guess in reality it is a bit more complex.
Not only that, but took out the radar, which while it has its own flaws, would have had no issue seeing the train through the fog. While they claimed it was because they had “solved vision” and didn’t need it anymore, it’s bullshit, and their engineering team knew it. They were in the middle of sourcing a new radar, but because of supply chain limitations (like everyone in 2021) with both their old and potential new supplier, they wouldn’t continue their “infinite growth” narrative and fElon wouldn’t get his insane pay package. They knew for a fact it would negatively affect performance significantly, but did it anyway so line could go up.
While no automotive company’s hands are particularly clean, the sheer level of willful negligence at Tesla is absolutely astonishing and have seen and heard so many stories about their shitty engineering practices that the only impressive thing is how relatively few people have died as a direct result of their lax attitude towards basic safety practices.
LIDAR would have similarly been degraded in the foggy conditions that this occurred in. Lasers are light too.
While I do think Tesla holds plenty of responsibility for their intentionally misleading branding in FSD, as well as cost saving measures to not include lidar and/or radar, this particular instance boils down to yet another shitty and irresponsible driver.
You should not be relying on FSD over train tracks. You should not be allowing FSD to be going faster than conditions allow. Dude was tearing down the road in thick fog, way faster than was safe for the conditions.
Maybe it shouldn’t be called full self driving if it’s not fully capable of self driving
A Tesla drover might get the impression that the cars “opinion” is better than their own, which could cause them to hesitate before intervening or to allow the car to drive in a way they are uncomfortable with.
The misinformation about the car reaches the level of negligence because even smart people are being duped by this.
Honestly I think some people just dont believe someone could lie so publicly and loudly and often, that it must be something else besides a grift.
One of the first things you learn to get your driver’s license is the Basic Speed Law, you must not drive faster than the driving conditions would allow. If only Full Self Driving followed the law and reduced its max speed based on the same.
If you were to strictly take that rule seriously, you should not allow FSD to drive at all, as at any speed its more dangerous than the person driving it (given an average driver who’s not intoxicated).
Well said.
AI said kill this guy right here
It noticed he didn’t like Elon’s tweet that day.
You know… I’d believe Elon is petty enough to actually put something like that in.
What? Every critic of tesla has been in a fata----
car bursts through wall
If he had time to notice it not slowing down he had time to brake and take it out of full self driving. I understand as someone who is sceptical about the fsd mode that I am more proactive at taking over than those who trust it a little bit more. I just feel if a company tells you to supervise it you should supervise it.
I still find fsd to be very finicky and vastly oversold
If something is sold as fully self driving, I would like to think it should be capable of fully self driving and not a feature that will drive me face first into a train.
I too come from a time where company’s had to sell functional products or go bankrupt, but alas those days are long gone.
To be fair, it could have fully driven itself into the train: “fully self driving” <> “fully safe driving” /s
Found the SQL developer!
Regardless of the naming, because everyone gets so stuck on fucking names and seems to ignore everything else because that makes for a quick comment with a ton of votes and feel good bullshit.
It is sold as a work in progress piece of software that is constantly being updated and still needs to be supervised. It has a ton of warnings about it’s capabilities, and lack thereof when activating it. There is no question when actually setting up FSD in the vehicle that it is something still in testing and not to be treated as a full replacement for paying attention. It constantly watches you and will warn you if you aren’t paying active attention for too long. If you ignore those warnings enough it will deactivate itself, forcing you to drive, and with enough deactivation will remove the capability entirely.
Image of the activation screen.
They’ve even updated the setting in the vehicle to be more specific, showing it as “Full Self-Driving (Supervised)”…
https://www.tesla.com/ownersmanual/modely/en_us/GUID-2CB60804-9CEA-4F4B-8B04-09B991368DC5.htmlAll of these reported situations are from people actively ignoring numerous safety and attention warnings, yet no one seems to ever put any blame on the driver in comments or articles. It’s always about blaming everything on Tesla when they’re actively telling every driver that it needs to be supervised because it will make mistakes.
You’re right that it’s just a name, which means it’s within Tesla’s power to not call it “full self driving”. Like maybe keep the word “full” for when it’s better than “full self driving brackets not really”.
The reason it’s called that is so when you’re buying the car, you can read “full self driving”, the salesman can call it “full self driving”, and then you can get excited and think you’re getting full self driving and pay stupid amounts of money for an iPhone on wheels.
It’s also so Musk can get up on stage and lie for years about how you’ll be able to go coast to coast while you sleep by the end of the year or whatever it is. Having a bunch of warnings in the software setup is not enough for someone gargling Musk’s jizz to cough it up and see it for the bullshit that it is.
You can give us all this extra info but you can’t change the core reality that the name is a lie.
The reason it’s called that is so when you’re buying the car, you can read “full self driving”, the salesman can call it “full self driving”, and then you can get excited and think you’re getting full self driving and pay stupid amounts of money for an iPhone on wheels.
Who exactly are you describing here? Like someone who’s been living under a rock for the past 5+ years and thinks cars just drive themselves now, but who also has $50k-$100k to drop on a car and also doesn’t do any research beforehand? I think you’re blowing this way out of proportion.
All these systems are flawed to some extent as nobody has cracked the code to L5 automation. There is some danger to it but there are many dangers to driving and this eliminates some of them. People have died in Teslas but many more have died in every other production vehicle that has ever existed. If this guy did the samr thing in a Toyota Camry, do you think we’d even be talking about it right now? These systems can only get better with a lot of real-world usage.
I totally get where the other person is coming from. These arguments are so tired and meaningless at this point. If you want Musk to go away, then stop bringing him up at every turn because it just makes people sound like bizzaro-world Musk fans as they hate the guy but can’t stop talking about him and following his every move. Some of us want to discuss and debate the actual technology and Musk had nothing to do with developing it.
Musk was involved in marketing and lying about it, though, and his extremely prolific public image is what gave it so much credibility. He’s lost a lot of that credibility now though, largely because people have spent a lot of time criticizing him. If you want him to disappear from the public eye that’s great, so do I, but he’s one of the most powerful men in the world, so that’s not going to happen.
Well why are you lending him further credibility by continuing to make him prolific? You’re just feeding into his goal of garnering more attention at every turn. Media outlets will continue to focus on all his insane ramblings because people like you are guaranteed to click on it.
but he’s one of the most powerful men in the world, so that’s not going to happen.
nobody has cracked the code to L5 automation.
especially if you remove LIDAR.
The systems with LIDAR aren’t L5 either. It’s impossible to claim what is and isn’t needed when nobody has actually come up with a solution.
Never said they were, but LIDAR is a big improvement over “cameras and AI”. Most other manufactures use it, Tesla stopped using it because of cost sav… because their AI kicks ass and if eyes are enough for humans they’re enough for their cars too.
Hell any system where lives are involved should be triple-redundancy.
So tired of the same arguments. They don’t mean anything in the real world. Complain all you want about what the shit is called, it makes no real world difference.
Legally, the driver is responsible for the fucking vehicle and these articles and comments like yours just give the impression you think they shouldn’t be responsible because of what it’s called. You’re giving shitty drivers a pass because they’re actively being stupid and you don’t like what Tesla names the software. That is the stupidest take in the world.
Go ahead in a court of law and claim you are not responsible for an accident that happens while FSD is activated and let’s see whether the name matters for your liability.
I notice you danced around the question of whether it was a lie to focus on bullshit legal stuff, which isn’t the arbiter of truth and reality. As I once heard a judge say, “You don’t come here for justice, you come here for a judgment.”
Anyway since you think it matters, the lawsuit you’re talking about is happening and a judge has ruled that the case has merit to continue: https://www.reuters.com/legal/tesla-must-face-vehicle-owners-lawsuit-over-self-driving-claims-2024-05-15/
Merit to continue just means it’s not clearly a bullshit lawsuit that should be thrown out to avoid wasting court time. Your linked lawsuit also is not about whether someone is legally responsible for a vehicle driving itself, it is again about the marketing.
I don’t give a shit whether their marketing is a lie. Marketing is not the issue at hand, as much as you all want it to be for whatever reason instead of actually blaming the shitty drivers. It has no bearing on whether someone is responsible for the vehicle they are in the driver seat of hitting something or someone.
Why do you not want to put any blame on these drivers? Drivers that ignore the warning when they turned on the function in the first place and that warns them every time they turn it on to still pay attention. Why are you so insistent that the blame should be on Tesla because of what they call it?
Makes me start to think you’re that type of driver and trying to justify your belief that you shouldn’t be responsible for the actions of a 2 ton murder machine traveling at high speed that you are in control of.
I don’t give a shit whether their marketing is a lie.
I mean clearly. “Full self-driving” is marketing, and it is a lie. That’s the point being made here. You don’t have to care about it, but that doesn’t mean it’s not a lie.
I’m not not blaming the drivers. They were foolish enough to buy a Tesla and trust it with their lives for a start. But I am also blaming the marketing. Two things can be true.
So tired of the same arguments. They don’t mean anything in the real world
The second sentence is a fallacy.
And it does matter that the company is calling it full self driving while it doesn’t fully self drive. That it would have that capability is something Musk has promised for many years. It’s also a reason that Tesla stock is worth so much.
Go ahead in a court of law and claim you are not responsible for an accident
That’s would be a very specific case. Tesla has been reminded multiple times that they need to take into account how people use their vehicles. The company is also under investigation for possible fraud because they are selling something that doesn’t do what people would respect it to do.
You’re focusing on one thing, but there’s multiple ways that the company could be liable. There’s been multiple articles explaining that the company is either under investigation or that the company has been warned to change things or else.
Well put. It’s a funny thing, words have meaning and if you advertise your products with those words, some portion of the population (gasp!) might believe you!
How does it not fully self drive? What’s your definition of full self driving then?
Mercedes Drive Pilot is Level 3 and even it will prompt you to take over when necessary, does it not fully self drive then either? What about Waymo/Cruze? They have remote operators controlling the vehicles when they get stuck. Not fully self driving either? Is the standard that it needs to be absolutely flawless and never fail or what is it?
Hey everyone, Elon’s come to lemmy!
They don’t mean anything in the real world.
Uh. They mean everything in the real world? You get sued for false advertising and fraud. Fox News got sued heavily for knowingly lying about voting machines. There’s a reason that companies have PR departments. Words matter a lot in the real world.
Are the drivers stupid? Sure. They believed the FSD claim after all. But that doesn’t mean Tesla is off the hook. Deceiving stupid people is still deceit.
Well said and thanks for posting the examples. It’s something that bothers me about any social media kind of site. Especially here on Lemmy. Nobody gives a damn about the incredible amount of negligence the drivers must have. It immediately becomes an anti Elon circlejerk every time.
It’s similar with news articles, which this post doesn’t even link to, most of the articles name drop Tesla or Elon just because otherwise it’s not a story. “Somebody hit a car / person / train because they weren’t paying attention to the road” isn’t story worthy. But as soon as doubt can be cast on an Elon company, it become a must post thing. I can’t stand Musks antics either, but he gets too much free rent in peoples mind. It’s wild
/rant
At the same time, there’s too many people who say that Full Self Driving obviously doesn’t mean that the vehicle still fully drive itself. Though for unknown reasons it is totally fine to keep using the name Full Self Driving.
And that is between Tesla and the NTSB to sort even though I agree. The car itself doesn’t mince words describing it to you, and at the time you’re driving it, the required supervision is unambiguous.
If it were called “Tesla Drive” or something else, everyone would still be here taking a shit on it nonetheless.
It’s called Full Self Driving (Supervised) nowdays. They changed the name.
The vehicle is capable of driving you to the grocery store on the other side of the city and back, sometimes with zero interventions from the driver. If that’s not Full Self Driving then I don’t know what is.
It’s like how they changed the cybertruck windows to Unbreakable Glass (Fragile).
I can’t find any source for that claim
That’s Full Self Driving (Sometimes).
Then what is full self driving to you? How good does the system need to be to qualify?
Even this “article” is about nothing happening. The driver was paying attention and took over when the vehicle was about to do something it should. Just as they should.
Also, even if FSD was 10x safer than a human driver and we replaced every single car on the roads with Teslas there would still be 8 people dying every single day in the US alone. Linking articles about these accidents does not prove it being unsafe. It only feeds the confirmation bias of the person posting it and the people upvoting it. People want it to be unsafe so that they can shit on Elon. The standards they apply to Tesla are ridiculous compared to that of other companies. The extremely limited Mercedes Drive Pilot is praised as revolutionary tech while FSD already checks most boxes for Level 4 self-driving.
Just some insight from my pov. Fsd is marketed as FSD (Supervised). I don’t agree with the jamming but it is what it is. I know it does janky stuff, it still forces you to pay attention. Do I believe this could happen, yes but do I doubt the driver always until proven otherwise.
I have had my model y yell at me to take control when I was already out of any auto/fsd mode. I have many downs and many ups. I agree that the car should actively steer you into the train. I was curious if anyone had a link to the dash footage or even to an article with it.
Fsd is marketed as FSD (Supervised)
it is, now, it was not marketed with any kind of parenthetical qualifier until recently.
Actually, it said Full Self Driving (BETA) until it was updated to (Supervised) recently.
If anything, the beta qualifier is actually better than just saying supervised since that term means not complete and still being developed.
never heard musk refer to it as anything but full self drive, no qualifiers.
companies concerned with safety wouldn’t market the shit until it’s safe (see mercedes apparent lead).
Mercedes Drive Pilot is hilariously limited system. It for example needs a car in front of it that it can follow or else it wont work. It also only works on limited number of hand-picked highways in California and Nevada.
There’s a video on YouTube comparing FSD to Mercedes’ equivalent driver assistant software (not the level 3 one) and it’s not even a competition. The Mercedes system is completely unusable.
Those should be your expectations when you are on the shop floor and that should allow you to reject the purchase if it’s s deal breaker for you. Not when you’re crossing railway tracks.
Weird how this notion of “personal responsibility” applies to every person except for those people who choose to intentionally misrepresenting the product by branding it in ways that are misleading. The people running this company aren’t responsible for their role in misleading the public, just because the fine print happens to indicate that the product isn’t actually what it’s marketed as?
Now you’ll probably say something to the effect of “I never said that! You’re putting words in my mouth!” except what other motivation can you have to jump to the defense of the liar and blame people for being misled, except that you want to put all the responsibility on individuals for being misled and not on the company that is systematically and intentionally misleading them? Maybe you just manage to derive a smug sense of superiority thinking of yourself as someone who is invulnerable to this kind of tactic so blaming the victims lets you feel good about yourself.
You literally cannot buy FSD without being told that it needs driver supervision. It also tells you that every single time you enable it and it’s constantly nagging to you when you take your hands off the wheel aswell as if you’re looking at your phone etc. and given enough warnings the system locks you out of it.
Has Musk been dishonest/misleading about it’s capabilities in the past? Yes. Is there a single Tesla owner with FSD who doesn’t know the truth? No.
I’m sure you’re on the shop floor for every one of those conversations.
But anyway, enjoy being confidently incorrect: https://www.consumerreports.org/cars/car-safety/tesla-driver-monitoring-fails-to-keep-driver-focus-on-road-a3964813328/
“Confidently incorrect”
Then proceeds to link over 2 year old article and even that aknowledges the existence of such system in the title.
It has an indoor camera that is constantly monitoring the driver and nags when they’re not paying attention. That’s a fact. Nothing what I said has been proven incorrect.
But anyway, enjoy being confidently incorrect
👆 This is what we call irony
full self driving
I just take serious issue with this label. It’s not fully self driving, it requires the user’s attention. the accidents and years of promises broken are just cherries on the shit sundae.
Elon here. Thanks for fighting the good fight. We will deposit 100 X bucks in your account.
I hear X bucks can be used in place of the three seashells. Does anyone know if that’s true?
Lol 😂
If he had time to notice it not slowing down he had time to brake and take it out of full self driving.
And that’s the way he survived. But wrecked that plastic box.
The good news is that we can finally see the light at the end of the tunnel…
“It’s just a freight train coming your waaaaayyyyyyy!” -metalilica
Seriously, you just had to throw that pun in there. LOL.
Every couple of months there’s a new story like this. And yet we’re supposed to believe this system is ready for use…
It is ready because Musk needs it to be ready. Watch out, this comment may bring the morale down, and Elron will be forced to … Cry like a baby 😆
Didn’t he recently claim Tesla robotaxi is only months away?
Well I suppose he didn’t say how many months, but the implication was less than a year, which has been his claim every year since 2016.He said that Teslas were an investment worth hundreds of thousands of dollars because owners would be able to use them as robot taxis when they weren’t using their car and charge a small fee by next year…in 2019. Back then he promised 1 million robot taxis nationwide in under a year. Recently he gave the date august 8 to reveal a new model of robot taxi. So, by Cybertruck estimates, I would say a Tesla robot taxi is a possibility by late 2030.
He is just spewing shit to keep the stock price afloat, as usual.
He also said they were ready to manufacture the 2nd generation Tesla Roaster “now,” which was back in 2014. No points for guessing that as of yet (despite taking in millions of dollars in preorders) they have not produced a single one.
Given this very early and still quite relevant warning, I’m astounded that anyone is dumb enough to believe any promise Elon makes about anything.
In what way is it not ready to use? Does cars have some other driver assistant features that are fool proof? You’re not supposed to blindly trust any of those. Why would FSD be an exception? The standards people are aplying to it are quite unreasonable.
Because it’s called Full Self Drive and Musk has said it will be able to drive without user intervention?
The naming is poor, but in no way does the car represent to you that no intervention is required. It also constantly asks you for input and even watches your eyes to make sure you pay attention.
The car maybe not, but the marketing sure does
Marketing besides the naming we have already established and Elon himself masturbating to it? Is there some other marketing that pushes this narrative, because I certainly have not seen it.
It’s called Full Self Driving (Supervised)
Yeah, it will be able to drive without driver intervention eventually. Atleast that’s their goal. Right now however, it’s level 2 and no-one is claiming otherwise.
In what way is it not ready to use?
Full Self Driving (sike!)
It’s unreasonable for FSD to see a train? … that’s 20ft tall and a mile long? Am I understanding you correctly?
Foolproof would be great, but I think most people would set the bar at least as high as not getting killed by a train.
Did you watch the video? It was insanely foggy there. It makes no difference how big the obstacle is if you can’t even see 50 meters ahead of you.
Also, the car did see the train. It just clearly didn’t understand what it was and how to react to it. That’s why the car has a driver who does. I’m sure this exact edge case will be added to the training data so that this doesn’t happen again. Stuff like this takes ages to iron out. FSD is not a finished product. It’s under development and receives constant updates and keeps improving. That’s why it’s classified as level 2 and not level 5.
Yes. It’s unreasonable to expect brand new technology to be able to deal with every possible scenario that a car can encounter on traffic. Just because the concept of train in a fog makes sense to you as a human doesn’t mean it’s obvious to the AI.
In what way is it not ready to use?
To me it seems you just spent three paragraphs answering your own question.
can’t even see 50 meters ahead
didn’t understand what it was and how to react to it
FSD is not a finished product. It’s under development
doesn’t mean it’s obvious to the AI
If I couldn’t trust a system not to drive into a train, I don’t feel like I would trust it to do even the most common tasks. I would drive the car like a fully attentive human and not delude myself into thinking the car is driving me with “FSD.”
You can’t see 50 meters ahead in that fog.
Completely true. And I would dictate my driving characteristics based on that fact.
I would drive at a speed and in a manner that would allow me to not almost crash into things. But especially trains.
I agree. In fact I’m surprised the vehicle even lets you enable FSD in that kind of poor visibility and based on the video it seemed to be going quite fast aswell.
LIDAR can
Yeah there’s a wide range of ways to map the surroundings. Road infrastructure, however is designed for vision so I don’t see why just cameras wouldn’t be sufficient. The issue here is not that it’s didn’t see the train - it’s on video, after all - but that it didn’t know how to react to it.
No, the standards people are applying to it are the bare minimum for a full self driving system like what musk claims.
It’s a level 2 self driving system which by definition requires driver supervision. It’s even stated in the name. What are the standards it doesn’t meet?
Of whiat words is FSD an acronym?
You’re not supposed to blindly trust any of those. Why would FSD be an exception?
Because that’s how Elon (and by extension Tesla) market it. Full self driving. If they’re saying I can blindly trust their product, then I expect it to be safe to blindly trust it.
And if the fine print says I can’t blindly trust it, they need to be sued or put under legal pressure to change the term, because it’s incredibly misleading.
Full Self Driving (Beta), nowdays Full Self Driving (Supervised)
Which of those names invokes trust to put your life in it’s hands?
It’s not in fine print. It’s told to you when you purchase FSD and the vehicle reminds you of it every single time you enable the system. If you’re looking at your phone it starts nagging at you eventually locking you out of the feature. Why would they put driver monitoring system in place if you’re supposed to put blind faith into it?
That is such an old, beat up strawman argument. Yes, Elon has said it would be fully autonomous in a year or so which turned out to be a lie but nobody today is claiming it can be blindly trusted. That simply just is not true.
It isn’t Full Self Driving if it is supervised.
It’s especially not Full Self Driving if it asks you to intervene.
It is false advertisement at best, deadly at worst.
It’s misleading advertising for sure. At no point have I claimed otherwise.
The meaning of what qualifies as “full self driving” is still up for debate however. There are worse human drivers on the roads than what the current version of FSD is capable of. It’s by no means flawless but it’s much better than most people even realize. It’s a vehicle capable of self driving even if not fully.
Unfortunately, companies also have to make their products safe for idiots. If the system is in beta or must be supervised, there should be inherently safe design that prevents situations like this from happening even if an idiot is at the wheel.
ESP is not idiot proof either just to name one such feature that’s been available for decades. It assists the driver but doesn’t replace them.
Hell, cars themselves are not idiot proof.
Hell, cars themselves are not idiot proof.
Yup, almost always, there’s an idiot in the driver seat.
Yeah and cars should have a system to stop idiots doing dumb things, best we have is a license so if it’s good enough for cars without added safety features is good enough for them with
Ever couple of months you hear about every issue like this, just like you hear about every airline malfunction. It ignores the base rate of accurate performances which is very high.
FSD is imperfect but still probably more ready for use than a substantial fraction of human drivers.
This isn’t actually true. The Tesla full self driving issues we hear about in the news are the ones that result in fatal and near fatal accidents, but the forums are chock full of reports from owners of the thing malfunctioning on a regular basis.
It IS actually true. It does goofy stuff in some situations, but on the whole is a little better than your typical relatively inexperienced driver. It gets it wrong about when to be assertive and when to wait sometimes, it thinks there’s enough space for a courteous merge but there isn’t (it does some Chicago style merges sometimes), it follows the lines on the road like they are gospel, and doesn’t always properly estimate how to come to a smooth and comfortable stop. These are annoying things, but not outrageous provided you are paying attention like you’re obliged to do.
I have it, I use it, and I make lots of reports to Tesla. It is way better than it used to be and still has plenty of room to improve, but a Tesla can’t reboot without having a disparaging article written about it.
Also fuck elon, because I don’t think it gets said enough.
typical relatively inexperienced driver
Look at the rates that teenagers crash, this is an indictment.
provided you are paying attention
It was advertised as fully autonomous dude. People wouldn’t have this much of a hard-on for trashing it if it wasn’t so oversold.
This fully autonomous argument is beat to death already. Every single Tesla owner knows you’re supposed to pay attention and be ready to take over when necessary. That is such a strawman argument. Nobody blames the car when automatic braking fails to see the car infront of it. It might save your ass if you’re distracted but ultimately it’s always the driver whose responsible. FSD is no different.
If it’s not fully capable of self driving then maybe they shouldn’t call it full self driving
Sure. Make then change the name to something different. I’m fine with that.
Though I still don’t know what most people actually mean by full self driving and how it’s different from what FSD can do right now.
You realize it can be true that the driver is at fault when they crash and that the crash was more likely to happen because you have Elon contradicting his own marketing team constantly and confusing people.
He literally would take reporters in his car and take his hands off the wheel. He just fundamentally doesn’t care about safety now. Probably doesn’t about safety later, just saw a way to make some money.
Seriously you sound like a Mac user in the '90s. “It only crashes 8 or 9 times a day, it’s so much better than it used to be. It’s got so many great features that I’m willing to deal with a little inconvenience…” Difference being that when a Mac crashes it just loses some data and has to reboot but when a Tesla crashes people die.
Seriously you sound like a
Mac user in the '90sLinux user today.FTFY
These are serious rate differences man.
Every driver, and even Tesla, will tell you it’s a work in progress, and you’d be hard pressed to find someone who has had an accident with it. I’d be willing to bet money that IF You find someone who has had an accident they have a driving record that’s shitty without it too.
If you want to talk stats, let’s talk stats, but “It seems like Tesla is in the news a lot for near crashes” is a pretty weak metric, even from your armchair.
Is 200ish crashes and 6 deaths per year too many?
I know its an absolute number but we are asking if its worth sacrificing people for the potential of safer driving later.
Can you explain why you are so confident that this will all be worth it in the end?
Evidence that teslas are more dangerous than other cars: https://www.thedrive.com/news/tesla-drivers-have-the-highest-crash-rate-of-any-brand-study
Evidence for the 200 crashes and 6 deaths a year claim for FSD: https://www.theverge.com/2024/4/26/24141361/tesla-autopilot-fsd-nhtsa-investigation-report-crash-death
You hear so much about the people Jeffrey Dahmer murdered, but never anything about all the people he didn’t murder!
Cute.
Here’s some actual information
People are terrible at probability estimation, and even with two fatal accidents a month FSD is likely still safer than most of the people on the road per million miles driven.
I see you’ve decided to be condescending, and also made a falsifiable claim. This is the part where you bring some actual data or STFU.
Whatever you say Mr Dahmer joke instead of content. I see that was really all in good faith and maybe I unintentionally hurt your feelings by citing a source on base rate biases?
What data would you like me to bring for discussion since you’ve been so open thus far? Do you want me to bring some data showing that teslas spend more time not having accidents than having accidents? I’m happy to go do some homework to enrich this interaction.
It’s not as though you can just ask Tesla for every case of an FSD crash. The falsifiable claim is just me tossing a number, the point is that memorable bad press and bad stats are not the same.
Why would you think there isnt data on Tesla crashes? Are they hiding their broken cars from bystanders and police or something now?
When you look at the development notes on the self-driving at Tesla, anyone with a brain wouldn’t trust that shit, not even a little bit. Most of what they did is placate Musk’s petty whims and delusions. Any real R&D issues were basically glazed over it given quick software fixes.
Are there notes available to read?
Here’s one of the few articles that wasn’t paywalled when I pulled up a Google search on it. It’s really not hard to find the stories are all over the place.
https://www.businessinsider.com/elon-musk-tesla-autopilot-fsd-engineers-unsettled-2021-8
Demonstrate what you mean because it really sounds like you’re describing what you feel should be true to justify your emotions about the bad Twitter man.
And to be clear, I mean link the documents you claim to have read and the parts of them you claim demonstrate this.
Just need to Google “Tesla self-driving development engineers and Elon Musk”, and you’ll find lots of articles. Here’s one of the few that wasn’t paywalled.
https://www.businessinsider.com/elon-musk-tesla-autopilot-fsd-engineers-unsettled-2021-8
This is an article from 2021 about a book researched in 2019.
Yeah, during development of the Tesla self driving system.
Read the development notes from the first years of any technology you use. The research you’re “referencing” is six years old at this point.
What’s next? You going to criticize an iPod Nano to make a point about the broken screen on your iPhone 8? Criticize Google assistant from 2019 to harangue OpenAI?
Look at what six years of development means: https://youtu.be/qTDlRLeDxxM?si=dFZzLcO_a8wfy2QS
It’s not about how well or badly it worked when they were developing it, it’s about the developing process. It’s about the fact that they had to appease Elon Musk’s ego in all aspects of developing the self drive system. To a disastrous degree.
And again there is a world of difference between the iPhone or Open AI or Google Assistant not working right and a car driving itself not working right because when those other things don’t work nobody dies or gets hurt. But a car can mame and kill people extremely easily.
That’s a very old article about even older opinions, now totally outdated as shown by statements like;
Almost five years on, Tesla still hasn’t completed its autonomous road trip — no car company has even come close.
You’re using unsubstantiated statements from the start of development which is totally different to what you claimed before being asked for a source.
Current development FSD has hit huge milestones which competitors have not.
He fortunately avoided the train, but unfortunately still owns a Tesla.
What a bunch of morons people were in 1912 to believe a ship could be unsinkable. Amirite guys?
The Titanic probably wouldn’t have sunk if it hit the iceberg head on. Clearly the Tesla simply mistook the train for an iceberg and itself for an ocean-liner and opted for a more ideal collision. The driver should have disabled ‘sea mode’ if they didn’t want that behavior, it’s all clearly spelled out in the owners manual.
As a frequent train passenger, I’m not overly concerned.
Seems a bit too weak to derail, probably only delay.
Feels like these things were more capable a decade ago when they had radar.
Not that they should be called “full self driving” either then or now, but at least radar can deal fog better than regular ass cameras
Those trains sure are weird and confusing, with their back and forth those tracks and all. Makes you wonder about train safety, it does!
// added a fun link
*a drunken human
I don’t want to disagree, but I would like a source to support this claim
That exclamation point in the comment you replied to should be your hint that it’s sarcasm.
No, Musk said this at one point at some press conference
Not the person I was replying to but okay… what has that got to do with the sarcasm of the comment I was referring to?
Look @ original comment, Musk has stated many a time that FSD is safer than human drivers and I’m pretty sure at one point he said it was 10x safer… No sarcasm in that statement.
Oh here
Musk has multiple times stated that FSD is safer than han driving. I’m not gonna bother finding the bids as I’m at work
Here ya go someone else did it for me