Tesla 'Self-Driving' NDA Hopes To Hide The Reality Of An Unfinished Product
from the I'm-sorry-Dave-I-can't-do-that dept
There isn't a day that goes by where Tesla hasn't found itself in the news for all the wrong reasons. Like last week, when Texas police sued Tesla because one of the company's vehicles going 70 miles per hour in self-driving mode failed to function properly, injuring five officers.
Five Montgomery County deputy constables were injured Saturday when the driver of a Tesla rear-ended a cruiser during a traffic stop, causing a chain-reaction crash, authorities said. https://t.co/FfteMQQ4zL
— Pooja Lodhia (@PoojaOnTV) February 27, 2021
If you hadn't been paying attention, Teslas in self-driving mode crashing into emergency vehicles is kind of a thing that happens more than it should. In this latest episode of "let's test unfinished products on public streets," the Tesla vehicle in "self-driving" mode's systems failed completely to detect not only the five officers, but their dog, according to the lawsuit filed against Tesla:
“The Tesla was completely unable to detect the existence of at least four vehicles, six people and a German Shepherd fully stopped in the lane of traffic,” reads the suit. “The Tahoes were declared a total loss. The police officers and the civilian were taken to the hospital, and Canine Officer Kodiak had to visit the vet."
Of course for Musk fans, a persecution complex is required for club membership, resulting in the belief that this is all one elaborate plot to ruin their good time. That belief structure extends to Musk himself, who can't fathom that public criticism and media scrutiny in the wake of repeated self-driving scandals is his own fault. It's also extended to the NDAs the company apparently forces Tesla owners to sign if they want to be included in the Early Access Program (EAP), a community of Tesla fans the company selects to beta test the company's unfinished self-driving (technically "Level 2" driver-assistance system) on public city streets.
The NDA frames the press and transparency as enemies, and urges participants not to share any content online that could make the company look bad, even if it's, you know, true:
"This NDA, the language of which Motherboard confirmed with multiple beta testers, specifically prohibits EAP members from speaking to the media or giving test rides to the media. It also says: "Do remember that there are a lot of people that want Tesla to fail; Don't let them mischaracterize your feedback and media posts." It also encourages EAP members to "share on social media responsibly and selectively...consider sharing fewer videos, and only the ones that you think are interesting or worthy of being shared."
Here's the thing: you don't need to worry about this kind of stuff if you're fielding a quality, finished product. And contrary to what Musk fans think, people concerned about letting fanboys test 5,000 pound automated robots that clearly don't work very well are coming from a valid place of concern. Clips like this one, for example, which show the Tesla self-driving system failing to perform basic navigational functions while in self-driving mode, aren't part of some elaborate conspiracy to make Tesla self-driving look bad and dangerous. There's plenty of evidence now clearly showing that Tesla self-driving, at least in its current incarnation, often is bad and dangerous:
Not sure why FSD is such a safety hazard - especially for pedestrians and cyclists?
Check out this video, posted last week. pic.twitter.com/Hg0tGfxXDT
— David Zipper (@DavidZipper) September 19, 2021
Ever since the 2018 Uber fatality in Arizona (which revealed the company had few if any meaningful safety protocols in place) it's been clear that current "self-driving" technology is extremely undercooked. It's also become increasingly clear that widely testing it on public streets (where other human beings have not consented to being used as Guinea pigs) is not a great idea. Especially if you're going to replace trained testers with criticism-averse fanboys you've carefully selected in the hopes they'll showcase only the most positive aspects of your products.
We've been so bedazzled by purported innovation we've buried common sense deep in the back yard. Wanting products to work, and executives to behave ethically, is not some grand conspiracy. It's a reasonable reaction to the reckless public testing of an unfinished, over-marketed product on public streets.
Thank you for reading this Techdirt post. With so many things competing for everyone’s attention these days, we really appreciate you giving us your time. We work hard every day to put quality content out there for our community.
Techdirt is one of the few remaining truly independent media outlets. We do not have a giant corporation behind us, and we rely heavily on our community to support us, in an age when advertisers are increasingly uninterested in sponsoring small, independent sites — especially a site like ours that is unwilling to pull punches in its reporting and analysis.
While other websites have resorted to paywalls, registration requirements, and increasingly annoying/intrusive advertising, we have always kept Techdirt open and available to anyone. But in order to continue doing so, we need your support. We offer a variety of ways for our readers to support us, from direct donations to special subscriptions and cool merchandise — and every little bit helps. Thank you.
–The Techdirt Team
Filed Under: cars, nda, self-driving, transparency
Companies: tesla
Reader Comments
Subscribe: RSS
View by: Time | Thread
An NDA that encourages you to share good information but strongly discourages sharing anything negative, yeah that's not an NDA so much as them paying to become PR shills for the company.
[ link to this | view in chronology ]
Re:
It’s not much different than the NDA for most beta level testing.
[ link to this | view in chronology ]
Re: Re:
"It’s not much different than the NDA for most beta level testing."
...where unethical companies are concerned, yes. That both these statements seem correct just tells us the distance between the bar of acceptability and Situation Normal.
[ link to this | view in chronology ]
Nothing really to worry about.
They haven't even started scoping out the turret controls. And you would think the engineers would not be testing impact assault techniques until they developed at least placeholder armor.
So far, I'm unimpressed by the Bolo mark 0.31.
[ link to this | view in chronology ]
Something something the hold my beer race again is unable to learn from its history.
Someone wanna mail Elon a copy of 'Unsafe at Any Speed'?
NDA might keep drivers from talking, up until they find out that they are being personally held responsible for harming/killing people.
Pretty sure auto insurance companies aren't listing shitty AI as an authorized driver, which will leave the fanbois in the very very poor house.
If only we had some sort of agency charged with protecting the public on the roads who could do something about this.
But obviously this isn't as important as the liberal conspiracy to silence conservatives online, so its not important enough to pursue.
I mean they already killed 700K people who didn't need to die, how many can Tesla's manage to kill until they get it fixed?
[ link to this | view in chronology ]
Re:
Oh for the insurance companies it's a godsend.
"Oh, you were letting the AI drive? Thanks for admitting negligence!"
[ link to this | view in chronology ]
Re: Re:
"Thanks for admitting negligence!" <-- perhaps there was a "to" missing there.
In the race car, the first thing we do after starting is turn of TCS, ABS, and other "helpful computer bits" that mess with track driving.
In the street cars we turn off TCS when four-wheeling or going through deep areas of water or loose sand.
Negligence (failure to use reasonable care) is not "letting the AI drive." It would be failing to consider whether the AI should drive or not.
Sorry, your weird assertion that using FSD is negligent... fails.
E
[ link to this | view in chronology ]
Re: Negligence is not "letting the AI drive"
An ABS system is not a self driving system. A mechanical system that distributes power or manages breaking does not decide when to brake, or when to turn. The AI "drives" these cars, it doesn't shuffle power from one wheel to another.
You don't know what you're talking about.
[ link to this | view in chronology ]
Re: Re:
You turned control of the vehicle over to a "beta" AI, that there are a few known issues with... after you signed an NDA that you wouldn't/couldn't tell anyone outside of Tesla if the AI decided to plow through a kindergarten playground???
Thats a nice Tesla you have... pity it has to sit in your backyard since no insurance company will cover it.
And even if they improve the AI, its going to be a while before any of them will want to touch them.
It ran down 6 people, vehicles, and a dog...
In the other clip the AI decided to pursue humans trying to legally cross the street.
Given the number of assholes doing the sleep thing while the car drives on the freeway, you can't convince me everyone trying out the full self driving will actually be paying full attention even knowing that other drivers have been screwed over it, they will assume they'll be okay (because it won't happen to them).
[ link to this | view in chronology ]
Your right to talk
Tesla made a deal: we'll give you FSD to beta... you don't talk about it publicly.
It's a fair deal.
FSD isn't there yet. We all know this. What's the big deal when the beta testers want to say "Hi this beta-test isn't working right?"
None. Don't sign it, or if you do, live by the contract you signed.
Hitting 5 cops -- definite no no. But was it FSD or a DUI in progress. The cops hate releasing any exculptory information ever, so we won't find out until it hits a courthouse...
E
[ link to this | view in chronology ]
Re: Your right to talk
It's not just 5 cops(one was under a vehicle) but 5 police vehicles. That's a pretty big blunder for any collision.
[ link to this | view in chronology ]
Re: Your right to talk
NDA does not apply when a government agency requests or demands information. Plow through 5 cop cars and wave an NDA flag and see how far that gets you.
[ link to this | view in chronology ]
It seems. The self driving software does not understand police flashing lights warning signs etc or police stop signs, it simply drives into police cars parked on a road there's no way the self driving software is ready for 1000s of new drivers to use it on public roads it just looks for other cars moving or maybe pedestrians crossing it has no understanding of a police checkpoint or emergency vechicles parked in the middle of a road the flashing lights maybe make it harder to
determine is that a car or just a traffic light on the road
[ link to this | view in chronology ]
Re:
It seems. The self driving software does not understand police flashing lights warning signs etc or police stop signs, [...]
It doesn't seem to understand object detected.
[ link to this | view in chronology ]
Re: Re:
There's a reason that most self driving vehicles ignore stopped objects. Any road side debris would cause the vehicle to stop, hence they program the vehicles to ignore them.
[ link to this | view in chronology ]
Re: Re: Re:
What does that even mean? "Not running into objects" is the single most important aspect of safe driving. Even without any knowledge of traffic laws or conventions, you could at least survive most situations by just obeying that rule.
That anyone would program vehicles to outright ignore objects in their path is batshit insane.
[ link to this | view in chronology ]
Re: Re: Re:
When cars "ignore" stopped objects, it's generally because the forward facing radar tells the system that there's nothing there. Radar is great for judging how fast objects are moving or differences in velocity, but when an object completely stops it blends into the background. Systems that rely too much on radar, trusting it over conflicting sensor inputs, are prone to these types of crash. They were designed to follow highway traffic at speed where, generally speaking, there aren't parked cars in the road.
[ link to this | view in chronology ]
Re: Re: Re: Re:
Oddly enough, though, my Ford Escape with cameras and radar can and does see non-moving objects in the roadway just fine and slows/stops accordingly. It isn't advertised or referred to as Full Self Driving but simply an "assist", which it is. Yet in thousands of miles of using it to drive across the country several times, it hasn't hit a stopped vehicle or a pedestrian, or even come close to doing so. Funny that.
[ link to this | view in chronology ]
Re: Re: Re: Re: Re:
Right, the system works reasonably well and doesn't always ignore all immobile objects. That's why they're generally beneficial. But it has the same kinds of warnings Tesla uses:
Ford Manual
Tesla Manual
[ link to this | view in chronology ]
Re: Re:
That's what they get for not using object-oriented programming.
[ link to this | view in chronology ]
Re: Re: Re:
OOPs
[ link to this | view in chronology ]
Re:
The AI deputized itself and did what cops do when they see a dog.
[ link to this | view in chronology ]
The ai should be programmed to stop or to slowly drive past any parked cars while avoiding any pedestrians eg do not drive into any cars parked in the way where ever they may be . It could also sound a horn to warn any people in front of its not sure how to proceed like reversing trucks have audio alarms to warn people of possible collisions
[ link to this | view in chronology ]
Someone Has To
So yes I one a Tesla, 2 in fact. However, someone has to lead and pointer this field. Tesla takes all the heat for self driving because they are so far out in front, spotlight is on them. Not because they have a bad product. If it GM or Mercedes leading this FSD, it would be them under the microscope. It's. double edged sword.
[ link to this | view in chronology ]
... cause five car pileups?
Tesla takes all the heat for self driving because they are so far out in front, spotlight is on them.
Well that and their cars seem to have a distressing tendency to plow into other vehicles more often that they should(that number being 'never'), but I'm sure you're right and it's just their super-duper advanced tech that's causing the criticism.
[ link to this | view in chronology ]
Re: ... cause five car pileups?
That is down to the human monitoring not taking action when the autopilot is not doing its job. Failure to detect stationary objects is a well known problem with the software, and a reason human monitoring is required.
[ link to this | view in chronology ]
Someone Has To
So yes I own a Tesla, 2 in fact. However, someone has to lead and pioneer this field. Tesla takes all the heat for self driving because they are so far out in front, spotlight is on them. Not because they have a bad product. If it GM or Mercedes were leading this FSD, it would be them under the microscope. It's. double edged sword.
[ link to this | view in chronology ]
Bovine excrement
Tesla takes the heat because Tesla calls its SAE level 2 driver assist "Full Self Driving" while GM calls its Level 2 system Blue Cruise, because level 2 is basically a fancy cruise control, while actual self driving is SAE level 5. which is several orders of magnitude beyond anything Tesla has released.
[ link to this | view in chronology ]
Re: Bovine excrement
Minor correction, GO'S system is Super Cruise, Blue Cruise is Ford's Level 2 offering, note that neither company so much as whisper "self driving".
[ link to this | view in chronology ]
Re: Bovine excrement
While you were out , you missed out on the day in school during math class when it was explained what "orders of magnitude" meant.
5 is not "orders of magnitude beyond" anything Tesla has released.
Google search is that way --> GO educate yourself and the people you are mis-educating.
[ link to this | view in chronology ]
Re: Re: Bovine excrement
Level is an arbitrary label, especially when applied to software concepts, and orders of magnitude is a good metaphor for how software complexity, and effort required to solve software problems scale.
[ link to this | view in chronology ]
Though the crash that injured the five police officers wasn't anything to do with Self-Driving.
The linked article states the car was in Autopilot mode which is the name for their cruise control system that explicitly warns drivers when enabling it it doesn't detect stationary objects.
[ link to this | view in chronology ]
Re:
Yes there's valid criticism to be found, though it's generally a good idea to understand the systems you're criticizing. This article is written with such a chip on its shoulder that it doesn't bother to distinguish between the limited access "FSD Beta" it wants to blame vs the standard "Autopilot" which was active during this crash. You also didn't seem to notice that they're also suing the bar which served alcohol to the driver, so he was drunk and not driving responsibly. But drunk drivers hitting things is so common it barely qualifies as news any more.
[ link to this | view in chronology ]
Re: Re:
Agreed. Karl's coverage of the incident is far from fair, it reads more like a hit-piece that conflates things to come a conclusion that's not supported by the facts.
There is no doubt that Tesla have some problems with the self-driving which can result it fatalities, but most of the cases so far have shown us that some drivers are shockingly negligent and/or stupid. Sadly, you can't cure stupid with even the most advanced AI - even if it happens to be a beta-version.
And when Karl wrote: "Of course for Musk fans, a persecution complex is required for club membership", I read that as "I don't like people who don't agree with my views on Musk, therefore I'll denigrate them".
There's one thing to have a strong opinion about something, but when someone goes down the rabbithole of conflating things to come up with ways to attack what they dislike they strayed into asshole territory and I have no problems at all calling them out for that behavior.
So Karl, I think you are a bit of a asshole when you write about Musk and his associated companies as evidenced by your conflation of a limited beta-program, it's NDA plus not mentioning the real circumstances and context of the crash compared with every other Tesla car/driver not enrolled in that program.
On the whole, I think it's good to highlight the risks and pitfalls of self-driving cars and the manufacturers responsibility in marketing the features correctly and you could have easily written a better piece with all the included facts, but you choose not to.
If you want to paint something in indignant faux black and white, don't be surprised if people call you out on it. Do better.
[ link to this | view in chronology ]
Re: Re:
Teslas are supposedly so advanced, yet their ironically-named "Autopilot" can't detect stationary objects? The assist system in my Ford Escape has no problem detecting stationary objects, or people, or dogs. It has stopped for those objects numerous times in my use.
[ link to this | view in chronology ]
Re: Re: Re:
Why did it need to stop for stationary objects? Ford has no self driving general release products, so you were using a driver assist product and YOU were legally obliged to drive the car. YOU should have been stopping for stationary objects.
[ link to this | view in chronology ]
Re: A Dance As old As 2005
Behold the majestic fall migration for the Tesla fanbois has begun. As they work their way from on tech website to another defending their progenitor from perceived insults. Their displays of rhetoric ratchet up as they attempt to outdo each other in feats of logical fallacy as they vie to impress an extremely limited pool of available mates.
[ link to this | view in chronology ]
Re: Re: Re:
Why's it ironic? Actual autopilot systems (as used in aviation) don't do that either. But most people don't have enough aviation knowledge to know that, which makes it an unfortunate name.
[ link to this | view in chronology ]
Having self-driving and relying only on cameras is flawed from the start. I think LIDAR should be a requirement.
Tesla seems to have issues with running into the back of stopped cars for whatever reasons. That should be the easiest thing to NOT do.
[ link to this | view in chronology ]
Musk keeps rebranding other ideas
Looks like Musk jacked Steam's early access model and ran with it.
[ link to this | view in chronology ]
happens more than it should?
So, how often should it happen to a self-driving cars compared to how often non-self driving cars rear-end another car?
Only then can we say if self-driving cars are more dangerous than idiot driven cars.
[ link to this | view in chronology ]
Oh, wait…what?
Oh, wait…what?
For starters the title of the article is generally ignored in a rant about crashes.
This type of NDA is par for the corse in beta anything. Do not publicly disclose problems during the beta period.
It’s why windows beta testers were long inside locked to the public MSDN discussions. It’s why Apple dev beta (pre public) discussion is behind the ADS program.
I’ve tested software, and hardware devices. Some of them dangerous on error. This is nothing unique.
Crashes?
This reeks of not-my-fault-ism!
How is Tesla at fault for the driver not retaking control?
Be it an OS or a car: the point of a beta test is to gage reaction to real world situations, make manual corrections, and report flaws; “bugs”.
So regardless of if it’s alt-esc when a random /rm * starts for no reason, or bus cars and SUVs and people in the road… ?
Why did the driver not do something?
Unless self drive or lvl 2 or what ever disengaged the steering MIB style: the driver is responsible. To some degree.
The damn manual/guide/agreement says the driver is responsible for controlling the vehicle!
Beta testing is self explanatory to anyone who reads the agreement they sign! It’s not finished!!!!!
—
The problem here, on Tesla’s end, is not better vetting the idiots who agree to beta testing. If the driver was doing the job they agreed to, actively testing an assistive technology, they wouldn’t be plowing into things while sleeping, reading, screwing, etc. —
[ link to this | view in chronology ]
Re: Oh, wait…what?
This drunk driver mentioned at the start was not part of the FSD Beta test group. The article didn't do you any favors in explaining the situation. There have been NO accidents involving the ~2,000 Beta testers since they started in October 2020.
[ link to this | view in chronology ]
Re: Re: Oh, wait…what?
I think were agreeing?
When you chose to beta test something it’s not finished. It isn’t expected that it work correctly.
It’s prerelease.
The responsibility falls on the tester to maintain a safe operating environment.
[ link to this | view in chronology ]
Re: Re: Re: Oh, wait…what?
Yes, we're agreeing. I just didn't want you to put too much effort into arguing the FUD in the article. Tesla is carefully vetting the beta testers and they've been very safe. Drunk drivers hitting cop cars while failing to supervise standard release Autopilot are less safe.
[ link to this | view in chronology ]
They sue Tesla and some bar the driver had been drinking in...but what about the driver? He's apparently not named in the suit. Was he at least arrested for DUI?
Or was the automation a convenient excuse to shift blame to someone with more money than an ordinary drunk driver?
[ link to this | view in chronology ]
how to buy/invest in tesla shares from India?
Tesla share value is increasing day by day because of the rapid growth in electronic Vehicle usage in all countries. Try stockal to buy tesla shares in India easily!
[ link to this | view in chronology ]