Why Tesla Won't Stop Despite Endless Software Recalls
Because cars have become rolling computers, Tesla treats recalls as software patches akin to something Microsoft or Apple would do for phones or laptops.
The potentially more significant safety issues do not appear to factor into its strategy. So far. JL
Rebecca Heilweil reports in Re/code:
Tesla announced two massive recalls this week related to issues with its vehicles’ safety-related software updates. The automaker has initiated nine recallswithin the past three months,
and federal safety regulators
now feel like this is a regular occurrence. Some of
the problems that the software updates are meant to address seem to be
ongoing or, worsening. In recent years, automakers have had the option of fixing some flaws
with a software update, just like Tesla has done with all the aforementioned recalls. Because these are akin to
releasing an app update, they’re easier and faster for car owners to
address. But they’re alsocheapfor automakers to execute.
Tesla announced two massive recalls this week related to issues with its vehicles’ software. One of the recalls ordered Tesla to roll back a self-driving feature that caused the company’s cars to break the law. While Elon Musksaid“there were no safety issues,” federal regulators disagree. But it’s also increasingly clear that government pressure is unlikely to change Tesla’s brazen approach.
In fact, these recently recalled features are just the latest examples of the company sellingpotentially dangerous software. This trend is especially concerning because the National Highway Traffic Safety Administration (NHTSA), the agency thatwill oversee the recalls, can onlytake actionafter the Tesla software has beenreleased to drivers. Tesla typically pushes these updates to its vehicles when they’re connected to the internet, so they don’t come ata significant cost to the company. That’s left Tesla free to release and test new features with drivers on the road — until the government catches up and intervenes, or Tesla identifies issues on its own.
Tesla’sfirst recall this weekinvolved a feature that enabled users to program their Teslas to slow down but not completely stop at all-way intersections with stop signs, under very specific conditions. The company first introduced the so-called “rolling stop” feature back in October 2020 as part of an update to its Full Self-Driving beta software. Thisadvanced versionof Tesla’s Autopilot driver-assistance technology is only available to a limited number of Tesla owners who pay$12,000for the upgrade andis famouslycontroversial. Despite the name, the Full Self-Driving software doesn’t actually turn Teslas into fully autonomous vehicles, but it does give drivers access to semi-autonomous features. The recent recall of the rolling stop feature affected all 53,822 Teslas that had upgraded to Full Self-Driving.
It’s unclear exactly how dangerous Tesla’s “rolling stop” feature was in practice. No warranty claims or injuries were reported to Tesla, and the system only worked if the car was driving at very slow speeds. Elon Musk claims that speed was2 miles per hour, while official NHTSA notes say it was5.6 miles per hour. Regardless, rolling through stop signs isillegalin most places, no matter how slow you’re going. As NHTSAnotedin a letter to Tesla, “Failing to stop at a stop sign can increase the risk of a crash.” Tesla has now released a software update to address the recall.
Tesla’ssecond recallcame on Thursday after NHTSA said that more than 817,000 Teslas could have a faulty seat belt feature. The agency said that these vehicles, in certain cases, wouldn’t produce an audible chime when people forget to put on their seat belts, a problem that Teslasaysit will, again, fix with a software update.
One software update released by Tesla allowed drivers to play video games while their cars were in motion.Andy Cross/The Denver Post via Getty Images
These are just the two most recent Tesla safety-related software updates. The automaker has initiated nine recallswithin the past three months, and federal safety regulators opening investigations into the company now feel like this is a regular occurrence. At the same time, some of the problems that the software updates are meant to address seem to be ongoing or, in some cases, worsening.
For example, Teslaissued a recall— and a software update — following reports last fall that vehicles using Autopilot would brake without warning after their computers detected hazards that didn’t actually exist. But this “phantom braking” issue actually appeared to get worse in the months following the recall, possibly because Tesla decided to stop using radar-based sensors, in favor of cameras, on some vehicle models, according toa recent report from the Washington Post. Federal regulators are again consulting the company about how to address the problem.
“NHTSA has kind of been playing a game of cat and mouse with Tesla for a number of years now,” explained Michael Brooks, the acting director of the Center for Auto Safety. “Right now, NHTSA seems to be playing a little more active version of a cat.”
Yet Tesla shows no sign of stopping. Again, car recalls don’t work the way they once did. Car owners used to take their recalled vehicles to a dealer or auto shop for a repair, or they would get a replacement or a refund. But in recent years, automakers have had the option of fixing some flaws with an over-the-air software update, just like Tesla has done with all of the aforementioned recalls. Because these recalls are akin to releasing an app update, they’re easier and faster for car owners to address. But they’re alsorelatively cheapfor automakers to execute. So some might argue that, if recalls simply lead to software updates, they’re little more than a slap on the wrist from regulators.
It looks as though Tesla is also using software updates to preempt potential recalls, too. This pastsummer, the company released a feature calledPassenger Playthat let people play video games on Teslas’ touchscreens while the vehicles were in motion. But after NHTSA announcedan investigation into the feature, Tesla issued a software update in December todisablePassenger Play while the car is in motion. Last August, the agency similarlyannounced a formal investigationinto 11 crashes involving Tesla vehicles with Autopilot or Traffic-Aware Cruise Control engaged that hit stopped emergency vehicles with their lights flashing, and then Tesla pusheda related software updatea few weeks later. That prompted NHTSA to ask the company in October why itdidn’t file a formal recall notice, implying that Tesla may have used the update as a “stealth recall.”
Other than formal recalls, regulatorsdon’t currently have many other toolsat their disposal.NHTSA could update itsFederal Motor Vehicle Safety Standards, but that’s ayearslong process, a fact that might help explain why regulations haven’t kept pace with new car technologies. The National Transportation Safety Board can also focus on car safety, but the agency is mostly limited to conductinginvestigations into crashesand making recommendations. While states areslowly developingtheir own rules, there still isn’t a nationwide consensus on how to approach vehicles like Teslas.
The Department of Transportation is still wrestling with how to approach increasingly autonomous vehicles.Drew Angerer/Getty Images
That’s why some think it’s time for thegovernment to do more. Last May, Reps. Bobby Rush (D-IL) and Larry Bucshon (R-IN) proposed theCrash Avoidance System Evaluation Act, which would force the Department of Transportation to study how well car crash avoidance tech, like Tesla’s Autopilot and Full Self-Driving, actually works. Meanwhile, Sens. Ed Markey (D-MA) and Richard Blumenthal (D-CT) urged the Federal Trade Commission in August toinvestigateTesla for false and misleading advertising related to those semi-autonomous features. Along with Sen. Amy Klobuchar (D-MN), they’ve also proposed theStay Aware for Everyone Act, which would mandate that the DOT analyze driver assistance systems and require carmakers to install driver-monitoring tools.
“These latest developments with Tesla’s Autopilot and Full Self-Driving systems are exactly why we’ve long expressed our concerns with this flawed technology,” Blumenthal and Markey told Recode. “Although automated driving and driver assistance systems have the potential to enhance safety, they must be implemented with strong safeguards that will ensure our cars follow the rules of the road and drivers are fully engaged.”
But the reason strong safeguards for semi-autonomous cars don’t exist is that there’s no consensus about how these vehicles should be regulated. Is the driver or the car manufacturer responsible when a software-assisted vehicle does something dangerous? Should autonomous and semi-autonomous cars be programmed to follow the letter of the law, or to drive the way most people actually drive? Regulators haven’t fully answered these questions, and Tesla is taking advantage of the ambiguity.
This all comes at what should be a moment of opportunity for Tesla and its commitment to safety. In the first nine months of last year, thenumber of US road fatalitiesrose at the fastest rate since the Transportation Department started tracking them in 1975. Musk has repeatedly argued that Tesla’s AI-powered vehicles save lives and that his vehicles could be a solution to worsening highway safety problems. But the company’s insistence on releasing features that deliberately play with safety, instead of promoting it, seems to be undermining this larger goal.
As a Partner and Co-Founder of Predictiv and PredictivAsia, Jon specializes in management performance and organizational effectiveness for both domestic and international clients. He is an editor and author whose works include Invisible Advantage: How Intangilbles are Driving Business Performance. Learn more...
0 comments:
Post a Comment