The U.S. National Transportation Safety Board ruled today that the “operational design” of Tesla’s Autopilot played a key role in a 2016 fatal crash that left the owner of a Model S dead. The vehicle’s semi-autonomous system permitted the driver, Joshua Brown to use Autopilot in ways “inconsistent” with warnings from Tesla that “permitted his prolonged disengagements from the driving task” and ultimately contributed to his death.>
On Tuesday, the National Transportation Safety Board is expected to find that Tesla’s…Read more
The nearly three-hour hearing offered a wide-ranging discussion on the limitations of automated systems on the road today. The NTSB suggested to any maker of semi-autonomous vehicles to prevent the use of the technology on roads where the vehicles aren’t suited to travel without human control of the vehicle.
Brown, a 40-year-old self-described Tesla fanatic, died in May 2016 when his Model S sedan rammed into the side of a trailer on a big-rig truck that was turning left in front of the vehicle. The Model S was traveling at 74 mph using Autopilot, which failed to identify the truck and never slowed the vehicle down, the NTSB said.
But Brown and the truck driver bore some fault, as well, in the Florida crash, the agency said.
The NTSB issued 13 findings related to the crash, including a pair of citations over Brown’s driving patterns. Brown relied too much on automation, and demonstrated a “lack of understanding of system limitations,” the agency said.
Part of that, the NTSB indicated, had to do with Tesla. Tesla has repeatedly said in the past that drivers must keep their hands on the wheel and pay attention to the road if they enable Autopilot.
But Tesla could’ve taken additional steps to prevent misuse of Autopilot, which worked as designed, the NTSB said, but the feature was used in ways that weren’t intended—and that Tesla didn’t go far enough to ensure drivers remained alert. (Tesla has previously said to a separate federal regulator that it evaluated the potential for some drivers to become inattentive while using the feature.)
“Tesla allowed the driver to use the system outside of the environment for which it was designed and the system gave far too much leeway to the driver to divert his attention,” said Robert Sumwalt, the NTSB chairman.
The truck driver, who refused to be interviewed by the NTSB, failed to yield the right of way to Brown, but the agency said both Tesla and the driver had “at least 10 seconds” to see each other, and there’s “no evidence” that either took evasive action to prevent the crash.
In a statement, a Tesla spokesperson said the “safety of its customers comes first” and “Autopilot significantly increases safety, as [the National Highway Traffic Safety Administration has found that it reduces accident rates by 40 percent.” (The conclusion of that 40 percent figure is the subject of an ongoing suit.)
“We appreciate the NTSB’s analysis of last year’s tragic accident and we will evaluate their recommendations as we continue to evolve our technology,” the spokesperson said. “We will also continue to be extremely clear with current and potential customers that Autopilot is not a fully self-driving technology and drivers need to remain attentive at all times.”
Brown’s death drew worldwide attention as it was the first known fatal crash to involve a vehicle traveling on its own using semi-automated technology. A preliminary report released earlier this year by the NTSB found that Brown kept his hands on the wheel of his 2015 Model S70D for only 25 seconds of an extended 37-minute period where his car was in Autopilot at 74 mph. The vehicle gave numerous audio and visual warnings before the crash.
Tesla CEO Elon Musk said himself that a later update to Autopilot—improving the system’s ability to differentiate a truck from a sign—“very likely” would’ve prevented the crash.
The NTSB staff said that Brown was knowledgable about the vehicle, but that the Tesla user manual offered sometimes conflicting accounts.
“A driver could have difficulties interpreting which roads it might be appropriate [to use Autopilot],” Ensar Becic, an NTSB human performance investigator, said during the hearing.
The NTSB also found that relying on sensors that determine whether a driver’s hands are on a steering wheel isn’t an effective method of determining whether they’re paying attention.
Above all, the NTSB’s 80-page report illustrated the limitations of tech currently commercially available semi-autonomous cars. The Tesla Model S was not a fully self-driving car, stressed Sumwalt, the NTSB chairman, in a warning call of sorts to consumers.
“It’s a long way from partially automated vehicles to self-driving cars,” he said. “And until we get there somebody still has to drive.”
Source : http://jalopnik.com/limits-of-teslas-autopilot-and-driver-error-cited-in-fa-1803806982