Canadian Manufacturing

U.S. ends probe into fatal Tesla crash, says Autopilot reduces accident rate

In win for self-driving technology, the NHSTA will not order a recall or impose a fine for a deadly Tesla crash last year involving Autopilot technology


Print this page


One component of Tesla’s Autopilot, Autosteer uses sensors to keep cars centred in their lane. PHOTO: Alexis Georgeson/Tesla

WASHINGTON—Tesla Motors Inc. won’t face a recall or fine as a result of a fatal crash involving its Autopilot system, but U.S. safety regulators are warning auto manufacturers and drivers not to treat semiautonomous cars as if they were fully self-driving.

The National Highway Traffic Safety Administration said Jan. 20 it found that the system had no safety defects at the time of the May 7 crash in Florida, and that it was primarily designed to prevent rear-end collisions rather than other crash scenarios.

Bryan Thomas, the agency’s chief spokesman, said automated driving systems still require a driver’s full attention. He warned that automakers need to keep tabs on how drivers use the technology and should design vehicles “with the inattentive driver in mind.”

The probe began June 28, nearly two months after a driver using Autopilot in a 2015 Tesla Model S died when it failed to spot a tractor-trailer crossing the car’s path on a highway in Williston, Florida, near Gainesville.

Tesla’s Autopilot uses cameras, radar and computers to detect objects and automatically brake if the car is about to hit something; it’s comprise mainly of Traffic-Aware Cruise Control and Autosteer, which can steer the car to keep it centred in its lane. The company has said that before Autopilot can be used, drivers must acknowledge that it’s an “assist feature” that requires both hands on the wheel at all times and that drivers must be ready to take control.

The agency’s criticism is likely to influence how automakers market semi-autonomous systems. Just about every company has or is working on similar systems as they move rapidly toward self-driving cars.

The investigation “helps clarify that cars are still supposed to be driven by attentive people, and if people behind the wheel aren’t attentive, it’s not the technology’s fault,” said Karl Brauer, executive publisher of Kelley Blue Book. That will help avoid the stigma that the technology causes accidents, he said.

NHTSA released guidelines last year that attempt to ensure safety without slowing development of semiautonomous and self-driving cars. The agency says self-driving features could dramatically reduce traffic deaths by eliminating human error, which plays a role in 94 per cent of fatal crashes.

Thomas said NHTSA wants to encourage innovation “to get the best answer to how we use these automated systems to the best effect and saving the most lives.”

In its probe, NHTSA evaluated how the system functions and looked into dozens of other crashes involving Teslas, including a July one on the Pennsylvania Turnpike that injured two people.

The Florida crash killed former Navy Seal Joshua Brown, 40, of Canton, Ohio. Tesla said at the time that the cameras on Brown’s Model S sedan failed to distinguish the white side of a turning tractor-trailer from a brightly lit sky and that neither the car nor Brown applied the brakes.

Thomas said Brown set the car’s cruise control at 74 mph—9 mph over the limit—less than two minutes before the crash. NHTSA’s crash reconstruction showed the tractor-trailer should have been visible to Brown at least 7 seconds before impact, enough time to react.

Detecting vehicles that cross in its path were beyond the capabilities of the Autopilot system, Thomas said on a conference call.

Meanwhile, investigators also analyzed crash data from before and after Autopilot software was installed in Tesla vehicles. It found Autosteer improved the vehicle crash rate by nearly 40 per cent, decreasing the crash rate to 0.8 per million miles from 1.3 per million miles.

In a statement Thursday, Tesla said it appreciated NHTSA’s thoroughness in reaching its conclusion.

When Tesla released Autopilot in 2015, some safety advocates questioned whether the Palo Alto, California, company and NHTSA allowed the public access to the system before testing was finished.

Consumer Reports magazine called on Tesla to drop the “Autopilot” name because it can give drivers too much trust in their car’s ability to drive itself.

In September, Tesla updated Autopilot software to rely more on radar sensors and less on cameras. The update also disabled the automatic steering if drivers don’t keep both hands on the wheel.

Another federal agency, the National Transportation Safety Board, has opened a broader investigation into the Tesla crash. It could be months before a final report that provides a probable cause for the collision is issued.

The company that made the camera and computer system for Tesla said in September that the company ignored its warnings about possible safety problems.

Israel-based Mobileye that before the release of Autopilot, it warned Tesla not to allow drivers to use the system without their hands on the steering wheel. Mobileye NV, a huge player in the self-driving business, has stopped supplying components to Tesla.

Thomas said investigators got information from Mobileye and evaluated the company’s statements, but still reached the no-defect conclusion. He wouldn’t comment on the company’s statements.

Tesla said at the time that Mobileye’s statements were inaccurate and stem from Tesla’s plans to develop its own vision system.

With a file from CanadianManufacturing.com Staff


Print this page

Related Posts from the network