Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

admin
Pinned April 30, 2018

<> Embed

@  Email

Report

Uploaded by user
Tesla insists Model X driver was at fault in fatal crash
<> Embed @  Email Report

Tesla insists Model X driver was at fault in fatal crash

Mariella Moon, @mariella_moon

April 12, 2018
Tesla insists Model X driver was at fault in fatal crash | DeviceDaily.com
 

When Tesla revealed that Autopilot was engaged during the fatal Model X crash in Mountain View, it only said that the vehicle’s “logs show[ed] that no action was taken” even though the driver had time to react. Now, the automaker has issued another statement much stronger than that, saying that the only way the accident could have happened was if the driver (identified as Apple engineer Walter Huang) wasn’t paying attention.

“We are very sorry for the family’s loss,” the statement sent to ABC 7 News reporter Dan Noyes started. “According to the family, Mr. Huang was well aware that Autopilot was not perfect and, specifically, he told them it was not reliable in that exact location, yet he nonetheless engaged Autopilot at that location. The crash happened on a clear day with several hundred feet of visibility ahead, which means that the only way for this accident to have occurred is if Mr. Huang was not paying attention to the road, despite the car providing multiple warnings to do so.” (Emphasis ours.)

The family’s lawyer believes that the company is doubling down on putting the blame on Huang in order to distract from their concerns about his vehicle’s Autopilot. Mike Fong, one of their attorneys, told ABC 7 that the sensors in Huang’s car misread the road’s lane lines and that its braking system failed to detect a stationary object on its way. Those are presumably the reasons why, on several occasions, his car veered towards the barrier where he crashed. While it sounds like the family is gearing up to sue, Fong said he doesn’t expect to file a lawsuit until after the National Transportation Safety Board is fully done with its investigation.

As for the rest of Tesla’s statement, it stresses that Autopilot requires drivers to have their hands on the wheel:

“The fundamental premise of both moral and legal liability is a broken promise, and there was none here. Tesla is extremely clear that Autopilot requires the driver to be alert and have hands on the wheel. This reminder is made every single time Autopilot is engaged. If the system detects that hands are not on, it provides visual and auditory alerts. This happened several times on Mr. Huang’s drive that day.

We empathize with Mr. Huang’s family, who are understandably facing loss and grief, but the false impression that Autopilot is unsafe will cause harm to others on the road. NHTSA found that even the early version of Tesla Autopilot resulted in 40% fewer crashes and it has improved substantially since then. The reason that other families are not on TV is because their loved ones are still alive.”

The Autopilot section on Tesla’s website does state that “every driver is responsible for remaining alert and active… and must be prepared to take action at any time.” However, it also features a video showing a person in the driver’s seat with their hands on their lap instead of the wheel, as Twitter product manager Patrick Traughber points out:

It’s far from the first time the automaker defended its Autopilot technology after a fatal crash. Tesla eyed its braking system instead of its self-driving tech for the 2016 incident wherein a Model S collided with a truck. The NTSB, however, blamed both the drivers in the incident and Autopilot itself. It cited previous findings that Autopilot’s ability to monitor drivers’ actions was extremely limited. The agency asked automakers to do more to ensure drivers are paying attention, and according to Bloomberg, Tesla told the NTSB it had already incorporated features to make that happen.

Engadget RSS Feed

(47)


Top