• Home
  • TESLA AUTOPILOT IS “CAUSING CRASHES”! –

TESLA AUTOPILOT IS “CAUSING CRASHES”! –

Tesla Autopilot crash caught on dashcam shows how not to use the system

Tesla AUTOPILOT CAUSES A CRASH , Caught On Camera!

ELON MUSK TRIES TO LIE HIS WAY OUT OF ANOTHER ONE OF THE HUNDREDS OF TESLA INJURIES, CRASHES AND DEATHS!!!

Earlier this week, a Tesla Model S hit a barrier on the highway near Dallas, Texas. The driver, who fortunately wasn’t injured, first blamed Tesla’s Autopilot for the crash.

We now have footage of the accident and it actually shows a situation that the Autopilot probably shouldn’t be expected to be able to handle, at least not yet. Ultimately, it serves as a reminder not to trust the system without paying attention.

Following our articles on a series of accidents last year where the Autopilot was activated during or right before the crashes, some readers were confused on whether the driver or the Autopilot should be considered at fault.

Since under its current form, Tesla’s Autopilot is only a “driver assist” system and drivers are asked to keep their hands on the steering wheel, the responsibility falls on the driver. Of course, that’s unless the Autopilot malfunctions and automatically steers away from the lane and into the side of the road, which is almost what we were led to believe with this latest accident, but that has so far never happened as far as we know.

The driver described the accident in a Reddit post on Monday:

“I was driving in the left lane of a two lane highway. The car is AP1 (first generation Autopilot) and I’ve never had any problems until today. Autopilot was on and didn’t give me a warning. It misread the road and hit the barrier. After the airbags deployed there was a bunch of smoke and my car rolled to a grinding stop. Thankfully no one was hurt and I walked away with only bruises.”

He attached pictures of the aftermath:

Fast forward to 3 days later. Another Redditor on the Tesla Motors subreddit found footage of the accident taken from the dashcam of a vehicle following the Tesla during the event.

The footage shows that the Tesla needed to merge or change lane in order to avoid the barrier – something the Autopilot should never be left to do without the driver intervening.autopilot-accident-2-gif

What is also clear from the footage is that the design of the road here is quite awful since even the driver in the vehicle with the dashcam almost hit the barrier and there presumably wasn’t any driver assist at play in this case.

As far as the Autopilot’s Autosteer feature, it did its job, which is to keep the vehicle in its lane which was still marked on the road leading right into the barrier.

What potentially didn’t work is the ‘Forward Collision Warning’ feature since the driver claims that there was no warning. Some would assume that Automatic Emergency Braking (AEB) should have kicked in, but it’s actually not designed to engage if there’s an alternative and in this case, the vehicle wasn’t supposed to brake in order to avoid the barrier – it could have been even more dangerous considering a vehicle was close behind and there was traffic to the right of the vehicle.

Tesla explains what the feature does:

“AEB does not engage when an alternative collision avoidance strategy (e.g., driver steering) remains viable. Instead, when a collision threat is detected, forward collision warning alerts the driver to encourage them to take appropriate evasive action.”
Of course, collision warning is no substitute to paying attention when driving.