Site icon Stuff South Africa

Drivers of self-driving cars can rely too much on autopilot, and that’s a recipe for disaster

self-driving tesla

We were promised a very near future where autonomous machines would be serving our needs and vehicle ownership would be rendered unnecessary: robots would quickly and efficiently deliver our orders and we could squeeze in a few more hours of work or sleep while being chauffeured around in self-driving cars.

Progress has been made, at least, on some of this. University campuses and cities across North America have indeed witnessed the growing presence of small food-delivery robots. Likewise, new partnerships have recently been announced to develop and test the safety of self-driving trucks.

The journey toward autonomous or self-driving consumer cars, on the other hand, has arguably come to a screeching halt. In 2021, top industry experts recognized that developing safe autonomous driving systems was not as simple as it was anticipated. Among them, Elon Musk himself conceded that developing the technology required to deliver safe self-driving cars has proved harder that he thought.

Automation paradox

More bad news came this week when the U.S. National Highway Traffic Safety Administration (NHTSA) released numbers that showed Tesla vehicles being responsible for nearly 70 per cent of the crashes involving so-called SAE Level 2 cars.

Some cars are completely autonomous and are capable of driving without any input from the human driver. For example, Waymo One, in Phoenix, Ariz., is a ride-hailing service that currently deploys autonomous cars on a test route.

SAE Level 2 autonomous systems, like Tesla Autopilot, require human drivers to stay alert at all times, even when the system temporarily takes control of steering and acceleration. As soon as the traffic or road conditions aren’t adequate for the system to operate, control is given back to the driver who needs to take over manual control of the vehicle.


Read more: Why Tesla’s Autopilot crashes spurred the feds to investigate driver-assist technologies – and what that means for the future of self-driving cars


Human factors engineering is a cross-disciplinary research field investigating how humans interact with vehicle technology. Its researchers have, for years, highlighted the safety risks of automated driving — especially when the system requires the driver to make up for technological shortcomings to operate safely.

This is the case in what is known as the automation paradox, wherein the more automated the vehicle, the harder it is for humans to operate it properly.

Underestimating vehicle capability

Among the most prominent risks of operating SAE Level 2 cars is when drivers misunderstand the capabilities of the automated system. The issue often leads to unsafe behaviours like reading a book or taking a nap while the vehicle is in motion.


Read more: Who’s to blame when a self-driving car has an accident?


YouTube player

 

In 2021, there were so many reports of unsafe behaviours at the wheel of Level 2 cars, that the NHTSA required manufacturers to start reporting crashes that had occurred when these systems were engaged.

The initial findings, released in June 2022, showed that since 2021, Tesla and Honda vehicles were, respectively, involved in 273 and 90 reported crashes when these systems were engaged. Most crashes occurred in Texas and California.

While these data paint a dismal picture of the safety of these systems, they pale in comparison to the over 40,000 reported fatal crashes that occurred in the United States in 2021 alone.

As part of the same report, NHTSA itself highlights some of the methodological limitations of the study: from the incompleteness of some of the source data to failing to account for individual manufacturers’ total vehicle volume or distance travelled by vehicles.

For the skeptics, this does not spell the end of autonomous cars. It does, however, confirm that widespread deployment of safe self-driving cars is not years, but decades, in the making.

Exit mobile version