Automakers’ Reliance on People to Take Emergency Control of Their Self-Driving Car May be Fatal Flaw

Thursday, July 21, 2016
Woman works while leaving driving to her car (photo: Volvo)

By Joan Lowy, Associated Press

 

WASHINGTON (AP) — Experts say the development of self-driving cars over the coming decade depends on an unreliable assumption by many automakers: that the humans in them will be ready to step in and take control if the car's systems fail.

 

Instead, experience with automation in other modes of transportation like aviation and rail suggests that the strategy will lead to more deaths like that of a Florida Tesla driver in May.

 

Decades of research shows that people have a difficult time keeping their minds on boring tasks like monitoring systems that rarely fail and hardly ever require them to take action. The human brain continually seeks stimulation. If the mind isn't engaged, it will wander until it finds something more interesting to think about. The more reliable the system, the more likely it is that attention will wane.

 

Automakers are in the process of adding increasingly automated systems that effectively drive cars in some or most circumstances, but still require the driver as a backup in case the vehicle encounters a situation unanticipated by its engineers.

 

Tesla's Autopilot, for example, can steer itself within a lane and speed up or slow down based on surrounding traffic or on the driver's set speed. It can change lanes with a flip of its signal, automatically apply brakes, or scan for parking spaces and parallel park on command.

 

Joshua Brown, a 40-year-old tech company owner from Canton, Ohio, who was an enthusiastic fan of the technology, was killed when neither he nor his Tesla Model S sedan's Autopilot braked for a truck making a left turn on a highway near Gainsville, according to federal investigators and the automaker.

 

Tesla warns drivers to keep their hands on the wheel even though Autopilot is driving, or the vehicle will automatically slow to a stop. A self-driving system Audi plans to introduce in its 2018 A7, which the company says will be the most advanced on the market, monitors drivers' head and eye movements, and automatically slows the car if the driver's attention is diverted.

 

But Brown's failure to brake means he either didn't see the truck in his path or saw it too late to respond — an indication he was relying on the automation and his mind was elsewhere, said Missy Cummings, director of Duke University's Humans and Autonomy Laboratory. The truck driver said he had heard a Harry Potter video playing in the car after the crash.

 

"Drivers in these quasi- and partial modes of automation are a disaster in the making," Cummings said. "If you have to rely on the human to see something and take action in anything less than several seconds, you are going to have an accident like we saw."

 

Operators — an airline pilot, a train engineer or car driver — can lose awareness of their environment when they turn control over to automation, said Rob Molloy, the National Transportation Safety Board's chief highway crash investigator.

 

He pointed to the crash of Air France Flight 447 into the Atlantic Ocean while flying from Brazil to France in 2007. A malfunction in equipment used to measure air speed caused the plane's autopilot to disconnect, catching pilots by surprise. Confused, they caused an otherwise flyable plane to stall and fall from the sky, killing 228 people.

 

Planes and trains have had automation "for 20, 30 years and there are still times when they're like, 'Wow, we didn't expect that to happen,'" Molloy said.

 

Part of the problem is overconfidence in the technology causes people to think they can check out.

 

Transportation Secretary Anthony Foxx warned automakers on Tuesday that they should realize that drivers will be tempted to use the technology in irresponsible ways and take that into account as they build their robotic systems.

 

"People are getting distracted by the coolness of the technology," Foxx said following a speech at a self-driving vehicle conference in San Francisco. "In many cases, they are going beyond what the technology is capable of doing."

 

Not long after Tesla introduced its Autopilot system, people were posting videos of car with the self-driving mode engaged cruising down tree-lined roads or even highways with no one in the driver's seat. Brown, for example, had posted videos lauding the Autopilot system and demonstrating it in action.

 

"There is a tendency of people to take one ride in one of these vehicles and then conclude that because they have not crashed over the course of 10 minutes that the system must be ready," said Bryant Walker Smith, a University of South Carolina professor who studies the technology.

 

Some experts think the ability of people to monitor autonomous systems may be getting worse. With the advent of smartphones, people are accustomed to having their desire for mental stimulation satisfied immediately.

 

"Go into Starbucks, for example," said Cummings. "No one can just patiently wait in line, they're all doing something on their phones. It's kind of pathetic."

 

Google, meanwhile, is aiming for a car that's fully self-driving and may not even have a steering wheel or brake pedals.

 

Associated Press writers Tom Krisher in Detroit and Michael Liedtke in San Francisco contributed to this report.

 

To Learn More:

Navy SEAL Becomes First Fatality in a Self-Driving Car (by Tom Krisher, Associated Press)

Computers are Drivers, Says NHTSA in Boost to Self-Driving Auto Industry (by Tom Krisher and Justin Pritchard, Associated Press

Ready or Not, Here Come Tesla Drivers and New Autosteer (by Ken Broder, AllGov California)

Human Drivers Create Headaches for Law-Abiding Driverless Cars (by Noel Brinkerhoff and Steve Straehley, AllGov)

Comments

Leave a comment