Credit: Human Factors and Ergonomics Society Automakers and their tech partners may be looking toward a future when they can offer fully autonomous cars, but Mica Endsley, the author of the recently released Human Factors paper, ”From Here to Autonomy,” contends that without considering the cognitive constraints of the driver, safety can’t be assured. People will still be needed to oversee, direct, and intervene in the actions of autonomous systems for the foreseeable future. Automobiles are only one among many areas of implementation to which this applies; the same can be said for cybersecurity, unmanned aircraft, space exploration, cargo delivery, health monitoring, and more. Endsley, who conceptualized and has worked extensively on human situation awareness, examines four decades of human factors research on human-automation interaction (aka autonomy) and has developed the Human-Autonomy System Oversight model (HASO). The model identifies critical cognitive processes and user interface features that must be considered when designing and testing any autonomous system.
The challenge for humans in monitoring autonomous systems is remaining aware of and attentive to the system state and the environment so they can respond when the system alone can’t do so safely and be able to act in partnership with the automation to achieve optimum results. Behind this challenge are a number of cognitive factors, including becoming complacent when nothing is going wrong, overly trusting the automation to take the correct action when needed, and the sheer boredom that can arise when not actively engaged in automated tasks, all of which get worse the better the automation is. This creates what Endsley posits is an automation conundrum:
”The more automation is added to a system, and the more reliable and robust that automation is, the less likely it is that human operators overseeing the automation will be aware of critical information and able to take over manual control when needed.”
According to Endsley, autonomous systems should be, among other things,Used only where necessary to keep human engagement highConsistent to help people understand the systemFocused on improving user awareness through information integrationMapped to users’ goals and mental models, minimizing cognitive loadSalient in making mode transitions to the human highly obviousTransparent, providing understandability and predictabilityNoncomplex to improve understanding and reduce user errors.Endsley notes, ”the article includes specific guidelines that designers can implement in building the operator interfaces for autonomous systems to better keep people in the loop and able to interact with the system autonomy effectively. This will help engineers and organizations that are trying to figure out what to do about the problem.”
Endsley concludes, ”As long as human oversight of autonomous systems and intervention is required to achieve successful performance, the automation conundrum will undermine performance and safety in many applications.”
Explore further:GM begins autonomous car tests in Scottsdale, Arizona
More information: ”From Here to Autonomy: Lessons Learned from Human-Automation Research” DOI: 10.1177/0018720816681350
Provided by:Human Factors and Ergonomics Society