As the self-driving car travels down a dark, rural road, a deer lingering among the trees up ahead looks poised to dart into the car’s path. Will the vehicle know exactly what to do to keep everyone safe?
Some computer scientists and engineers aren’t so sure. But researchers at the University of Virginia School of Engineering and Applied Science are hard at work developing methods they hope will bring greater confidence to the machine-learning world – not only for self-driving cars, but for planes that can land on their own and drones that can deliver your groceries.
The heart of the problem stems from the fact that key functions of the software that guides self-driving cars and other machines through their autonomous motions are not written by humans – instead, those functions are the product of machine learning. Machine-learned functions are expressed in a form that makes it essentially impossible for humans to understand the rules and logic they encode, thereby making it very difficult to evaluate whether the software is safe and in the best interest of humanity.
Researchers in UVA’s Leading Engineering for Safe Software Lab – the LESS Lab, as it’s commonly known – are working to develop the methods necessary to provide society with the confidence to trust emerging autonomous systems.
The team’s researchers are Matthew B. Dwyer, Robert Thomson Distinguished Professor; Sebastian Elbaum, Anita Jones Faculty Fellow and Professor; Lu Feng, assistant professor; Yonghwi Kwon, John Knight Career Enhancement Assistant Professor; Mary Lou Soffa, Owens R. Cheatham Professor of Sciences; and Kevin Sullivan, associate professor. All hold appointments in the UVA Engineering Department of Computer Science. Feng holds a joint appointment in the Department of Engineering Systems and Environment.
Since its creation in 2018, the LESS Lab has rapidly grown to support more than 20 graduate students, publish more than 50 publications and obtain competitive external awards totaling more than $10 million in research funding. The awards have come from such agencies as the National Science Foundation, the Defense Advanced Research Projects Agency, the Air Force Office of Scientific Research and the Army Research Office.
The lab’s growth trajectory matches the scope and urgency of the problem these researchers are trying to solve.
The Machine-Learning Explosion
An inflection point in the rapid rise of machine learning happened just a decade ago when computer vision researchers won the ImageNet Large Scale Visual Recognition Challenge – to identify objects in photos – using a machine-learning solution. Google took notice and quickly moved to capitalize on the use of data-driven algorithms.
Other tech companies followed suit, and public demand for machine-learning applications snowballed. Last year, Forbes estimated that the global machine-learning market grew at a compound annual growth rate of 44% and is on track to become a $21 billion market by 2024.
But as the technology ramped up, computer scientists started sounding the alarm that mathematical methods to validate and verify the software were lagging.
Government agencies like the Defense Advanced Research Projects Agency responded to the concerns. In 2017, DARPA launched the Assured Autonomy program to develop “mathematically verifiable approaches” for assuring an acceptable level of safety with data-driven, machine-learning algorithms.
UVA Engineering’s Department of Computer Science also took action, extending its expertise with a strategic cluster of hires in software engineering, programming languages and cyber-physical systems. Experts combined efforts in cross-cutting research collaborations, including the LESS Lab, to focus on solving a global software problem in need of an urgent solution.
A Canary in the Coal Mine
By this time, the self-driving car industry was firmly in the spotlight and the problems with machine learning were becoming more evident. A fatal Uber crash in 2018 was recorded as the first human death from an autonomous vehicle.
“The progress toward practical autonomous driving was really fast and dramatic, shaped like a hockey stick curve, and much faster than the rate of growth for the techniques that can ensure their safety,” Elbaum said.
This August, tech blog Engadget reported that one self-driving car company, Waymo, had put in 20 million miles of on-the-road testing. Yet, devastating failures continue to occur; as the software has gotten more and more complex, no amount of real-world testing would be enough to find all the bugs.
And the software failures that have occurred have made many wary of on-the-road testing.
“Even if a simulation works 100 times, would you jump into an autonomous car and let it drive you around? Probably not,” Dwyer said. “Probably you want some significant on-the-street testing to further increase your confidence. And at the same time, you don’t want to be the one in the car when that test goes on.”
And then there is the need to anticipate – and test – every single obstacle that might come up in the 4-million-mile public roads network.
The LESS Lab’s team of researchers are Matthew B. Dwyer, Lu Feng, Yonghwi Kwon, Mary Lou Soffa, Kevin Sullivan and Sebastian Elbaum. (Contributed photos)
“Think about the complexities of a particular scenario, like driving on the highway with hauler trucks on either side so that an autonomous vehicle needs to navigate crosswinds at high speeds and around curves,” Dwyer said. “Getting a physical setup that matches that challenging scenario would be difficult to do in real-world testing. But in simulation that can get that much easier.”
And that’s one area where LESS Lab researchers are making real headway. They are developing sophisticated virtual reality simulations that can accurately create such difficult scenarios that might otherwise be impossible to test on the road.
“There is a huge gap between testing in the real world versus testing in simulation,” Elbaum said. “We want to close that gap and create methods in which we can expose autonomous systems to the many complex scenarios they will face. This will substantially reduce testing time and cost, and it is a safer way to test.”

