Assistant Professor Dhruv Batra Earns Prestigious ECASE-Army Award
School of Interactive Computing Assistant Professor Dhruv Batra was recently selected as a recipient of the prestigious Early Career Award for Scientists and Engineers (ECASE-Army) by the Army Research Office, providing five years’ worth of research funding to make artificial intelligence (AI) systems more transparent, explainable, and trustworthy.
The award, which provides a total of $1 million over the course of the grant, comes as a result of Batra’s selection for a similar early-career award by the Army Research Office Young Investigator Program in 2014.
The research Batra’s lab will pursue with the funding addresses a fundamental challenge in development of AI – their “black-box” nature, the consequent difficulty humans face in identifying why or how AI systems fail, and how to improve upon those technologies. When a self-driving car from a major tech company, for example, suffered its first fatality in 2015, legal and regulatory agencies understandably questioned what went wrong. The challenge at the time was providing a sufficient answer to that question.
“Your response can’t just be, ‘Well, there was this machine learning box in there, and it just didn’t detect the car. We don’t know why,’” said Batra, who is also a member of the Machine Learning and GVU Centers.
Batra’s research aims to create AI systems that can more readily explain what they do and why. This could come in the form of natural language or visual explanations, both of which – computer vision and natural language processing – are central areas of focus in Batra’s lab. The machine could, for example, identify regions in image that provide support for its predictions, potentially assisting a user’s understanding of what the machine can or cannot do.
It’s an important area of study for a few reasons, Batra said. He classifies AI technology into three levels of maturity:
- Level 1 is technology that is in its infancy. It is not near deployment to everyday users, and the consumers of the technology are researchers. The goal for transparency and explanation is to help researchers and developers to understand the failure modes and current limitations, and deduce how to improve the technology – “actionable insight,” as Batra called it.
- Level 2 is when things are working to a degree, enough so that the technology can and has been deployed.
“The technology may be mature in a narrow range, and you can ship the product,” Batra said. “Like face detection or fingerprint technology. It’s built into products and being used at agencies, airports, or other places.”
In such cases, you want explanations and interpretability that helps build appropriate trust with users. Users can understand when the system reliably works and when it might not work – face detection in bad lighting, for example – and make efforts to use in a more appropriate setting.
- Level 3 is typically a fairly narrow category where the AI is better – sometimes significantly so – than the human. Batra used chess-playing and Go-playing bots as an example. The best chess-playing bots convincingly outperform the best humans and reliably hand a resounding defeat to the average human player.
“We already know bots play much better than humans,” he said. “In such cases, you don’t need to improve the machine and you already trust its skill level. You want the machine to give you explanations not so that you can improve the AI, but so that you can improve yourself.”
Batra envisions scenarios where the techniques his lab develops could assist at all three levels, but the experiments will take place between Levels 1 and 2. They will work in Visual Question Answering, which are agents that answer natural language questions about visual content, and other areas of maturity that may reach the product level in five or more years.
The funding will begin for Batra in January. Batra has served as an assistant professor at Georgia Tech since Fall 2016. Visit his website for more information about his research.