Machine Learning Times
Machine Learning Times
EXCLUSIVE HIGHLIGHTS
The Rise Of Large Database Models
 Originally published in Forbes Even as large language models have...
3 Predictions For Predictive AI In 2025
 Originally published in Forbes GenAI’s complementary sibling, predictive AI, makes...
The Quant’s Dilemma: Subjectivity In Predictive AI’s Value
 Originally published in Forbes This is the third of a...
To Deploy Predictive AI, You Must Navigate These Tradeoffs
 Originally published in Forbes This is the second of a...
SHARE THIS:

4 years ago
Algorithm Helps Artificial Intelligence Systems Dodge “Adversarial” Inputs

 
Originally published in Massachusetts Institute of Technology, March 8, 2021. 

Method builds on gaming techniques to help autonomous vehicles navigate in the real world, where signals may be imperfect.

In a perfect world, what you see is what you get. If this were the case, the job of artificial intelligence systems would be refreshingly straightforward.

Take collision avoidance systems in self-driving cars. If visual input to on-board cameras could be trusted entirely, an AI system could directly map that input to an appropriate action — steer right, steer left, or continue straight — to avoid hitting a pedestrian that its cameras see in the road.

But what if there’s a glitch in the cameras that slightly shifts an image by a few pixels? If the car blindly trusted so-called “adversarial inputs,” it might take unnecessary and potentially dangerous action.

A new deep-learning algorithm developed by MIT researchers is designed to help machines navigate in the real, imperfect world, by building a healthy “skepticism” of the measurements and inputs they receive.

The team combined a reinforcement-learning algorithm with a deep neural network, both used separately to train computers in playing video games like Go and chess, to build an approach they call CARRL, for Certified Adversarial Robustness for Deep Reinforcement Learning.

The researchers tested the approach in several scenarios, including a simulated collision-avoidance test and the video game Pong, and found that CARRL performed better — avoiding collisions and winning more Pong games — over standard machine-learning techniques, even in the face of uncertain, adversarial inputs.

To continue reading this article, click here.

 

Leave a Reply