Artificial intelligence learns a neural network when it should not be trusted


Neural network confidence

MIT researchers have developed a neural network pathway of deep learning to quickly estimate confidence levels in their output. Advanced AI can increase safety and efficiency in assistive decision making. Credit: M.I.T.

A.I. A quick way to estimate uncertainty in supportive decision making can lead to safer results.

Increasingly, artificial intelligence systems, known as neural networks of deep learning, are used to inform decisions about human health and safety, such as autonomous driving or medical diagnosis. These networks are good at identifying patterns in large, complex datasets to help make decisions. But how do we know they are true? Alexander Amini and his allies MIT And Harvard University wanted to find out.

They have developed a fast way to crush data for neural networks, and have not only predicted the model’s confidence level based on the quality of the data available, but also created the output. Advance can save lives, as worldwide education is already being organized in the real world today. The level of certainty of the network may be the difference between an independent health vehicle that determines that “it is clear to proceed through the intersection” and “it is clear, so just hang in the case.”

Current methods of estimating uncertainty for neural networks are costly and relatively slow on a computational basis for split-second decisions. But Amini’s approach, known as “deep clear regression”, speeds up the process and can lead to safer results. “We need to understand not only the high-performance models, but also the ones,” says Amini, a PhD student at the MIT Computer Science and Artificial Intelligence Laboratory (CSAL) Professor Daniel Russ Group. .

“This idea is important and widely applied. It can be used to assess products that rely on learned models. By estimating the uncertainty of the scholarly model, we will also learn how much error should be expected from the model, and what missing data can improve the model, ”says Rousseau.

Amini will present research at next month’s Neurips conference, with Russ, who is Andrew and Erna Viterby, professors of electrical engineering and computer science, director of CSAL, and Stephen A. of MIT. Schwarzmann, Deputy Dean of Research in Computing; And Wilco Schwarting of MIT and graduates of AV Solima from MIT and Harvard.

Efficient uncertainty

After up-and-down history, deep education has shown remarkable performance on various tasks, in some cases even surpassing humans. Accuracy. And nowadays, wherever the computer goes, there seems to be deep learning. It fuels search engine results, social media feeds and facial recognition. “We have had great success using Deep Danda education,” says Amini. “Knowing the correct answer 99 percent of the time, the neural network is really good.” 99 percent of people won’t cut it when life is on the line.

“One thing that excludes researchers is the ability to know these models and tell us when they might be wrong.” “We really care about 1 percent of that time, and how we can detect those situations reliably and effectively.”

Neural networks can be massive, sometimes with billions of dimensions in a row. So, just to get the answer it can be a heavy calculation lift, leave a confidence level. Uncertainty analysis is not new in neural networks. But previous approaches, arising from Baisian deep learning, have often relied on running a neural network or sampling to understand its beliefs. It takes time and memory in the process, a luxury that does not exist in high speed traffic.

Researchers found a way to estimate uncertainty with just one run of the neural network. They designed the network with bulk up output, which not only made the decision but also distributed a new possibility to get evidence in support of that decision. These distributions, called explicit evidence distributions, directly predict the model’s credibility. This includes any uncertainty present in the underlying input data, as well as in the final decision of the model. This difference may indicate whether the uncertainty can be reduced by tweaking the neural network manually, or whether the input data is just noisy.

Confidence check

To put their approach to testing, the researchers began with a challenging computer vision task. They trained their neural network to analyze a monochromatic color image and to estimate the depth value (i.e. distance from the camera lens) for each pixel. An autonomous vehicle can use similar calculations to estimate the proximity of a pedestrian or other vehicle, which is not an easy task.

The performance of their network was comparable to previous state-of-the-art models, but it also had the ability to estimate its own uncertainties. As the researchers had hoped, the network predicts high uncertainty for pixels where it predicts incorrect depth and depth. “It was highly calibrated by the errors that the network makes, which we believe was one of the most important factors in determining the quality of the new uncertainty estimates,” says Amini.

To stress-test their calibration, the team also demonstrated that the network predicts high uncertainty for “out-of-distribution” data – completely new types of images were never encountered during training. After training the network on indoor scenes, they fed him a bunch of outdoor driving scenes. The network constantly warned that its responses to the novel’s outdoor scenes were uncertain. The test highlighted the network’s ability to flag when users should not have full confidence in its decisions. In these cases, “if this is a health care application, we may not trust the diagnosis given by the doctor, and seek a different opinion instead,” says Emini.

The network also knew when the photos were docked, potential hedging against data-manipulation attacks. In another experiment, researchers increased the level of opposing noise in a group of images fed to them on the network. The effect was subtle – hardly noticeable to the human eye – but the network sniffed those images and tagged its output with a high degree of uncertainty. This ability to sound the alarm on false information can help detect and deter counter-attacks, a growing concern in the age of deepfax.

Deep investigative regression is “a simple and elegant approach that moves forward in the field of estimating uncertainty, which is important for robotics and other real-world control systems,” says Repo Headsell, who was not involved with the work. “This is done in an innovative way that avoids some of the awkward aspects of other approaches – e.g. patterns or stripes – which makes it not only elegant but also computationally more efficient – a winning combination.”

Deep Unda clear offenses can increase the safety of AI-assisted decision making. “We are starting to see a lot of this [neural network] Models leave the research lab and in the real world result in situations that touch humans with potentially life-threatening consequences, ”says Amini. “Any user of the method, whether he is a doctor or a person in the passenger seat of a vehicle, needs to be aware of any risk or uncertainty associated with that decision.” He imagines the system to quickly flag uncertainty, but he also uses it to make more informed decisions in dangerous situations, such as an autonomous vehicle approaching an intersection.

“Any field that is going to do deployed machine learning ultimately needs to have an awareness of reliable uncertainty,” he says.

The work was supported in part by the Toyota-CSAL Joint Research Center, the National Science Foundation and the Toyota Research Institute.