Press "Enter" to skip to content

Do neural networks need rest like human brains to perform well?

Leslie 0

Androids may or may not have to count digital sheep to catch a wink, but they will almost certainly need periods of rest to perform consistently.

The reason: When rested, neural networks perform better, similar to how human brains benefit from sleep. This is the hypothesis of a new research from Los Alamos National Laboratory.

“We study spiking neural networks, which are systems that learn much as living brains do,” said Los Alamos National Laboratory computer scientist Yijing Watkins.

“We were fascinated by the prospect of training a neuromorphic processor in a manner analogous to how humans and other biological systems learn from their environment during childhood development.”

Watkins and her research team found that the network simulations became unstable after continuous periods of unsupervised learning.

When they exposed the networks to states that are analogous to the waves that living brains experience during sleep, stability was restored. “It was as though we were giving the neural networks the equivalent of a good night’s rest,” said Watkins.

The goal

The researchers are now planning to implement their algorithm on Intel’s Loihi neuromorphic chip. They hope allowing Loihi to sleep from time to time will enable it to stably process information from a silicon retina camera in real time.

If the findings confirm the need for sleep in artificial brains, we can probably expect the same to be true of androids and other intelligent machines that may come about in the future, according to the researchers.

Watkins will be presenting the research at the Women in Computer Vision Workshop on June 14 in Seattle.

Dawning of the realisation

The discovery came about as the research team worked to develop neural networks that closely approximate how humans and other biological systems learn to see.

The group initially struggled with stabilizing simulated neural networks undergoing unsupervised dictionary training, which involves classifying objects without having prior examples to compare them to.

“The issue of how to keep learning systems from becoming unstable really only arises when attempting to utilize biologically realistic, spiking neuromorphic processors or when trying to understand biology itself,” said Los Alamos computer scientist and study coauthor Garrett Kenyon.

“The vast majority of machine learning, deep learning, and AI researchers never encounter this issue because in the very artificial systems they study, they have the luxury of performing global mathematical operations that have the effect of regulating the overall dynamical gain of the system.”

The researchers characterize the decision to expose the networks to an artificial analog of sleep as nearly a last ditch effort to stabilize them.

They experimented with various types of noise, roughly comparable to the static you might encounter between stations while tuning a radio.

Result

The best results came when they used waves of so-called Gaussian noise, which includes a wide range of frequencies and amplitudes.

The researchers hypothesize that the noise mimics the input received by biological neurons during slow-wave sleep. The results suggest that slow-wave sleep may act, in part, to ensure that cortical neurons maintain their stability and do not hallucinate.