**Sandia explores neural computing to extend Moore’s Law**

**Sandia explores neural computing to extend Moore’s Law**

Computation is stuck in a rut. The integrated circuits that powered the past 50 years of technological revolution are reaching their physical limits.

This predicament has computer scientists scrambling for new ideas: new devices built using novel physics, new ways of organizing units within computers and even algorithms that use new or existing systems more efficiently. To help coordinate new ideas, Sandia National Laboratories has assisted organizing the Institute of Electrical and Electronics Engineers (IEEE) International Conference on Rebooting Computing held Oct. 17-19.

Researchers from Sandia’s Data-driven and Neural Computing Dept. will present three papers at the conference, highlighting the breadth of potential non-traditional neural computing applications.

“We’re taking a stab at the scope of what neural algorithms can do. We’re not trying to be exhaustive, but rather we’re trying to highlight the kind of application over which algorithms may be impactful,” said Brad Aimone, a computational neuroscientist and co-author of one paper. Historically, neural computing has been seen as approximate and fuzzy, he added; however, Sandia researchers in their papers aim to extend neural algorithms so they incorporate rigor and predictability, which shows they may have a role in high performance scientific computing.

The three papers are entitled “Overcoming the Static Learning Bottleneck — the Need for Adaptive Neural Learning” by Craig Vineyard and Steve Verzi; “Computing with Dynamical Systems” by Fred Rothganger; and “Spiking Network Algorithms for Scientific Computing” by William Severa, Ojas Parekh, Kris Carlson, Conrad James and Aimone.

**Troubles and benefits of continuously learning**

The brain is continually learning. “While we do learn in school, our learning doesn’t stop when school ends. Instead, our brains are continually adapting through processes, such as synaptic modifications. However, most machine-learning algorithms learn once and are done,” said Vineyard, a computer scientist.

Most so-called machine-learning algorithms have a learning phase and a separate testing and operation phase. This is really time consuming. Ambitious — and challenging — attempts to develop algorithms that learn continuously also run the risk of the algorithm “learning” something that’s wrong, Vineyard said.

His paper argues for continual learning and suggests the use of game theory — the mathematics of logical decisions, such as when to take out the trash and when to hope your roommate will do it for you — to bring precision to the decision of when an algorithm should learn.

**What are dynamical systems anyway?**

A dynamical system is an equation that describes how things change with time. A simple dynamical system is a function that describes the movement of a grandfather clock’s pendulum. “The idea behind using dynamical systems for computation is to build a machine such that its dynamics — which has to do with the structure of the machine or the structure of the math — will lead it to the answer based on feeding it the question,” said Rothganger, a computer scientist.

Both our brains and, in a way, conventional computers are dynamical systems: They find answers just based on the question and how the computers are constructed, said Rothganger. His paper proposes that if researchers think of a traditional scientific computing problem, matrix decomposition, as a dynamical system, they could solve them rigorously on neuro-inspired systems.

“There’s a lot of potential and also a lot of risk in the idea I’m working on,” said Rothganger. If his idea works, “it would provide a point of unification between neural algorithms and traditional numerical algorithms.”

**Artisan mathematicians craft spiking network algorithms**

The third paper identifies three hand-crafted algorithms that use the careful arrangement of spiking neuron-like nodes to perform precise computations. In the brain, each neuron is connected to many other neurons and uses spikes of electricity to communicate. Severa, a mathematician, and his co-authors took inspiration from these aspects of the brain.

An example of these innovative algorithms is a kind of flow estimation called particle image velocimetry. By taking two pictures of dust motes moving through the air and figuring out how far they moved in the time between photos, researchers can determine the speed of the air and any local eddies. This can be done on a conventional computer using fancy math, but Severa’s method uses the massively parallel nature of neurons to calculate all the possible shifts efficiently, he said.

“By carefully designing your networks and the properties of your neurons, you can do exact things,” said Severa. “You can push the envelope of what you can expect a neural network to do.”

Whether the future holds neuro-inspired computers in your cellphone that understand phrases like “Show me a cute picture of Fluffy” and “Order my favorite Chinese food,” or if neural computers can also work alongside future quantum computers in solving tough math problems quickly, computing needs to be reinvented, and soon, said Aimone. By bringing together experts in many different disciplines, he said the International Conference on Rebooting Computing aims to nurture new ideas and spur this revolution.

Funding for all the projects was provided by Sandia’s Laboratory Directed Research and Developmentoffice. Two projects also were part of the Hardware Acceleration of Adaptive Neural Algorithms (HAANA) Grand Challenge.

**The broader rebooting computing effort**

Sandia employees are among the organizers of the IEEE’s Rebooting Computing initiative and the conference. Sandia’s Chief Technology Officer Rob Leland will give the conference kickoff talk on the history of innovation in computing. Sandia researchers Erik DeBenedictis and Matt Marinella are members of the conference program committee.

DeBenedictis, Sapan Agarwal, Jeanine Cook and Michael Frank also are presenting four papers on low-energy logic and memory. Christopher DeRose and Tony Lentine are presenting a paper on optical communications.

Learn more: Turning to the brain to reboot computing

##### The Latest on: Neural algorithms

*via Google News*

##### The Latest on: Neural algorithms

- Guitar virtuoso Plini and Neural DSP release amp modelling plugin on April 17, 2019 at 7:13 am
Finally, last in the signal chain is the Time Effects section, featuring a tape delay algorithm modeled using neural networks for vintage warmth, and a multidimensional reverb algorithm. Plini: ... […]

- Animal testing exercise to be put to an end with this new algorithm on April 17, 2019 at 4:58 am
The algorithm compares tested and untested chemical fragment ... (ANI) Washington D.C. [USA], Apr 17 (ANI): A recent study helps to decode the neural basis of dream emotions. As part of the project, ... […]

- AI Is Better Than People At Tracking Biological Neural Networks on April 16, 2019 at 11:50 am
Our deep learning-based algorithm is fast, and is demonstrated to be as accurate as human experts To measure neural activity, researchers typically use a process known as two-photon calcium ... […]

- Algorithm could spoil Game of Thrones character deaths on April 16, 2019 at 4:10 am
Firstly with a fairly straightforward statistical approach known as Bayesian inference and the second relied on techniques involving machine learning and neural networks ... Using the TUM team's ... […]

- BSC Researchers Shrink Floating Point Formats to Accelerate Deep Neural Network Training on April 15, 2019 at 11:01 am
A case in point is with deep learning. In early stages of training a Deep Neural Network (DNN), a lot of guesswork goes on. The algorithm assigns random values to the weights and computes the error. ... […]

- Artificial Intelligence BREAKTHROUGH: 'Better than human' AI algorithm beats brain in test on April 15, 2019 at 5:00 am
The algorithm, according to its lead author ... “This improved performance in active neuron detection should provide more information about the neural network and behavioural states, and open the door ... […]

- Google Content Algorithms and Ranking Effects on April 9, 2019 at 7:12 pm
Bill Slawski and I had an email discussion about a recent algorithm. Bill suggested a specific research paper and patent might be of interest to look at. What Bill suggested challenged me to think ... […]

- New algorithms and tools accelerate embedded vision adoption on April 9, 2019 at 6:55 am
Instead, developer focus can shift to selecting among available algorithms, and then to obtaining the necessary quantities of training data. Deep neural networks (DNNs) have transformed computer ... […]

- Convolutional Neural Network Deep Learning Techniques for Crowd Counting on April 8, 2019 at 8:38 pm
CNN algorithms use density-based estimations to preserve spatial ... InfoQ: How can Cascaded Convolutional Neural Networks help to solve crowd-counting type of problems? Kesari: We used Cascaded CNNs ... […]

*via Bing News*