How neural networks simulate symbolic reasoning
Hear from CIOs, CTOs, and other C-level and senior execs on data and AI strategies at the Future of Work Summit this January 12, 2022. Learn more
Researchers at the University of Texas have discovered a new way for neural networks to simulate symbolic reasoning. This discovery sparks an exciting path toward uniting deep learning and symbolic reasoning AI.
In the new approach, each neuron has a specialized function that relates to specific concepts. “It opens the black box of standard deep learning models while also being able to handle more complex problems than what symbolic AI has typically handled,” Paul Blazek, University of Texas Southwestern Medical Center researcher and one of the authors of the Nature paper, told VentureBeat.
This work complements previous research on neurosymbolic methods such as MIT’s Clevrer, which has shown some promise in predicting and explaining counterfactual possibilities more effectively than neural networks. Additionally, DeepMind researchers previously elaborated on another neural network approach that outperformed state-of-the-art neurosymbolic approaches.
Essence neural networks mimic human reasoning
The team at the University of Texas coined the term, “essence neural network” (ENN) to characterize its approach, and it represents a way of building neural networks rather than a specific architecture. For example, the team has implemented this approach with popular architectures such as convolutional neural net and recurrent neural net (RNN) architectures.
The big difference is that they did away with backpropagation, which is a cornerstone of many AI processes. “Backpropagation famously opened deep neural networks to efficient training using gradient descent optimization methods, but this is not generally how the human mind works,” Blazek said. ENNs don’t use backpropagation or gradient descent. Rather, ENNs mimic the human reasoning process, learn the structure of concepts from data, and then construct the neural network accordingly.
Blazek said the new technique could have practical commercial applications in the next few years. For example, the team has demonstrated a few ENN applications to automatically discover algorithms and generate novel computer code. “Standard deep learning took several decades of development to get where it is now, but ENNs will be able to take shortcuts by learning from what has worked with deep learning thus far,” he said.
Promising applications of the new technique include the following:
- Cognitive science: The researchers designed ENNs as a proof-of-principle for their new neurocognitive theory. It integrates ideas from the philosophy of mind, psychology, neuroscience, and artificial intelligence to explore how the human mind processes information. The theoretical framework could prove beneficial in exploring various theories and models from all these fields.
- Algorithm discovery: The researchers found that ENNs can discover new algorithms, similarly to how people can.
- High-stakes applications: The research establishes basic building blocks for explainable deep learning systems that can be better understood before deployment and post hoc analysis.
- Robust AI: There has been great concern about adversarial attacks against black-box AI systems. ENNs are naturally more robust to adversarial attacks, particularly for symbolic reasoning use-cases.
- Machine teaching with limited data: An ENN can train on limited, idealistic data and then generalize to much more complex examples that it has never seen.
Working backward from biology to understand the brain
In contrast to most AI research, the researchers approached the problem from a biological perspective. “The original purpose of our work was to understand how the neuronal structure of the brain processes information,” Blazek said.
The team ultimately proposed a generalized framework for understanding how the brain processes information and encodes cognitive processes. The core idea is that each neuron makes a specialized distinction, either signifying a specific concept or differentiating between two opposing concepts. In other words, one type of neuron makes the distinction “like A” versus “not like A,” and the other kind of neuron makes the distinction “more like A” versus “more like B.”.
These neurons are arranged in an appropriate hierarchy to integrate these distinctions and arrive at more sophisticated conclusions. There are many ways to design the specialized distinction made by each neuron and to arrange the neurons to make complex decisions.
This theory of understanding neural information processing agrees with various theories and observations from philosophy of mind, psychology, and neuroscience. “The surprising thing about this framework is that the neurons reason about ideas in the exact same way that philosophers have always described our reasoning process,” Blazek said.
- up-to-date information on the subjects of interest to you
- our newsletters
- gated thought-leader content and discounted access to our prized events, such as Transform 2021: Learn More
- networking features, and more
Source: Read Full Article