Nearly all the neural networks that power modern artificial intelligence tools such as ChatGPT are based on a 1960s-era computational model of a living neuron. A new model developed at the Flatiron Institute’s Center for Computational Neuroscience (CCN) suggests that this decades-old approximation doesn’t capture all the computational abilities that real neurons possess and that this older model is potentially holding back AI development.
The new model developed at CCN posits that individual neurons exert more control over their surroundings than previously thought. The updated neuron model could ultimately lead to more powerful artificial neural networks that better capture the powers of our brains, the model developers say.
The researchers present the model in a paper published in the journal Proceedings of the National Academy of Sciences.
“Neuroscience has advanced quite a bit in these past 60 years, and we now recognize that previous models of neurons are rather rudimentary,” says Dmitri Chklovskii, a group leader at the CCN and senior author of the new paper. “A neuron is a much more complex device—and much smarter—than this overly simplified model.”
Artificial neural networks aim to mimic the way in which the human brain processes information and makes decisions, albeit in a much more simplified manner. These networks are built of ordered layers of “nodes” based on the 1960s neuron model. The network starts with an input layer of nodes that receives information, then has middle layers of nodes that process the information, and then ends with an output layer of nodes that sends out the results.
Typically, a node will only pass information to the next layer if the total input it receives from the previous layer’s nodes exceeds a certain threshold. When current artificial neural networks are trained, information goes through a node in only one direction, and there’s no way for nodes to influence the information they receive from nodes earlier in the chain.
In contrast, the newly published model treats neurons as tiny “controllers,” an engineering term for devices that can influence their surroundings based on information gathered about those surroundings. Not merely passive relays of input, our brain cells may actually work to control the state of their fellow neurons.
Chklovskii believes that this more realistic model of a neuron-as-controller could be a significant step toward improving the performance and efficiency of many machine learning applications.
“Although AI’s achievements are very impressive, there are still many problems,” he says. “The current applications can give you wrong answers, or hallucinate, and training them requires a lot of energy; they’re very expensive. There are all these problems that the human brain seems to avoid. If we were to understand how the brain actually does this, we could build better AI.”
The neuron-as-controller model was inspired by what scientists do understand about large-scale circuits in the brain made up of many neurons. Most brain circuits are thought to be organized into feedback loops, where cells later in the processing chain influence what happens earlier in the chain. Much like a thermostat maintaining the temperature of a house or building, brain circuits need to keep themselves stable to avoid overwhelming the body’s system with activity.
Chklovskii says it was not entirely intuitive that this kind of feedback control could also be realized by an individual brain cell. He and his colleagues realized that a novel form of control, known as direct data-driven control, is straightforward and efficient enough to be biologically plausible as taking place in individual cells.
“People thought of the brain as a whole or even parts of the brain as being a controller, but no one suggested that a single neuron could do that,” Chklovskii says. “Control is a computationally intensive task. It’s hard to think of a neuron as having enough computational capacity.”
Viewing neurons as mini-controllers also explains several previously unexplained biological phenomena, Chklovskii says. For example, it’s long been appreciated that there’s a lot of noise in the brain, and the purpose of this biological randomness has been debated, but the CCN team found through their modeling that certain types of noise could actually enhance neurons’ performance.
More specifically, at junctions where one neuron connects to another (called ‘synapses’), there are often instances where a neuron transmits an electrical signal but the downstream coupled neuron does not get the message. Whether and when the downstream neuron receives a synaptic signal seems to be governed largely by chance.
While other scientists have speculated that such randomness was just the nature of small biological systems and not important to neuron behavior, the Flatiron team found that adding noise to their model neuron allowed it to adapt to a constantly changing environment. The randomness appears to be important in replicating how real neurons function, the team found.
Chklovskii next plans to analyze types of neurons that don’t fit their new model. For example, neurons in the retina receive direct inputs from the visual environment. These neurons may not be able to control their inputs the way neurons deeper in the brain can, but they might use some of the same principles Chklovskii and his team identified: namely, these neurons might be able to predict their inputs, even if they can’t influence them.
“Control and prediction are actually very related,” Chklovskii says. “You cannot control efficiently without predicting the impact of your actions in the world.”
More information:
Chklovskii, Dmitri B., The neuron as a direct data-driven controller, Proceedings of the National Academy of Sciences (2024). DOI: 10.1073/pnas.2311893121. doi.org/10.1073/pnas.2311893121
Simons Foundation
New computational model of real neurons could lead to better AI (2024, June 24)
retrieved 24 June 2024
from https://techxplore.com/news/2024-06-real-neurons-ai.html
part may be reproduced without the written permission. The content is provided for information purposes only.
Nearly all the neural networks that power modern artificial intelligence tools such as ChatGPT are based on a 1960s-era computational model of a living neuron. A new model developed at the Flatiron Institute’s Center for Computational Neuroscience (CCN) suggests that this decades-old approximation doesn’t capture all the computational abilities that real neurons possess and that this older model is potentially holding back AI development.
The new model developed at CCN posits that individual neurons exert more control over their surroundings than previously thought. The updated neuron model could ultimately lead to more powerful artificial neural networks that better capture the powers of our brains, the model developers say.
The researchers present the model in a paper published in the journal Proceedings of the National Academy of Sciences.
“Neuroscience has advanced quite a bit in these past 60 years, and we now recognize that previous models of neurons are rather rudimentary,” says Dmitri Chklovskii, a group leader at the CCN and senior author of the new paper. “A neuron is a much more complex device—and much smarter—than this overly simplified model.”
Artificial neural networks aim to mimic the way in which the human brain processes information and makes decisions, albeit in a much more simplified manner. These networks are built of ordered layers of “nodes” based on the 1960s neuron model. The network starts with an input layer of nodes that receives information, then has middle layers of nodes that process the information, and then ends with an output layer of nodes that sends out the results.
Typically, a node will only pass information to the next layer if the total input it receives from the previous layer’s nodes exceeds a certain threshold. When current artificial neural networks are trained, information goes through a node in only one direction, and there’s no way for nodes to influence the information they receive from nodes earlier in the chain.
In contrast, the newly published model treats neurons as tiny “controllers,” an engineering term for devices that can influence their surroundings based on information gathered about those surroundings. Not merely passive relays of input, our brain cells may actually work to control the state of their fellow neurons.
Chklovskii believes that this more realistic model of a neuron-as-controller could be a significant step toward improving the performance and efficiency of many machine learning applications.
“Although AI’s achievements are very impressive, there are still many problems,” he says. “The current applications can give you wrong answers, or hallucinate, and training them requires a lot of energy; they’re very expensive. There are all these problems that the human brain seems to avoid. If we were to understand how the brain actually does this, we could build better AI.”
The neuron-as-controller model was inspired by what scientists do understand about large-scale circuits in the brain made up of many neurons. Most brain circuits are thought to be organized into feedback loops, where cells later in the processing chain influence what happens earlier in the chain. Much like a thermostat maintaining the temperature of a house or building, brain circuits need to keep themselves stable to avoid overwhelming the body’s system with activity.
Chklovskii says it was not entirely intuitive that this kind of feedback control could also be realized by an individual brain cell. He and his colleagues realized that a novel form of control, known as direct data-driven control, is straightforward and efficient enough to be biologically plausible as taking place in individual cells.
“People thought of the brain as a whole or even parts of the brain as being a controller, but no one suggested that a single neuron could do that,” Chklovskii says. “Control is a computationally intensive task. It’s hard to think of a neuron as having enough computational capacity.”
Viewing neurons as mini-controllers also explains several previously unexplained biological phenomena, Chklovskii says. For example, it’s long been appreciated that there’s a lot of noise in the brain, and the purpose of this biological randomness has been debated, but the CCN team found through their modeling that certain types of noise could actually enhance neurons’ performance.
More specifically, at junctions where one neuron connects to another (called ‘synapses’), there are often instances where a neuron transmits an electrical signal but the downstream coupled neuron does not get the message. Whether and when the downstream neuron receives a synaptic signal seems to be governed largely by chance.
While other scientists have speculated that such randomness was just the nature of small biological systems and not important to neuron behavior, the Flatiron team found that adding noise to their model neuron allowed it to adapt to a constantly changing environment. The randomness appears to be important in replicating how real neurons function, the team found.
Chklovskii next plans to analyze types of neurons that don’t fit their new model. For example, neurons in the retina receive direct inputs from the visual environment. These neurons may not be able to control their inputs the way neurons deeper in the brain can, but they might use some of the same principles Chklovskii and his team identified: namely, these neurons might be able to predict their inputs, even if they can’t influence them.
“Control and prediction are actually very related,” Chklovskii says. “You cannot control efficiently without predicting the impact of your actions in the world.”
More information:
Chklovskii, Dmitri B., The neuron as a direct data-driven controller, Proceedings of the National Academy of Sciences (2024). DOI: 10.1073/pnas.2311893121. doi.org/10.1073/pnas.2311893121
Simons Foundation
New computational model of real neurons could lead to better AI (2024, June 24)
retrieved 24 June 2024
from https://techxplore.com/news/2024-06-real-neurons-ai.html
part may be reproduced without the written permission. The content is provided for information purposes only.