Login / Signup

Network model with internal complexity bridges artificial intelligence and neuroscience.

Linxuan HeYunhui XuWeihua HeYihan LinYang TianYujie WuWenhui WangZiyang ZhangJunwei HanYonghong TianBo XuGuoqi Li
Published in: Nature computational science (2024)
Artificial intelligence (AI) researchers currently believe that the main approach to building more general model problems is the big AI model, where existing neural networks are becoming deeper, larger and wider. We term this the big model with external complexity approach. In this work we argue that there is another approach called small model with internal complexity, which can be used to find a suitable path of incorporating rich properties into neurons to construct larger and more efficient AI models. We uncover that one has to increase the scale of the network externally to stimulate the same dynamical properties. To illustrate this, we build a Hodgkin-Huxley (HH) network with rich internal complexity, where each neuron is an HH model, and prove that the dynamical properties and performance of the HH network can be equivalent to a bigger leaky integrate-and-fire (LIF) network, where each neuron is a LIF neuron with simple internal complexity.
Keyphrases
  • artificial intelligence
  • big data
  • machine learning
  • deep learning
  • mental health
  • preterm birth