{"id":6103,"date":"2022-11-15T21:15:00","date_gmt":"2022-11-15T21:15:00","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2022\/11\/15\/solving-brain-dynamics-gives-rise-to-flexible-machine-learning-models\/"},"modified":"2022-11-15T21:15:00","modified_gmt":"2022-11-15T21:15:00","slug":"solving-brain-dynamics-gives-rise-to-flexible-machine-learning-models","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2022\/11\/15\/solving-brain-dynamics-gives-rise-to-flexible-machine-learning-models\/","title":{"rendered":"Solving brain dynamics gives rise to flexible machine-learning models"},"content":{"rendered":"<p>Author: Rachel Gordon | MIT CSAIL<\/p>\n<div>\n<p>Last year, MIT researchers announced that they had <a href=\"https:\/\/news.mit.edu\/2021\/machine-learning-adapts-0128\">built \u201cliquid\u201d neural networks<\/a>, inspired by the brains of small species: a class of flexible, robust machine learning models that learn on the job and can adapt to changing conditions, for real-world safety-critical tasks, like driving and flying. The flexibility of these \u201cliquid\u201d neural nets meant boosting the bloodline to our connected world, yielding better decision-making for many tasks involving time-series data, such as brain and heart monitoring, weather forecasting, and stock pricing.<\/p>\n<p>But these models become computationally expensive as their number of neurons and synapses increase and require clunky computer programs to solve their underlying, complicated math. And all of this math, similar to many physical phenomena, becomes harder to solve with size, meaning computing lots of small steps to arrive at a solution.\u00a0<\/p>\n<p>Now, the same team of scientists has discovered a way to alleviate this bottleneck by solving the differential equation behind the interaction of two neurons through synapses to unlock a new type of fast and efficient artificial intelligence algorithms. These modes have the same characteristics of liquid neural nets \u2014 flexible, causal, robust, and explainable\u00a0\u2014 but are orders of magnitude faster, and scalable. This type of neural net could therefore be used for any task that involves getting insight into data over time, as they\u2019re compact and adaptable even after training \u2014 while many traditional models are fixed.\u00a0<\/p>\n<p>The models,\u00a0dubbed a \u201cclosed-form continuous-time\u201d (CfC) neural network, outperformed state-of-the-art counterparts on a slew of tasks, with considerably higher speedups and performance in recognizing human activities from motion sensors, modeling physical dynamics of a simulated walker robot, and event-based sequential image processing. On a medical prediction task, for example, the new models were 220 times faster on a sampling of 8,000 patients.\u00a0<\/p>\n<p>A new <a href=\"https:\/\/www.nature.com\/articles\/s42256-022-00556-7\" target=\"_blank\" rel=\"noopener\">paper on the work<\/a> is published today in <em>Nature Machine Intelligence.<\/em><\/p>\n<p>\u201cThe new machine-learning models we call \u2018CfC\u2019s\u2019 replace the differential equation defining the computation of the neuron with a closed form approximation, preserving the beautiful properties of liquid networks without the need for numerical integration,\u201d says MIT Professor Daniela Rus, director of the Computer Science and Artificial Intelligence Laboratory (CSAIL) and senior author on the new paper. \u201cCfC models are causal, compact, explainable, and efficient to train and predict. They open the way to trustworthy machine learning for safety-critical applications.\u201d<\/p>\n<p><strong>Keeping things liquid\u00a0<\/strong><\/p>\n<p>Differential equations enable us to compute the state of the world or a phenomenon as it evolves, but not all the way through time \u2014 just step-by-step. To model natural phenomena through time and understand previous and future behavior, like human activity recognition or a robot\u2019s path, for example, the team reached into a bag of mathematical tricks to find just the ticket: a \u201cclosed form&#8217;\u201d solution that models the entire description of a whole system, in a single compute step.\u00a0<\/p>\n<p>With their models, one can compute this equation at any time in the future, and at any time in the past. Not only that, but the speed of computation is much faster because you don\u2019t need to solve the differential equation step-by-step.\u00a0<\/p>\n<p>Imagine an end-to-end neural network that receives driving input from a camera mounted on a car. The network is trained to generate outputs, like the car&#8217;s steering angle. In 2020, the team solved this by using liquid neural networks with 19 nodes, so 19 neurons plus a small perception module could drive a car. A differential equation describes each node of that system. With the closed-form solution, if you replace it inside this network, it would give you the exact behavior, as it\u2019s a good approximation of the actual dynamics of the system. They can thus solve the problem with an even lower number of neurons, which means it would be faster and less computationally expensive.\u00a0<\/p>\n<p>These models can receive inputs as time series (events that happened in time), which could be used for classification, controlling a car, moving a humanoid robot, or forecasting financial and medical events. With all of these various modes, it can also increase accuracy, robustness, and performance, and, importantly, computation speed\u00a0\u2014 which sometimes comes as a trade-off.\u00a0<\/p>\n<p>Solving this equation has far-reaching implications for advancing research in both natural and artificial intelligence systems. \u201cWhen we have a closed-form description of neurons and synapses\u2019 communication, we can build computational models of brains with billions of cells, a capability that is not possible today due to the high computational complexity of neuroscience models. The closed-form equation could facilitate such grand-level simulations and therefore opens new avenues of research for us to understand intelligence,\u201d says MIT CSAIL Research Affiliate Ramin Hasani, first author on the new paper.<\/p>\n<p><strong>Portable learning<\/strong><\/p>\n<p>Moreover, there is early evidence of Liquid CfC models in learning tasks in one environment from visual inputs, and transferring their learned skills to an entirely new environment without additional training. This is called out-of-distribution generalization, which is one of the most fundamental open challenges of artificial intelligence research.\u00a0\u00a0<\/p>\n<p>\u201cNeural network systems based on differential equations are tough to solve and scale to, say, millions and billions of parameters. Getting that description of how neurons interact with each other, not just the threshold, but solving the physical dynamics between cells enables us to build up larger-scale neural networks,\u201d says Hasani. \u201cThis framework can help solve more complex machine learning tasks \u2014 enabling better representation learning \u2014 and should be the basic building blocks of any future embedded intelligence system.\u201d<\/p>\n<p>\u201cRecent neural network architectures, such as neural ODEs and liquid neural networks, have hidden layers composed of specific dynamical systems representing infinite latent states instead of explicit stacks of layers,\u201d says Sildomar Monteiro, AI and Machine Learning Group lead at Aurora Flight Sciences, a Boeing company, who was not involved in this paper. \u201cThese implicitly-defined models have shown state-of-the-art performance while requiring far fewer parameters than conventional architectures. However, their practical adoption has been limited due to the high computational cost required for training and inference.\u201d He adds that this paper \u201cshows a significant improvement in the computation efficiency for this class of neural networks \u2026 [and] has the potential to enable a broader range of practical applications relevant to safety-critical commercial and defense systems.\u201d<\/p>\n<p>Hasani and Mathias Lechner, a postdoc at MIT CSAIL, wrote the paper supervised by Rus, alongside MIT Alexander Amini, a CSAIL postdoc; Lucas Liebenwein SM &#8217;18, PhD &#8217;21; Aaron Ray, an MIT electrical engineering and computer science PhD student and CSAIL affiliate; Max Tschaikowski, associate professor in computer science at Aalborg University in Denmark; and Gerald Teschl, professor of mathematics at the University of Vienna.<\/p>\n<\/div>\n<p><a href=\"https:\/\/news.mit.edu\/2022\/solving-brain-dynamics-gives-rise-flexible-machine-learning-models-1115\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: Rachel Gordon | MIT CSAIL Last year, MIT researchers announced that they had built \u201cliquid\u201d neural networks, inspired by the brains of small species: [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2022\/11\/15\/solving-brain-dynamics-gives-rise-to-flexible-machine-learning-models\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":473,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[24],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/6103"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=6103"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/6103\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/467"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=6103"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=6103"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=6103"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}