Back to Explorer
Research PaperResearchia:202602.10050[Data Science > Machine Learning]

When do neural ordinary differential equations generalize on complex networks?

Moritz Laber

Abstract

Neural ordinary differential equations (neural ODEs) can effectively learn dynamical systems from time series data, but their behavior on graph-structured data remains poorly understood, especially when applied to graphs with different size or structure than encountered during training. We study neural ODEs (nODE\mathtt{nODE}s) with vector fields following the Barabási-Barzel form, trained on synthetic data from five common dynamical systems on graphs. Using the S1\mathbb{S}^1-model to generate graphs with realistic and tunable structure, we find that degree heterogeneity and the type of dynamical system are the primary factors in determining nODE\mathtt{nODE}s' ability to generalize across graph sizes and properties. This extends to nODE\mathtt{nODE}s' ability to capture fixed points and maintain performance amid missing data. Average clustering plays a secondary role in determining nODE\mathtt{nODE} performance. Our findings highlight nODE\mathtt{nODE}s as a powerful approach to understanding complex systems but underscore challenges emerging from degree heterogeneity and clustering in realistic graphs.


Source: arXiv:2602.08980v1 - http://arxiv.org/abs/2602.08980v1 PDF: https://arxiv.org/pdf/2602.08980v1 Original Link: http://arxiv.org/abs/2602.08980v1

Submission:2/10/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!