Recognition: unknown
Interpretable Relational Inference with LLM-Guided Symbolic Dynamics Modeling
Pith reviewed 2026-05-10 15:59 UTC · model grok-4.3
The pith
COSINE jointly recovers interaction graphs and sparse symbolic equations by letting an LLM adapt the mathematical library during optimization.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
COSINE is a differentiable co-optimization framework that simultaneously learns interaction graphs and sparse symbolic dynamical expressions; an outer-loop large language model adaptively prunes and expands the function library using feedback from the inner optimization, yielding both accurate structural recovery and compact, mechanism-aligned equations on synthetic benchmarks and large-scale epidemic data.
What carries the argument
The COSINE co-optimization loop that alternates between updating network edges and symbolic coefficients while an LLM uses inner-loop performance to revise the candidate function library.
If this is right
- Accurate recovery of hidden interaction structures becomes possible without assuming a fixed topology in advance.
- Dynamical equations remain sparse and directly readable, aligning with known mechanisms in the data-generating process.
- The same pipeline works on both small synthetic systems and large real-world epidemic records.
- Interpretability is achieved without sacrificing the accuracy previously available only from black-box neural surrogates.
Where Pith is reading between the lines
- The approach could be tested on climate or neural population data where partial mechanistic knowledge exists but full symbolic forms are unknown.
- If LLM bias remains small across domains, the need for hand-crafted function libraries in scientific modeling would decrease.
- Scaling tests on higher-dimensional systems would reveal whether sparsity is preserved when the number of variables grows.
Load-bearing premise
The true dynamics admit a sparse symbolic representation whose quality can be judged reliably by LLM feedback without systematic bias in the library suggestions.
What would settle it
Apply the method to a known dynamical system whose governing terms lie outside both the initial library and the LLM's subsequent proposals; recovery should then produce either dense non-sparse expressions or visibly incorrect interaction graphs.
Figures
read the original abstract
Inferring latent interaction structures from observed dynamics is a fundamental inverse problem in many-body interacting systems. Most neural approaches rely on black-box surrogates over trainable graphs, achieving accuracy at the expense of mechanistic interpretability. Symbolic regression offers explicit dynamical equations and stronger inductive biases, but typically assumes known topology and a fixed function library. We propose \textbf{COSINE} (\textbf{C}o-\textbf{O}ptimization of \textbf{S}ymbolic \textbf{I}nteractions and \textbf{N}etwork \textbf{E}dges), a differentiable framework that jointly discovers interaction graphs and sparse symbolic dynamics. To overcome the limitations of fixed symbolic libraries, COSINE further incorporates an outer-loop large language model that adaptively prunes and expands the hypothesis space using feedback from the inner optimization loop. Experiments on synthetic systems and large-scale real-world epidemic data demonstrate robust structural recovery and compact, mechanism-aligned dynamical expressions. Code: https://anonymous.4open.science/r/COSINE-6D43.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces COSINE, a differentiable co-optimization framework for jointly inferring latent interaction graphs and sparse symbolic dynamical equations from time-series observations. An inner optimization loop fits graph edges and symbolic expressions from a function library, while an outer-loop LLM adaptively prunes and expands that library using performance feedback. The central claim is that this yields robust structural recovery and compact, mechanism-aligned expressions on synthetic systems and large-scale epidemic data.
Significance. If the claims hold after addressing the noted gaps, the work would offer a meaningful advance in interpretable relational inference by relaxing the fixed-library assumption common in symbolic regression while retaining explicit dynamical forms. The joint graph-symbolic optimization and LLM-guided adaptation represent a novel combination with potential applicability to epidemic modeling and other interacting systems. Credit is due for providing anonymous code and for targeting both topology and mechanism discovery in one framework.
major comments (3)
- [Abstract] Abstract: the claim of 'robust structural recovery' is presented without any quantitative metrics, error bars, baseline comparisons, or ablation results, which is load-bearing for evaluating performance given the variability introduced by LLM guidance.
- [Method] Method section (outer LLM loop): the assumption that LLM feedback reliably selects an unbiased, high-quality function library is central to the interpretability benefit, yet no controls, inter-LLM consistency checks, prompt-robustness tests, or comparisons against fixed libraries are described; without these the reported compactness on epidemic data could be an artifact of LLM bias toward familiar or short expressions.
- [Experiments] Experiments: the abstract states results on synthetic systems and epidemic data but provides no details on library sizes, sparsity regularization values, or how LLM temperature/selection criteria were chosen, leaving the free parameters unexamined and the robustness claim difficult to reproduce or falsify.
minor comments (2)
- [Method] Notation for the joint objective and LLM feedback signal should be defined more explicitly with equations to clarify the inner/outer loop interaction.
- [Experiments] The epidemic dataset description would benefit from a table summarizing size, variables, and ground-truth interaction structure if available.
Simulated Author's Rebuttal
We thank the referee for their constructive feedback and positive assessment of COSINE's potential contribution to interpretable relational inference. We agree that strengthening the abstract, adding robustness controls for the LLM component, and providing explicit experimental details will improve clarity and reproducibility. Below we respond point-by-point to the major comments and indicate the planned revisions.
read point-by-point responses
-
Referee: [Abstract] Abstract: the claim of 'robust structural recovery' is presented without any quantitative metrics, error bars, baseline comparisons, or ablation results, which is load-bearing for evaluating performance given the variability introduced by LLM guidance.
Authors: We agree that the abstract would be stronger with concrete quantitative support for the 'robust structural recovery' claim. The Experiments section already reports these metrics (including error bars, baselines, and ablations), but the abstract summarizes at a high level without them. We will revise the abstract to include key results such as structural recovery accuracies with standard deviations and comparisons to baselines, making the performance claims directly evaluable. revision: yes
-
Referee: [Method] Method section (outer LLM loop): the assumption that LLM feedback reliably selects an unbiased, high-quality function library is central to the interpretability benefit, yet no controls, inter-LLM consistency checks, prompt-robustness tests, or comparisons against fixed libraries are described; without these the reported compactness on epidemic data could be an artifact of LLM bias toward familiar or short expressions.
Authors: This correctly highlights a gap in validating the LLM-guided adaptation. The manuscript presents the outer-loop LLM as enabling adaptive library pruning/expansion, but lacks explicit controls or comparisons. We will add a new subsection with: (i) direct comparisons of COSINE against fixed-library baselines, (ii) inter-run consistency results across multiple LLM calls, and (iii) prompt-sensitivity analysis. These will help confirm that compactness and performance gains are not due to LLM bias. revision: yes
-
Referee: [Experiments] Experiments: the abstract states results on synthetic systems and epidemic data but provides no details on library sizes, sparsity regularization values, or how LLM temperature/selection criteria were chosen, leaving the free parameters unexamined and the robustness claim difficult to reproduce or falsify.
Authors: We acknowledge the need for full hyperparameter transparency to support reproducibility. While the Methods and Experiments sections describe the overall library construction, optimization, and LLM integration, specific values (library sizes per dataset, sparsity regularization coefficients, LLM temperature, and selection criteria) for the epidemic experiments were not tabulated. We will add a detailed table and accompanying text specifying all these parameters, along with any sensitivity checks. revision: yes
Circularity Check
No significant circularity in derivation chain
full rationale
The paper introduces the COSINE framework as a joint optimization procedure combining graph inference with sparse symbolic regression, augmented by an outer LLM loop for adaptive library pruning based on inner-loop feedback. The abstract and method description present this as an algorithmic construction whose outputs (recovered graphs and expressions) are validated on held-out synthetic and real epidemic data. No equations, uniqueness theorems, or self-citations are invoked that would reduce the claimed structural recovery or mechanism alignment to a fitted parameter or input by construction. The LLM component is treated as an external oracle rather than a self-referential definition, and experimental claims remain falsifiable against independent benchmarks.
Axiom & Free-Parameter Ledger
free parameters (2)
- sparsity regularization weight
- LLM prompt temperature and selection criteria
axioms (1)
- domain assumption Observed trajectories are generated by a system whose interactions and dynamics admit a sparse symbolic representation.
Reference graph
Works this paper leans on
-
[1]
Brunton, S
Pmlr, 2021. Brunton, S. L., Proctor, J. L., and Kutz, J. N. Discovering governing equations from data by sparse identification of nonlinear dynamical systems.Proceedings of the national academy of sciences, 113(15):3932–3937, 2016a. Brunton, S. L., Proctor, J. L., and Kutz, J. N. Sparse iden- tification of nonlinear dynamics with control (sindyc). IFAC-Pa...
2021
-
[2]
Learningsymbolic physics with graph networks.arXiv:1909.05862,
World Scientific, 1999. Castellano, C., Fortunato, S., and Loreto, V . Statistical physics of social dynamics.Reviews of Modern Physics, 81(2):591, 2009. Chen, S., Wang, J., and Li, G. Neural relational inference with efficient message passing mechanisms. InProceed- ings of the AAAI Conference on Artificial Intelligence, volume 35, pp. 7055–7063, 2021. Cr...
-
[3]
KAN: Kolmogorov-Arnold Networks
Springer, 1975. Lai, Y .-C. Finding nonlinear system equations and complex network structures from data: A sparse optimization ap- proach.Chaos: An Interdisciplinary Journal of Nonlinear Science, 31(8), 2021. Li, Y ., Meng, C., Shahabi, C., and Liu, Y . Structure- informed graph auto-encoder for relational inference and simulation. InICML Workshop on Lear...
work page internal anchor Pith review arXiv 1975
-
[4]
Respond with a single JSON object and nothing else (no markdown or explanatory text)
-
[5]
message_terms
The JSON must contain exactly two array fields: "message_terms" and "update_terms". 13 Co-Optimization of Symbolic Interactions and Network Edges - Functions in message_terms are used to model interactions between nodes. - Functions in update_terms are used to model each node’s own state update
-
[6]
name", "expr
Each item must contain exactly three fields: {"name", "expr", "type"}, where type in {"vector", "scalar"}. - The field name must be exactly "expr". - No extra fields are allowed, no empty field is allowed
-
[7]
msg" or
Within the same stream, names must be unique. Available tensors and helpers: - Message stream only: xi, xj shape: [B, N, N, D]; diff = xj - xi - Update stream only: x in [B, N, D]; h in [B, N, D]; deg in [B, N, 1] - Shared helpers: torch, F, pi, inf Strict variable scoping rules (NO EXCEPTIONS): - Expressions in message_terms may only use: xi, xj, diff (a...
-
[8]
Message function (message_terms): - Start with simple and plausible interaction mechanisms (e.g., linear difference ‘xj - xi‘)
-
[9]
- Keep terms low-order (linear or quadratic) initially
Update function (update_terms): - Include core baseline terms (e.g., ‘x‘, ‘h‘) to represent basic self-dynamics and neighbor influence. - Keep terms low-order (linear or quadratic) initially
-
[10]
Multi-channel (D>1): - Use simple channel-wise terms before exploring cross-channel interactions. Advanced Exploration (if the system description suggests high complexity): - If the system is known to be highly non-linear (e.g., biological, chemical), you may include terms from the following categories: - Trigonometric: sin(.) / cos(.) on diff or state. -...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.