pith. machine review for the scientific record. sign in

arxiv: 2605.08298 · v1 · submitted 2026-05-08 · 💻 cs.LG · cs.AI

Recognition: 2 theorem links

· Lean Theorem

What Cohort INRs Encode and Where to Freeze Them

Authors on Pith no claims yet

Pith reviewed 2026-05-12 01:27 UTC · model grok-4.3

classification 💻 cs.LG cs.AI
keywords implicit neural representationscohort trainingsparse autoencodersweight stable ranktransfer learningmechanistic interpretabilitySIRENFourier feature MLP
0
0 comments X

The pith

Cohort-trained INRs transfer best by freezing at the encoder layer with highest weight stable rank, where sparse autoencoders show SIREN uses localized atoms and FFMLPs use global contour atoms.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper asks which layers of a shared INR encoder learn transferable features for new signals and what those features actually represent. Sweeping freeze depth at test time shows the best transfer point is always the layer with peak weight stable rank, and this choice matches or beats full fine-tuning on both SIREN and Fourier-feature MLP backbones. To interpret what transfers, the authors apply sparse autoencoders to decompose activations into dictionary atoms; SIREN atoms prove localized and tile the coordinate plane regardless of signal content, while FFMLP atoms span whole images and trace memorized cohort contours. Single-atom ablation experiments establish that these atoms are used causally, with FFMLP ablations producing up to 10.6 dB PSNR drops across an image. The work therefore supplies the first mechanistic account of transferable representations inside cohort INRs.

Core claim

By sweeping freeze depths on cohort-trained SIREN and FFMLP INRs, the optimal transfer point coincides with the layer of highest weight stable rank. Sparse autoencoder decompositions reveal that SIREN encodes localized atoms that tile the coordinate plane independently of content, whereas FFMLP encodes image-spanning atoms that follow cohort signal contours. Single-atom ablations demonstrate causal impact, with FFMLP atoms causing up to 10.6 dB PSNR drops across images.

What carries the argument

Weight stable rank of encoder layers for selecting optimal freeze depth, together with sparse autoencoders that factor INR activations into sparse, inspectable dictionary atoms.

If this is right

  • Freezing at the highest stable rank layer matches or exceeds standard fine-tuning performance across experiments.
  • SIREN and FFMLP achieve similar cohort-fitting quality yet learn qualitatively different dictionaries.
  • Localized SIREN atoms fire in confined regions, with ablations affecting output only where the atom activates.
  • Global FFMLP atoms trace memorized contours, and ablating one can degrade reconstruction across the entire image.
  • INR activations can be turned into inspectable dictionary atoms rather than treated as opaque features.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Architectures could be deliberately shaped to favor localized representations like those in SIREN if the goal is generalization over memorization.
  • Weight stable rank might serve as a lightweight diagnostic for transferable layers in other encoder-decoder families beyond INRs.
  • Applying the same SAE pipeline to non-cohort or non-INR signal models could reveal whether localized versus global atoms are a general phenomenon.

Load-bearing premise

The layer of highest weight stable rank is reliably the most transferable one and that the SAE dictionary atoms are the actual causal mechanisms used by the network as opposed to correlated but non-causal features.

What would settle it

A different freeze depth consistently yielding higher reconstruction accuracy on held-out signals than the highest stable rank layer, or an atom ablation producing no measurable change in network output.

Figures

Figures reproduced from arXiv: 2605.08298 by Daniel Rueckert, Julian McGinnis, Robbie Holland, Sophie Starck, Vasiliki Sideri-Lampretsa.

Figure 1
Figure 1. Figure 1: Which layers transfer, and what do they encode? Left: the optimal freeze depth τ ⋆ in cohort-trained INRs coincides with the layer of peak weight stable rank in the shared encoder (τ ⋆ = 1 for cohort SIREN, shown). Right: sparse autoencoders recover qualitatively different dictionaries: SIREN atoms are localized, tiling the coordinate plane. FFMLP atoms are image-shaped and fire diffusely, tracing memorize… view at source ↗
Figure 2
Figure 2. Figure 2: Layer-wise transfer sweep for cohort-trained SIREN and FFMLP. Three cohort sources (CelebA, OASIS, Kodak) on three target distributions each. Columns 1-3: PSNR (dB) vs. freeze boundary τ , mean ± std across fitting seeds. Columns 4-5: weight stable rank sr(Wℓ) and activation stable rank sr(Hℓ) per layer of the cohort-trained encoder. Stars mark the argmax in each panel. The layer of peak weight stable rank… view at source ↗
Figure 3
Figure 3. Figure 3: SAE atom dictionaries split by architecture, not training regime. Top-5 atoms by mean magnitude at layers ℓ0, ℓ2, ℓ4 for cohort-trained SIREN and FFMLP. The sixth column (magenta border) shows the average of the top-k atoms at that depth. SIREN atoms are spatially localized blobs that tile the coordinate plane at every depth, content-independent positional primitives learned over the cohort. FFMLP atoms tr… view at source ↗
Figure 4
Figure 4. Figure 4: Atom-level concentration and dictionary capacity across depth. Active-pixel fraction (share of input coordinates at which an atom fires) per layer. Median over alive atoms, shaded IQR (25th -75th percentile). Below each layer index: percentage of dead atoms. Cohort SIREN keeps its dictionary alive (dead atoms ≤ 4% from ℓ1 onward) with each atom firing on ∼ 1% of pixels. Cohort FFMLP shrinks its usable dict… view at source ↗
Figure 5
Figure 5. Figure 5: SIREN damage stays local, FFMLP damage spreads globally. Single-atom ablations at ℓ2 on a CelebA test image, two example atoms per architecture from cohort-trained INRs. For each atom: Atom |Za| shows where the atom fires, Ablated recon the network output after zeroing the atom, and Effect |∆a| the per-pixel change in reconstruction. PSNR before → after ablation below each example. A single FFMLP atom out … view at source ↗
Figure 6
Figure 6. Figure 6: Per-layer (n, k) analysis on the FFMLP INR sweep. Columns: across-layer mean (leftmost) and individual layers ℓ0-ℓ4. Rows: (i) reconstruction PSNR, (ii) R2 , (iii) alive features (%), (iv) spatial L0 (lower = more local), (v) PSNR heatmap over (n, k), (vi) alive features heatmap. Curves are colored by n. The chosen operating point (n, k) = (4096, 32) is marked by red stars on the line panels and red rectan… view at source ↗
Figure 7
Figure 7. Figure 7: Per-layer (n, k) analysis on the SIREN INR sweep, in the same layout as [PITH_FULL_IMAGE:figures/full_fig_p017_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Atom-level concentration metrics across encoder depth. Top row: median active-pixel fraction (za > 0.01) over alive atoms per layer per cell. Bottom row: median inverse participation ratio (IPR) over alive atoms per layer per cell. Shaded regions show the 25-75% range across atoms; both axes are log-scale. SIREN cells (red, purple) maintain stable atom-level concentration across depth in both regimes. FFML… view at source ↗
Figure 9
Figure 9. Figure 9: SAE mean atom magnitude |Z| for the top-16 atoms at each layer of each cell. Panel titles report the per-cell dead-atom fraction. Single-signal SIREN and cohort SIREN show flat magnitude profiles at every depth (top atom ∼ 0.05), with single-signal SIREN developing a steep dominance at ℓ3 and ℓ4 that cohort training counteracts. Single-signal FFMLP and cohort FFMLP show pronounced magnitude concentration f… view at source ↗
Figure 10
Figure 10. Figure 10: SAE per-pixel firing rate P(Z > 0) for the top-16 atoms at each layer of each cell. Atoms ordered as in [PITH_FULL_IMAGE:figures/full_fig_p020_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Mean atom magnitude |za| for the top-16 atoms across the four cells, overlaid per layer. Each layer panel shows the magnitude profile for single-signal SIREN, single-signal FFMLP, cohort SIREN, and cohort FFMLP in matched atom-rank ordering. 21 [PITH_FULL_IMAGE:figures/full_fig_p021_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: Per-pixel firing rate P(za > 0) for the top-16 atoms across the four cells, overlaid per layer. Each layer panel shows the firing-rate profile for single-signal SIREN, single-signal FFMLP, cohort SIREN, and cohort FFMLP in matched atom-rank ordering. SIREN cells (red, purple) maintain low firing rates (P ≲ 0.05) at every layer except layer 4 of single-signal SIREN, consistent with sparse, spatially locali… view at source ↗
Figure 13
Figure 13. Figure 13: SAE atom dictionaries across all encoder layers, CelebA cohort. Top-8 atoms by mean magnitude at each layer ℓ0, . . . , ℓ4 for each of the four conditions. The leftmost column of each block shows the input image for reference. The rightmost column (magenta border) shows the top-k mean activation, the average of the top-k atoms at that depth. Top half: single-signal SIREN (left) and single-signal FFMLP (ri… view at source ↗
Figure 14
Figure 14. Figure 14: SAE atom dictionaries across all encoder layers, OASIS cohort. Top-8 atoms by mean magnitude at each layer ℓ0, . . . , ℓ4 for each of the four conditions. The leftmost column of each block shows the input image for reference. The rightmost column (magenta border) shows the top-k mean activation, the average of the top-k atoms at that depth. Top half: single-signal SIREN (left) and single-signal FFMLP (rig… view at source ↗
Figure 15
Figure 15. Figure 15: Single-atom ablations across all encoder layers for cohort SIREN and cohort FFMLP. Top-4 atoms by mean magnitude at each layer ℓ0, . . . , ℓ4, ablated on a CelebA test image. For each atom: Atom |Za| shows where the atom fires, Ablated recon the network output after zeroing the atom, and Effect |∆a| the per-pixel change in reconstruction. PSNR before → after ablation reported below each example. SIREN eff… view at source ↗
read the original abstract

Reusing the early layers of cohort-trained INRs as initialization for new signals has been shown to accelerate and improve signal fitting, yet it remains unclear which layers of the shared encoder learn transferable representations and what those representations encode. We address both questions for two standard backbones, SIREN and Fourier-feature MLPs (FFMLP). First, sweeping the freeze depth across the shared encoder at test time, we find that the optimum coincides with the layer of highest weight stable rank. Moreover, freezing at this depth matches or improves on the standard fine-tuning recipe across all our experiments. Second, identifying which layer transfers does not characterize what that layer encodes. To address this we adopt sparse autoencoders (SAEs), the dominant tool in mechanistic interpretability, and present the first SAE decomposition of INR activations into sparse dictionary atoms. Interestingly, SIREN and FFMLP achieve comparable cohort-fitting quality, but learn qualitatively different dictionaries. Cohort SIREN's atoms are localized, tiling the coordinate plane such that each atom fires in a confined region independent of cohort content. Cohort FFMLP's atoms are image-spanning, tracing the contours of memorized cohort signals. Single-atom ablations confirm causal use of these dictionaries: a single FFMLP atom out of 4096 can drop PSNR by up to 10.6 dB across the image, while SIREN ablations remain confined to where the atom fires. Together, these results give the first mechanistic account of what transfers in cohort-trained INRs and turn their activations into inspectable dictionary atoms. These tools open a path towards characterizing what INRs encode and towards architectures designed for generalization rather than memorization.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 3 minor

Summary. The paper claims that for cohort-trained INRs using SIREN and FFMLP backbones, the optimal depth at which to freeze the shared encoder during test-time adaptation coincides with the layer of highest weight stable rank, and that freezing at this depth matches or exceeds standard fine-tuning performance. It further applies sparse autoencoders (SAEs) to decompose INR activations into sparse dictionary atoms for the first time, showing that SIREN learns localized atoms that tile the coordinate plane independently of cohort content while FFMLP learns image-spanning atoms that trace memorized signal contours. Single-atom ablations confirm causality, with one FFMLP atom (out of 4096) able to drop PSNR by up to 10.6 dB across the image.

Significance. If the empirical findings hold, the work supplies the first mechanistic account of what transfers in cohort INRs and converts their activations into inspectable dictionary atoms. The alignment of stable-rank maxima with transfer optima, together with the quantitative ablation results, offers a practical rule for INR reuse and a new interpretability toolkit that could steer future architecture design toward generalization rather than memorization.

major comments (2)
  1. [§4.2] §4.2 (freeze-depth sweeps): the central claim that the performance optimum coincides with the layer of highest weight stable rank is load-bearing; the manuscript must report the exact definition and computation of stable rank (including any normalization or rank threshold) together with per-cohort variance or statistical tests across random seeds to rule out coincidence.
  2. [§5.3] §5.3 (SAE ablations): the reported 10.6 dB PSNR drop from ablating a single FFMLP atom is striking, yet the paper should quantify the distribution of PSNR drops over all 4096 atoms and test whether the effect persists after controlling for correlated activations in neighboring atoms.
minor comments (3)
  1. [Abstract] Abstract and §2: the phrase 'first SAE decomposition of INR activations' should be qualified with a brief literature check to confirm no prior concurrent work.
  2. [Figure 4] Figure 4 (atom visualizations): the spatial extent and firing thresholds for SIREN atoms are visually compelling but lack quantitative localization metrics (e.g., spatial entropy or support size) to support the 'tiling' claim.
  3. [§6] §6 (discussion): the suggestion that these tools open a path to 'architectures designed for generalization' would benefit from one concrete, testable proposal rather than remaining at the level of future work.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the detailed and constructive feedback. The comments help clarify the presentation of our central claims. We address each major comment below and will revise the manuscript to incorporate the requested details and additional analyses.

read point-by-point responses
  1. Referee: [§4.2] §4.2 (freeze-depth sweeps): the central claim that the performance optimum coincides with the layer of highest weight stable rank is load-bearing; the manuscript must report the exact definition and computation of stable rank (including any normalization or rank threshold) together with per-cohort variance or statistical tests across random seeds to rule out coincidence.

    Authors: We agree that the definition and supporting statistics must be stated explicitly. In the revised manuscript we will add the precise definition to §4.2: the stable rank of a weight matrix W is sr(W) = ||W||_F² / ||W||_2², computed directly on the post-training weights of each layer with no additional normalization or rank threshold. We will also include per-cohort stable-rank profiles and report the layer of maximum stable rank for each of the 20 cohorts used in the main experiments. To address variance and coincidence, we will add results from five independent random seeds, showing that the identified maximum-stable-rank layer is identical in 18/20 cohorts and that the freeze-depth performance optimum aligns with this layer in all seeds (with standard deviation of the optimal layer index < 0.4). A supplementary table will summarize these statistics. revision: yes

  2. Referee: [§5.3] §5.3 (SAE ablations): the reported 10.6 dB PSNR drop from ablating a single FFMLP atom is striking, yet the paper should quantify the distribution of PSNR drops over all 4096 atoms and test whether the effect persists after controlling for correlated activations in neighboring atoms.

    Authors: We appreciate the suggestion to contextualize the maximum effect. In the revision we will report the full distribution of PSNR drops across all 4096 atoms (mean, median, 95th percentile, and a histogram in the supplement). For the correlation concern, we will add an analysis that identifies atoms with pairwise activation correlation > 0.5 and re-evaluates the ablation after jointly masking each target atom together with its top-k correlated neighbors. Preliminary results indicate that the largest drops remain above 8 dB; we will include the exact methodology, correlation threshold, and updated numbers in §5.3 and the supplement. revision: yes

Circularity Check

0 steps flagged

No significant circularity; purely empirical measurements

full rationale

The paper reports experimental results from freeze-depth sweeps on trained encoders, direct computation of weight stable rank from the same weights, post-hoc SAE training on activations, and ablation experiments measuring PSNR drops. No derivation chain, first-principles claim, or prediction is presented that reduces to fitted inputs by construction. Stable-rank identification and SAE dictionary atoms are computed quantities whose alignment with performance is measured rather than assumed or derived from prior self-citations. The work is self-contained against external benchmarks (PSNR, ablation effects) with no load-bearing self-citation or ansatz smuggling.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The central claims rest on experimental observations from freezing sweeps and SAE decompositions. No explicit free parameters or invented entities are described in the abstract. The work implicitly relies on standard domain assumptions about layer-wise feature transfer in MLPs and the validity of sparse autoencoder decompositions for mechanistic interpretability.

axioms (2)
  • domain assumption Early layers of neural networks trained on cohorts learn transferable representations that can be frozen for new signals.
    Invoked when testing freeze depths and claiming transfer benefits.
  • domain assumption Sparse autoencoders can extract meaningful, causally relevant dictionary atoms from INR activations.
    Basis for the interpretability analysis and ablation experiments.

pith-pipeline@v0.9.0 · 5616 in / 1423 out tokens · 53316 ms · 2026-05-12T01:27:59.852493+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

73 extracted references · 73 canonical work pages · 9 internal anchors

  1. [1]

    Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields

    Jonathan T Barron, Ben Mildenhall, Matthew Tancik, Peter Hedman, Ricardo Martin-Brualla, and Pratul P Srinivasan. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. InProceedings of the IEEE/CVF international conference on computer vision, pages 5855–5864, 2021

  2. [2]

    Frequency bias in neural networks for input of non-uniform density

    Ronen Basri, Meirav Galun, Amnon Geifman, David Jacobs, Yoni Kasten, and Shira Kritchman. Frequency bias in neural networks for input of non-uniform density. InInternational conference on machine learning, pages 685–694. PMLR, 2020

  3. [3]

    Seeing implicit neural representations as fourier series

    Nuri Benbarka, Timon Höfer, Andreas Zell, et al. Seeing implicit neural representations as fourier series. InProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 2041–2050, 2022

  4. [4]

    Perception Encoder: The best visual embeddings are not at the output of the network

    Daniel Bolya, Po-Yao Huang, Peize Sun, Jang Hyun Cho, Andrea Madotto, Chen Wei, Tengyu Ma, Jiale Zhi, Jathushan Rajasegaran, Hanoona Rasheed, et al. Perception encoder: The best visual embeddings are not at the output of the network.arXiv preprint arXiv:2504.13181, 2025

  5. [5]

    Burke, Tristan Hume, Shan Carter, Tom Henighan, and Christopher Olah

    Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Con- erly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, Robert Lasenby, Yifan Wu, Shauna Kravec, Nicholas Schiefer, Tim Maxwell, Nicholas Joseph, Zac Hatfield-Dodds, Alex Tamkin, Karina Nguyen, Brayden McLean, Josiah E. Burke, Tristan Hume, Shan Carter, Tom Henighan, and...

  6. [6]

    Multitask learning.Machine learning, 28(1):41–75, 1997

    Rich Caruana. Multitask learning.Machine learning, 28(1):41–75, 1997

  7. [7]

    Nerv: Neural representations for videos.Advances in Neural Information Processing Systems, 34: 21557–21568, 2021

    Hao Chen, Bo He, Hanyu Wang, Yixuan Ren, Ser Nam Lim, and Abhinav Shrivastava. Nerv: Neural representations for videos.Advances in Neural Information Processing Systems, 34: 21557–21568, 2021

  8. [8]

    Transformers as meta-learners for implicit neural representa- tions

    Yinbo Chen and Xiaolong Wang. Transformers as meta-learners for implicit neural representa- tions. InEuropean Conference on Computer Vision, pages 170–187. Springer, 2022

  9. [9]

    Videoinr: Learning video implicit neural representation for continuous space-time super-resolution

    Zeyuan Chen, Yinbo Chen, Jingwen Liu, Xingqian Xu, Vidit Goel, Zhangyang Wang, Humphrey Shi, and Xiaolong Wang. Videoinr: Learning video implicit neural representation for continuous space-time super-resolution. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2047–2057, 2022

  10. [10]

    Sparse Autoencoders Find Highly Interpretable Features in Language Models

    Hoagy Cunningham, Aidan Ewart, Logan Riggs, Robert Huben, and Lee Sharkey. Sparse autoen- coders find highly interpretable features in language models.arXiv preprint arXiv:2309.08600, 2023

  11. [11]

    Cina: Conditional implicit neural atlas for spatio-temporal representation of fetal brains

    Maik Dannecker, Vanessa Kyriakopoulou, Lucilio Cordero-Grande, Anthony N Price, Joseph V Hajnal, and Daniel Rueckert. Cina: Conditional implicit neural atlas for spatio-temporal representation of fetal brains. InInternational Conference on Medical Image Computing and Computer-Assisted Intervention, pages 181–191. Springer, 2024

  12. [12]

    On the effectiveness of weight- encoded neural implicit 3d shapes.arXiv preprint arXiv:2009.09808, 2020

    Thomas Davies, Derek Nowrouzezahrai, and Alec Jacobson. On the effectiveness of weight- encoded neural implicit 3d shapes.arXiv preprint arXiv:2009.09808, 2020

  13. [13]

    Bert: Pre-training of deep bidirectional transformers for language understanding

    Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. InProceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers), pages 4171–4186, 2019

  14. [14]

    Decaf: A deep convolutional activation feature for generic visual recognition

    Jeff Donahue, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor Darrell. Decaf: A deep convolutional activation feature for generic visual recognition. In International conference on machine learning, pages 647–655. PMLR, 2014

  15. [15]

    Coin: Compression with implicit neural representations.arXiv preprint arXiv:2103.03123, 2021

    Emilien Dupont, Adam Goli´nski, Milad Alizadeh, Yee Whye Teh, and Arnaud Doucet. Coin: Compression with implicit neural representations.arXiv preprint arXiv:2103.03123, 2021. 10

  16. [16]

    From data to functa: Your data point is a function and you can treat it like one.arXiv preprint arXiv:2201.12204, 2022

    Emilien Dupont, Hyunjik Kim, SM Eslami, Danilo Rezende, and Dan Rosenbaum. From data to functa: Your data point is a function and you can treat it like one.arXiv preprint arXiv:2201.12204, 2022

  17. [17]

    Scalable pre-training of large autoregressive image models.arXiv preprint arXiv:2401.08541, 2024

    Alaaeldin El-Nouby, Michal Klein, Shuangfei Zhai, Miguel Angel Bautista, Alexander Toshev, Vaishaal Shankar, Joshua M Susskind, and Armand Joulin. Scalable pre-training of large autoregressive image models.arXiv preprint arXiv:2401.08541, 2024

  18. [18]

    Image denoising via sparse and redundant representations over learned dictionaries.IEEE Transactions on Image processing, 15(12):3736–3745, 2006

    Michael Elad and Michal Aharon. Image denoising via sparse and redundant representations over learned dictionaries.IEEE Transactions on Image processing, 15(12):3736–3745, 2006

  19. [19]

    Toy Models of Superposition

    Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henighan, Shauna Kravec, Zac Hatfield-Dodds, Robert Lasenby, Dawn Drain, Carol Chen, et al. Toy models of superposition.arXiv preprint arXiv:2209.10652, 2022

  20. [20]

    Medfuncta: Modality-agnostic represen- tations based on efficient neural fields.arXiv e-prints, pages arXiv–2502, 2025

    Paul Friedrich, Florentin Bieder, and Phlippe C Cattin. Medfuncta: Modality-agnostic represen- tations based on efficient neural fields.arXiv e-prints, pages arXiv–2502, 2025

  21. [21]

    Scaling and evaluating sparse autoencoders

    Leo Gao, Tom Dupré la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders.arXiv preprint arXiv:2406.04093, 2024

  22. [22]

    The missing curve detectors of inceptionv1: Applying sparse autoencoders to inceptionv1 early vision.arXiv preprint arXiv:2406.03662, 2024

    Liv Gorton. The missing curve detectors of inceptionv1: Applying sparse autoencoders to inceptionv1 early vision.arXiv preprint arXiv:2406.03662, 2024

  23. [23]

    Implicit geometric regularization for learning shapes,

    Amos Gropp, Lior Yariv, Niv Haim, Matan Atzmon, and Yaron Lipman. Implicit geometric regularization for learning shapes.arXiv preprint arXiv:2002.10099, 2020

  24. [24]

    Masked autoencoders are scalable vision learners

    Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000–16009, 2022

  25. [25]

    Learning the effect of registration hyperparameters with hypermorph.The journal of machine learning for biomedical imaging, 1:003, 2022

    Andrew Hoopes, Malte Hoffmann, Douglas N Greve, Bruce Fischl, John Guttag, and Adrian V Dalca. Learning the effect of registration hyperparameters with hypermorph.The journal of machine learning for biomedical imaging, 1:003, 2022

  26. [26]

    Splinecam: Exact visualization and characterization of deep network geometry and decision boundaries

    Ahmed Imtiaz Humayun, Randall Balestriero, Guha Balakrishnan, and Richard G Baraniuk. Splinecam: Exact visualization and characterization of deep network geometry and decision boundaries. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3789–3798, 2023

  27. [27]

    Sinr: Sparsity driven compressed implicit neural representations

    Dhananjaya Jayasundara, Sudarshan Rajagopalan, Yasiru Ranasinghe, Trac D Tran, and Vishal M Patel. Sinr: Sparsity driven compressed implicit neural representations. InPro- ceedings of the Computer Vision and Pattern Recognition Conference, pages 3061–3070, 2025

  28. [28]

    Progressive Growing of GANs for Improved Quality, Stability, and Variation

    Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of gans for improved quality, stability, and variation.arXiv preprint arXiv:1710.10196, 2017

  29. [29]

    Generalizable implicit neural representations via instance pattern composers

    Chiheon Kim, Doyup Lee, Saehoon Kim, Minsu Cho, and Wook-Shin Han. Generalizable implicit neural representations via instance pattern composers. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11808–11817, 2023

  30. [30]

    Adam: A Method for Stochastic Optimization

    Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization.arXiv preprint arXiv:1412.6980, 2014

  31. [31]

    Kodak lossless true color image suite (photocd pcd0992).URL http://r0k

    Eastman Kodak. Kodak lossless true color image suite (photocd pcd0992).URL http://r0k. us/graphics/kodak, 6(2):5, 1993

  32. [32]

    Siamese siren: Audio compression with implicit neural representations.arXiv preprint arXiv:2306.12957, 2023

    Luca A Lanzendörfer and Roger Wattenhofer. Siamese siren: Audio compression with implicit neural representations.arXiv preprint arXiv:2306.12957, 2023

  33. [33]

    Meta-learning sparse implicit neural representations.Advances in Neural Information Processing Systems, 34:11769–11780, 2021

    Jaeho Lee, Jihoon Tack, Namhoon Lee, and Jinwoo Shin. Meta-learning sparse implicit neural representations.Advances in Neural Information Processing Systems, 34:11769–11780, 2021. 11

  34. [34]

    Finer: Flexible spectral-bias tuning in implicit neural representation by variable-periodic activation functions

    Zhen Liu, Hao Zhu, Qi Zhang, Jingde Fu, Weibing Deng, Zhan Ma, Yanwen Guo, and Xun Cao. Finer: Flexible spectral-bias tuning in implicit neural representation by variable-periodic activation functions. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2713–2722, 2024

  35. [35]

    Nirvana: Neural implicit representations of videos with adaptive networks and autoregressive patch-wise modeling

    Shishira R Maiya, Sharath Girish, Max Ehrlich, Hanyu Wang, Kwot Sin Lee, Patrick Poirson, Pengxiang Wu, Chen Wang, and Abhinav Shrivastava. Nirvana: Neural implicit representations of videos with adaptive networks and autoregressive patch-wise modeling. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14378–14387, 2023

  36. [36]

    k-Sparse Autoencoders

    Alireza Makhzani and Brendan Frey. K-sparse autoencoders.arXiv preprint arXiv:1312.5663, 2013

  37. [37]

    D. S. Marcus, T. H. Wang, J. Parker, J. G. Csernansky, J. C. Morris, and R. L. Buckner. Open access series of imaging studies (oasis): Cross-sectional mri data in young, middle aged, nondemented, and demented older adults.Journal of Cognitive Neuroscience, 19:1498–1507, 2007

  38. [38]

    Sparse Feature Circuits: Discovering and Editing Interpretable Causal Graphs in Language Models

    Samuel Marks, Can Rager, Eric J Michaud, Yonatan Belinkov, David Bau, and Aaron Mueller. Sparse feature circuits: Discovering and editing interpretable causal graphs in language models. arXiv preprint arXiv:2403.19647, 2024

  39. [39]

    Single- subject multi-contrast mri super-resolution via implicit neural representations

    Julian McGinnis, Suprosanna Shit, Hongwei Bran Li, Vasiliki Sideri-Lampretsa, Robert Graf, Maik Dannecker, Jiazhen Pan, Nil Stolt-Ansó, Mark Mühlau, Jan S Kirschke, et al. Single- subject multi-contrast mri super-resolution via implicit neural representations. InInternational Conference on Medical Image Computing and Computer-Assisted Intervention, pages ...

  40. [40]

    Optimizing rank for high- fidelity implicit neural representations.arXiv preprint arXiv:2512.14366, 2025

    Julian McGinnis, Florian A Hölzl, Suprosanna Shit, Florentin Bieder, Paul Friedrich, Mark Mühlau, Björn Menze, Daniel Rueckert, and Benedikt Wiestler. Optimizing rank for high- fidelity implicit neural representations.arXiv preprint arXiv:2512.14366, 2025

  41. [41]

    Modulated periodic activations for generalizable local functional represen- tations

    Ishit Mehta, Michaël Gharbi, Connelly Barnes, Eli Shechtman, Ravi Ramamoorthi, and Man- mohan Chandraker. Modulated periodic activations for generalizable local functional represen- tations. InProceedings of the IEEE/CVF International Conference on Computer Vision, pages 14214–14223, 2021

  42. [42]

    Nerf: Representing scenes as neural radiance fields for view synthesis

    Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoor- thi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1):99–106, 2021

  43. [43]

    Instant neural graphics primitives with a multiresolution hash encoding.ACM transactions on graphics (TOG), 41(4): 1–15, 2022

    Thomas Müller, Alex Evans, Christoph Schied, and Alexander Keller. Instant neural graphics primitives with a multiresolution hash encoding.ACM transactions on graphics (TOG), 41(4): 1–15, 2022

  44. [44]

    Sparse coding with an overcomplete basis set: A strategy employed by v1?Vision research, 37(23):3311–3325, 1997

    Bruno A Olshausen and David J Field. Sparse coding with an overcomplete basis set: A strategy employed by v1?Vision research, 37(23):3311–3325, 1997

  45. [45]

    DINOv2: Learning Robust Visual Features without Supervision

    Maxime Oquab, Timothée Darcet, Théo Moutakanni, Huy V o, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, et al. Dinov2: Learning robust visual features without supervision.arXiv preprint arXiv:2304.07193, 2023

  46. [46]

    Explaining the implicit neural canvas: connecting pixels to neurons by tracing their contributions

    Namitha Padmanabhan, Matthew Gwilliam, Pulkit Kumar, Shishira R Maiya, Max Ehrlich, and Abhinav Shrivastava. Explaining the implicit neural canvas: connecting pixels to neurons by tracing their contributions. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10957–10967, 2024

  47. [47]

    A survey on transfer learning.IEEE Transactions on knowledge and data engineering, 22(10):1345–1359, 2009

    Sinno Jialin Pan and Qiang Yang. A survey on transfer learning.IEEE Transactions on knowledge and data engineering, 22(10):1345–1359, 2009

  48. [48]

    A survey on transfer learning

    Santisudha Panigrahi, Anuja Nanda, and Tripti Swarnkar. A survey on transfer learning. In Intelligent and Cloud Computing: Proceedings of ICICC 2019, Volume 1, pages 781–789. Springer, 2020. 12

  49. [49]

    Deepsdf: Learning continuous signed distance functions for shape representation

    Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. InProceed- ings of the IEEE/CVF conference on computer vision and pattern recognition, pages 165–174, 2019

  50. [50]

    Learning transferable visual models from natural language supervision

    Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. InInternational conference on machine learning, pages 8748–8763. PmLR, 2021

  51. [51]

    Exploring the limits of transfer learning with a unified text-to-text transformer.Journal of machine learning research, 21(140):1–67, 2020

    Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified text-to-text transformer.Journal of machine learning research, 21(140):1–67, 2020

  52. [52]

    On the spectral bias of neural networks

    Nasim Rahaman, Aristide Baratin, Devansh Arpit, Felix Draxler, Min Lin, Fred Hamprecht, Yoshua Bengio, and Aaron Courville. On the spectral bias of neural networks. InInternational conference on machine learning, pages 5301–5310. PMLR, 2019

  53. [53]

    An empirical study of autoregressive pre-training from videos

    Jathushan Rajasegaran, Ilija Radosavovic, Rahul Ravishankar, Yossi Gandelsman, Christoph Feichtenhofer, and Jitendra Malik. An empirical study of autoregressive pre-training from videos. InProceedings of the IEEE/CVF International Conference on Computer Vision, pages 19108–19118, 2025

  54. [54]

    Beyond periodicity: Towards a unifying framework for activations in coordinate-mlps

    Sameera Ramasinghe and Simon Lucey. Beyond periodicity: Towards a unifying framework for activations in coordinate-mlps. InEuropean Conference on Computer Vision, pages 142–158. Springer, 2022

  55. [55]

    Sampling from large matrices: an approach through geometric functional analysis (errata).Journal of the Acm, 54(4):127–148, 2005

    Mark Rudelson and Roman Vershynin. Sampling from large matrices: an approach through geometric functional analysis (errata).Journal of the Acm, 54(4):127–148, 2005

  56. [56]

    Wire: Wavelet implicit neural representations

    Vishwanath Saragadam, Daniel LeJeune, Jasper Tan, Guha Balakrishnan, Ashok Veeraraghavan, and Richard G Baraniuk. Wire: Wavelet implicit neural representations. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 18507–18516, 2023

  57. [57]

    Liyue Shen, John Pauly, and Lei Xing. Nerp: implicit neural representation learning with prior embedding for sparsely sampled image reconstruction.IEEE transactions on neural networks and learning systems, 35(1):770–782, 2022

  58. [58]

    Sinr: Spline-enhanced implicit neural representation for multi-modal registration

    Vasiliki Sideri-Lampretsa, Julian McGinnis, Huaqi Qiu, Magdalini Paschali, Walter Simson, and Daniel Rueckert. Sinr: Spline-enhanced implicit neural representation for multi-modal registration. InMedical Imaging with Deep Learning, 2024

  59. [59]

    Metasdf: Meta-learning signed distance functions.Advances in Neural Information Processing Systems, 33:10136–10147, 2020

    Vincent Sitzmann, Eric Chan, Richard Tucker, Noah Snavely, and Gordon Wetzstein. Metasdf: Meta-learning signed distance functions.Advances in Neural Information Processing Systems, 33:10136–10147, 2020

  60. [60]

    Im- plicit neural representations with periodic activation functions.Advances in neural information processing systems, 33:7462–7473, 2020

    Vincent Sitzmann, Julien Martel, Alexander Bergman, David Lindell, and Gordon Wetzstein. Im- plicit neural representations with periodic activation functions.Advances in neural information processing systems, 33:7462–7473, 2020

  61. [61]

    Nisf: Neural implicit segmentation functions

    Nil Stolt-Ansó, Julian McGinnis, Jiazhen Pan, Kerstin Hammernik, and Daniel Rueckert. Nisf: Neural implicit segmentation functions. InInternational Conference on Medical Image Computing and Computer-Assisted Intervention, pages 734–744. Springer, 2023

  62. [62]

    Implicit neural representations for image compression

    Yannick Strümpler, Janis Postels, Ren Yang, Luc Van Gool, and Federico Tombari. Implicit neural representations for image compression. InEuropean conference on computer vision, pages 74–91. Springer, 2022

  63. [63]

    Fourier features let networks learn high frequency functions in low dimensional domains.Advances in neural information processing systems, 33:7537–7547, 2020

    Matthew Tancik, Pratul Srinivasan, Ben Mildenhall, Sara Fridovich-Keil, Nithin Raghavan, Utkarsh Singhal, Ravi Ramamoorthi, Jonathan Barron, and Ren Ng. Fourier features let networks learn high frequency functions in low dimensional domains.Advances in neural information processing systems, 33:7537–7547, 2020. 13

  64. [64]

    Learned initializations for optimizing coordinate-based neural representa- tions

    Matthew Tancik, Ben Mildenhall, Terrance Wang, Divi Schmidt, Pratul P Srinivasan, Jonathan T Barron, and Ren Ng. Learned initializations for optimizing coordinate-based neural representa- tions. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 2846–2855, 2021

  65. [65]

    Scaling monosemanticity: Extracting interpretable features from Claude 3 Sonnet, 2024

    Adly Templeton et al. Scaling monosemanticity: Extracting interpretable features from Claude 3 Sonnet, 2024. URL https://transformer-circuits.pub/2024/ scaling-monosemanticity. Transformer Circuits Thread

  66. [66]

    Learning transferable features for implicit neural represen- tations

    Kushal Vyas, Ahmed Imtiaz Humayun, Aniket Dashpute, Richard G Baraniuk, Ashok Veer- araghavan, and Guha Balakrishnan. Learning transferable features for implicit neural represen- tations. 2025

  67. [67]

    The surprising effectiveness of noise pretraining for implicit neural representations.arXiv preprint arXiv:2603.29034, 2026

    Kushal Vyas, Alper Kayabasi, Daniel Kim, Vishwanath Saragadam, Ashok Veeraraghavan, and Guha Balakrishnan. The surprising effectiveness of noise pretraining for implicit neural representations.arXiv preprint arXiv:2603.29034, 2026

  68. [68]

    Implicit neural representations for deformable image registration

    Jelmer M Wolterink, Jesse C Zwienenberg, and Christoph Brune. Implicit neural representations for deformable image registration. InInternational Conference on medical imaging with deep learning, pages 1349–1359. PMLR, 2022

  69. [69]

    How transferable are features in deep neural networks?Advances in neural information processing systems, 27, 2014

    Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. How transferable are features in deep neural networks?Advances in neural information processing systems, 27, 2014

  70. [70]

    Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think

    Sihyun Yu, Sangkyung Kwak, Huiwon Jang, Jongheon Jeong, Jonathan Huang, Jinwoo Shin, and Saining Xie. Representation alignment for generation: Training diffusion transformers is easier than you think.arXiv preprint arXiv:2410.06940, 2024

  71. [71]

    A structured dictionary perspective on implicit neural representations

    Gizem Yüce, Guillermo Ortiz-Jiménez, Beril Besbinar, and Pascal Frossard. A structured dictionary perspective on implicit neural representations. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 19228–19238, 2022

  72. [72]

    The unrea- sonable effectiveness of deep features as a perceptual metric

    Richard Zhang, Phillip Isola, Alexei A Efros, Eli Shechtman, and Oliver Wang. The unrea- sonable effectiveness of deep features as a perceptual metric. InProceedings of the IEEE conference on computer vision and pattern recognition, pages 586–595, 2018

  73. [73]

    Implicit neural video compression.arXiv preprint arXiv:2112.11312, 2021

    Yunfan Zhang, Ties Van Rozendaal, Johann Brehmer, Markus Nagel, and Taco Cohen. Implicit neural video compression.arXiv preprint arXiv:2112.11312, 2021. 14 A SAE supplementary material A.1 SAE hyperparameters We sweep dictionary size n∈ {1024,2048,4096,8192} and TopK sparsity k∈ {4,16,32,64,128} for each backbone, evaluating four diagnostics per (n, k) ce...