pith. machine review for the scientific record. sign in

arxiv: 2605.12649 · v1 · submitted 2026-05-12 · 💻 cs.CV

Recognition: unknown

DIVER:Diving Deeper into Distilled Data via Expressive Semantic Recovery

Authors on Pith no claims yet

Pith reviewed 2026-05-14 21:11 UTC · model grok-4.3

classification 💻 cs.CV
keywords dataset distillationdiffusion modelssemantic recoverycross-architecture generalizationlatent space filteringdata synthesiscomputer vision
0
0 comments X

The pith

A dual-stage framework uses a pre-trained diffusion model to recover expressive semantics from distilled datasets and improve performance across different neural architectures.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Dataset distillation creates compact proxy datasets but prior single-stage methods overfit to one architecture and lose intrinsic semantics. DIVER adds a second stage that runs distilled images through a pre-trained diffusion model. Semantic inheritance moves high-level semantics into the latent space to remove architecture-specific noise. Semantic guidance and fusion then steer the reverse diffusion process only in its concrete phase to restore semantics without introducing artifacts. The result is distilled data that trains models of many architectures more effectively while keeping memory use low.

Core claim

DIVER performs semantic inheritance to embed high-level semantics of abstract distilled images into latent space, applies semantic guidance to direct the reverse diffusion procedure, and restricts semantic fusion to the concrete phase of the reverse process so that architecture-specific noise is filtered while original semantics are preserved.

What carries the argument

The three-step semantic recovery process (inheritance into latent space, guidance of reverse diffusion, and late-stage fusion) that filters architecture-specific noise using a pre-trained diffusion model.

If this is right

  • Distilled datasets become usable across a wider range of model architectures without retraining the distillation step.
  • The same compact dataset can support privacy-preserving training on both convolutional and transformer-based networks.
  • Processing overhead stays comparable to a single forward pass of a diffusion transformer on 256x256 images.
  • Memory consumption remains under 4 GB, allowing the method to run on modest hardware.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same latent-space filtering step might be applied to other forms of synthetic data such as generated images or text embeddings to remove model-specific artifacts.
  • Late-stage semantic fusion could be tested as a general technique for stabilizing guidance in any conditional diffusion process.
  • If the assumption holds, the method suggests a route to architecture-agnostic dataset distillation that does not require joint optimization over multiple target models.

Load-bearing premise

The pre-trained diffusion model can separate architecture-specific noise from the intrinsic semantics of the distilled images in latent space.

What would settle it

Training a different architecture on DIVER-processed distilled data yields no accuracy gain over the original single-stage distilled data on the same task.

Figures

Figures reproduced from arXiv: 2605.12649 by Guoming Lu, Jiawei Du, Jielei Wang, Qianxin Xia, Wenbo Jiang, Zhiyong Shu.

Figure 1
Figure 1. Figure 1: Comparison between classical single-stage DD and our proposed dual-stage DIVER. In stage I, DIVER is the same as classical DD. Mainly in stage II, we employ the pre-trained generative model to directly refine the distilled dataset, thereby synthesizing a new dataset termed synthetic dataset, significantly enhancing the generalization capabilities of traditional techniques across various paradigms. The quan… view at source ↗
Figure 2
Figure 2. Figure 2: The overview of DIVER. Semantic inheritance filters out architecture-specific “noise” and distills high-level semantics of distilled images into the latent space, retaining the initial semantics. Semantic guidance enhances the preservation of original semantics by directing the sampling procedure to generate realistic and informative images. Semantic fusion fuses conditional labels with inherited and guide… view at source ↗
Figure 3
Figure 3. Figure 3: Semantic evolution of the entire process. 3.3.2. SEMANTIC GUIDANCE During the image synthesis phase, our goal is to generate a high-quality image that satisfies the semantics of the specific distilled image. However, due to the continuous injection of conditional label information during the reverse process, la￾tent code using semantic inheritance inevitably suffers from information degradation. To compens… view at source ↗
Figure 4
Figure 4. Figure 4: (Left) The effect of guidance factor on performance on ImageYellow (IPC=10) with EDF. (Medium) The effect of applying different forward steps to the inherited latent on performance on ImageFruit (IPC=10) with NCFM (only SI). (Right) Performance of DD (MTT), GLaD and our DIVER under IPC 1 on the specific ConvNet and across heterogeneous architectures [PITH_FULL_IMAGE:figures/full_fig_p008_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: (Left) Comparison of synthetic images from different methods with and without our DIVER on ImageFruit. Our approach recovers expressive semantics and is more realistic. (Right) Comparison of images generated with (semantic-phase fusion) and without (full-phase fusion) our SF on ImageMeow. SF enhances category clarity and fidelity. Original Images Distilled Images Reconstructed Images Our Synthetic Images (… view at source ↗
Figure 6
Figure 6. Figure 6: t-SNE visualization of original images, distilled images, reconstructed images (obtained through direct VAE encoding and decoding without the diffusion), and our synthetic images. The results are presented for the first class of ImageNet under 10 IPC. For our method, SI only needs to encode distilled images into latent codes, incurring negligible computational over￾head. SG employs Eqn. 8 to compute gradie… view at source ↗
Figure 7
Figure 7. Figure 7: Comparison of distilled images, reconstructed images (obtained through direct VAE encoding and decoding without DiT), and our synthetic images with DM(left) and MTT(right) [PITH_FULL_IMAGE:figures/full_fig_p018_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: The pixel difference between the images synthesized by adjacent forward steps. 18 [PITH_FULL_IMAGE:figures/full_fig_p018_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Visualization comparison between raw DD and DIVER on ImageFruit [PITH_FULL_IMAGE:figures/full_fig_p019_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Visualization comparison between raw DD and DIVER on ImageWoof. 19 [PITH_FULL_IMAGE:figures/full_fig_p019_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Visualization comparison between raw DD and DIVER on ImageMeow [PITH_FULL_IMAGE:figures/full_fig_p020_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: Visualization comparison between raw DD and DIVER on ImageSquawk. 20 [PITH_FULL_IMAGE:figures/full_fig_p020_12.png] view at source ↗
Figure 13
Figure 13. Figure 13: Visualization comparison between raw DD and DIVER on ImageNette [PITH_FULL_IMAGE:figures/full_fig_p021_13.png] view at source ↗
Figure 14
Figure 14. Figure 14: Visualization comparison between raw DD and DIVER on ImageYellow. 21 [PITH_FULL_IMAGE:figures/full_fig_p021_14.png] view at source ↗
read the original abstract

Dataset distillation aims to synthesize a compact proxy dataset that is unreadable or non-raw from the original dataset for privacy protection and highly efficient learning. However, previous approaches typically adopt a single-stage distillation paradigm, which suffers from learning specific patterns that overfit on a prior architecture, consequently suppressing the expression of semantics and leading to performance degradation across heterogeneous architectures. To address this issue, we propose a novel dual-stage distillation framework called ${\textbf{DIVER}}$, which leverages the pre-trained diffusion model to dive deeper into $\textbf{DI}$stilled data $\textbf{V}$ia $\textbf{E}$xpressive semantic $\textbf{R}$ecovery, an entire process of semantic inheritance, guidance, and fusion. Semantic inheritance distills high-level semantics of abstract distilled images into the latent space to filter out architecture-specific ``noise" and retain the intrinsic semantics. Furthermore, semantic guidance improves the preservation of the original semantics by directing the reverse procedure. Finally, semantic fusion is designed to provide semantic guidance only during the concrete phase of the reverse process, preventing semantic ambiguity and artifacts while maintaining the guidance information. Extensive experiments validate the effectiveness and efficiency of DIVER in improving classical distillation techniques and significantly improving cross-architecture generalization, requiring processing time comparable to raw DiT on ImageNet (256$\times$256) with only 4 GB of GPU memory usage. Code is available: https://github.com/einsteinxia/DIVER.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The manuscript proposes DIVER, a dual-stage dataset distillation framework that leverages a pre-trained diffusion model to perform semantic inheritance (distilling high-level semantics into latent space to filter architecture-specific noise), semantic guidance (directing the reverse diffusion process), and semantic fusion (applying guidance only in the concrete phase of the reverse process). The central claim is that this recovers expressive semantics from distilled data, overcoming the architecture-specific overfitting of single-stage methods and yielding significantly better cross-architecture generalization, while requiring only DiT-comparable runtime and 4 GB GPU memory on ImageNet (256×256).

Significance. If the diffusion-based semantic recovery mechanism is shown to cleanly separate intrinsic semantics from architecture-specific patterns without introducing new biases, the work would meaningfully advance dataset distillation by providing a practical route to architecture-agnostic proxies. The reported efficiency (DiT-level time at 4 GB) would further strengthen its utility for privacy-preserving and resource-constrained learning scenarios.

major comments (2)
  1. [Abstract] Abstract: The central claim that semantic inheritance via the pre-trained diffusion model 'filters out architecture-specific noise' while retaining intrinsic semantics is load-bearing, yet the manuscript provides no explicit verification such as latent-space distance metrics, t-SNE visualizations, or failure-mode analysis demonstrating that the diffusion prior does not itself inject dataset-specific biases.
  2. [Abstract] Abstract: The assertion that restricting semantic guidance to the 'concrete phase' of the reverse process prevents ambiguity and artifacts is presented without a precise definition of how the concrete phase is identified or an ablation showing that earlier guidance stages produce the claimed artifacts.
minor comments (1)
  1. [Abstract] The efficiency claim of 'processing time comparable to raw DiT' and 'only 4 GB of GPU memory' should be supported by a dedicated runtime/memory table with exact hardware specifications and batch sizes.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback on our manuscript. We address each major comment point by point below and indicate the planned revisions.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim that semantic inheritance via the pre-trained diffusion model 'filters out architecture-specific noise' while retaining intrinsic semantics is load-bearing, yet the manuscript provides no explicit verification such as latent-space distance metrics, t-SNE visualizations, or failure-mode analysis demonstrating that the diffusion prior does not itself inject dataset-specific biases.

    Authors: We agree that explicit verification would strengthen the central claim. In the revised manuscript, we will add t-SNE visualizations of latent representations before and after semantic inheritance, quantitative metrics such as average cosine similarity and Euclidean distances in latent space across architectures, and a failure-mode analysis discussing potential biases from the diffusion prior. These will be included in Section 4 and the supplementary material. revision: yes

  2. Referee: [Abstract] Abstract: The assertion that restricting semantic guidance to the 'concrete phase' of the reverse process prevents ambiguity and artifacts is presented without a precise definition of how the concrete phase is identified or an ablation showing that earlier guidance stages produce the claimed artifacts.

    Authors: We acknowledge the need for a precise definition and ablation. The concrete phase will be defined as the final 400 timesteps (t ≤ 400) of the reverse process. We will add an ablation study comparing guidance at different stages, showing artifacts and performance drops for earlier guidance, with quantitative metrics on image quality and downstream accuracy. This will be added to Section 3.3 and the experiments. revision: yes

Circularity Check

0 steps flagged

No significant circularity; framework adds independent stages on external pre-trained models

full rationale

The paper's core contribution is a dual-stage distillation framework (DIVER) that applies semantic inheritance, guidance, and fusion using an external pre-trained diffusion model. The abstract and description introduce these as novel additions to filter architecture-specific noise while preserving semantics, without any equations, fitted parameters renamed as predictions, or self-citations that reduce the claims to tautologies or prior author work. The process is presented as building directly on independent diffusion priors, with no self-definitional loops or load-bearing internal citations visible. This qualifies as a self-contained derivation against external benchmarks, consistent with a normal non-circular finding.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The approach relies on existing pre-trained diffusion models as a foundation; no new free parameters or invented entities are introduced in the abstract. The key assumption is that diffusion latent spaces can separate architecture-specific noise from intrinsic semantics.

axioms (1)
  • domain assumption Pre-trained diffusion models can recover high-level semantics from distilled images by operating in latent space
    Invoked as the basis for semantic inheritance and guidance steps.

pith-pipeline@v0.9.0 · 5569 in / 1235 out tokens · 38942 ms · 2026-05-14T21:11:44.677198+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

62 extracted references · 62 canonical work pages · 5 internal anchors

  1. [1]

    Distilling the Knowledge in a Neural Network

    Distilling the knowledge in a neural network , author=. arXiv preprint arXiv:1503.02531 , year=

  2. [2]

    Dataset distillation.arXiv preprint arXiv:1811.10959, 2018

    Dataset distillation , author=. arXiv preprint arXiv:1811.10959 , year=

  3. [3]

    Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision , pages=

    Dataset condensation with distribution matching , author=. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision , pages=

  4. [4]

    Dataset condensation with gradient matching.arXiv preprint arXiv:2006.05929, 2020

    Dataset condensation with gradient matching , author=. arXiv preprint arXiv:2006.05929 , year=

  5. [5]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Dataset distillation by matching training trajectories , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  6. [6]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Generalizing dataset distillation via deep generative prior , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  7. [7]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Generalized large-scale data condensation via various backbone and statistical matching , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  8. [8]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    On the diversity and realism of distilled dataset: An efficient dataset distillation paradigm , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  9. [9]

    arXiv preprint arXiv:2406.05704 , year=

    Hierarchical features matter: A deep exploration of gan priors for improved dataset distillation , author=. arXiv preprint arXiv:2406.05704 , year=

  10. [10]

    Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

    From head to tail: Towards balanced representation in large vision-language models through adaptive data calibration , author=. Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

  11. [11]

    Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

    GazeGene: Large-scale Synthetic Gaze Dataset with 3D Eyeball Annotations , author=. Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

  12. [12]

    Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

    Dataset distillation with neural characteristic function: A minmax perspective , author=. Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

  13. [13]

    Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

    Emphasizing discriminative features for dataset distillation in complex scenarios , author=. Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

  14. [14]

    Su, Duo and Hou, Junjie and Gao, Weizhi and Tian, Yingjie and Tang, Bowen , booktitle=. D

  15. [15]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Efficient dataset distillation via minimax diffusion , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  16. [16]

    Chan-Santiago, Jeffrey A and Tirupattur, Praveen and Nayak, Gaurav Kumar and Liu, Gaowen and Shah, Mubarak , journal=. MGD ^

  17. [17]

    The Thirteenth International Conference on Learning Representations , year=

    Influence-guided diffusion for dataset distillation , author=. The Thirteenth International Conference on Learning Representations , year=

  18. [18]

    Advances in neural information processing systems , volume=

    Denoising diffusion probabilistic models , author=. Advances in neural information processing systems , volume=

  19. [19]

    Denoising Diffusion Implicit Models

    Denoising diffusion implicit models , author=. arXiv preprint arXiv:2010.02502 , year=

  20. [20]

    Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

    High-resolution image synthesis with latent diffusion models , author=. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

  21. [21]

    Classifier-Free Diffusion Guidance

    Classifier-free diffusion guidance , author=. arXiv preprint arXiv:2207.12598 , year=

  22. [22]

    Semantic image synthesis via diffusion models.arXiv preprint arXiv:2207.00050,

    Semantic image synthesis via diffusion models , author=. arXiv preprint arXiv:2207.00050 , year=

  23. [23]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Universal guidance for diffusion models , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  24. [24]

    Proceedings of the IEEE/CVF International Conference on Computer Vision , pages=

    Freedom: Training-free energy-guided conditional diffusion model , author=. Proceedings of the IEEE/CVF International Conference on Computer Vision , pages=

  25. [25]

    arXiv preprint arXiv:2501.14216 , year=

    TFG-Flow: Training-free Guidance in Multimodal Generative Flow , author=. arXiv preprint arXiv:2501.14216 , year=

  26. [26]

    Computer Vision--ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13 , pages=

    Visualizing and understanding convolutional networks , author=. Computer Vision--ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13 , pages=. 2014 , organization=

  27. [27]

    2009 IEEE conference on computer vision and pattern recognition , pages=

    Imagenet: A large-scale hierarchical image database , author=. 2009 IEEE conference on computer vision and pattern recognition , pages=. 2009 , organization=

  28. [28]

    URL https://github

    A smaller subset of 10 easily classified classes from imagenet, and a little more french , author=. URL https://github. com/fastai/imagenette , volume=

  29. [29]

    Advances in Neural Information Processing Systems , volume=

    Squeeze, recover and relabel: Dataset condensation at imagenet scale from a new perspective , author=. Advances in Neural Information Processing Systems , volume=

  30. [30]

    Proceedings of the IEEE/CVF International Conference on Computer Vision , pages=

    Datadam: Efficient dataset distillation with attention matching , author=. Proceedings of the IEEE/CVF International Conference on Computer Vision , pages=

  31. [31]

    Proceedings of the IEEE/CVF international conference on computer vision , pages=

    Scalable diffusion models with transformers , author=. Proceedings of the IEEE/CVF international conference on computer vision , pages=

  32. [32]

    Advances in neural information processing systems , volume=

    Generative modeling by estimating gradients of the data distribution , author=. Advances in neural information processing systems , volume=

  33. [33]

    ICML , year=

    Taming Diffusion for Dataset Distillation with High Representativeness , author=. ICML , year=

  34. [34]

    arXiv preprint arXiv:2505.13300 , year=

    Dd-ranking: Rethinking the evaluation of dataset distillation , author=. arXiv preprint arXiv:2505.13300 , year=

  35. [35]

    Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

    A style-based generator architecture for generative adversarial networks , author=. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

  36. [36]

    Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

    Analyzing and improving the image quality of stylegan , author=. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

  37. [37]

    Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

    Minimizing the accumulated trajectory error to improve dataset distillation , author=. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

  38. [38]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Exploiting inter-sample and inter-feature relations in dataset distillation , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  39. [39]

    Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

    Towards stable and storage-efficient dataset distillation: Matching convexified trajectory , author=. Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

  40. [40]

    arXiv preprint arXiv:2406.01063 , year=

    Dance: Dual-view distribution alignment for dataset condensation , author=. arXiv preprint arXiv:2406.01063 , year=

  41. [41]

    An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

    An image is worth 16x16 words: Transformers for image recognition at scale , author=. arXiv preprint arXiv:2010.11929 , year=

  42. [42]

    URL https://arxiv

    Dataset distillation with convexified implicit gradients (2023) , author=. URL https://arxiv. org/abs/2302.06755 , year=

  43. [43]

    Proceedings of the 17th ACM Conference on Recommender Systems , pages=

    Gradient matching for categorical data distillation in ctr prediction , author=. Proceedings of the 17th ACM Conference on Recommender Systems , pages=

  44. [44]

    Advances in Neural Information Processing Systems , volume=

    An efficient dataset condensation plugin and its application to continual learning , author=. Advances in Neural Information Processing Systems , volume=

  45. [45]

    arXiv preprint arXiv:2305.16645 , volume=

    Summarizing stream data for memory-restricted online continual learning , author=. arXiv preprint arXiv:2305.16645 , volume=

  46. [46]

    Proceedings of the AAAI Conference on Artificial Intelligence , volume=

    Fedvck: Non-iid robust and communication-efficient federated learning via valuable condensed knowledge for medical image analysis , author=. Proceedings of the AAAI Conference on Artificial Intelligence , volume=

  47. [47]

    arXiv preprint arXiv:2405.11525 , year=

    Overcoming data and model heterogeneities in decentralized federated learning via synthetic anchors , author=. arXiv preprint arXiv:2405.11525 , year=

  48. [48]

    arXiv preprint arXiv:2405.17535 , year=

    Calibrated dataset condensation for faster hyperparameter search , author=. arXiv preprint arXiv:2405.17535 , year=

  49. [49]

    International Conference on Machine Learning , pages=

    Generative teaching networks: Accelerating neural architecture search by learning to generate synthetic training data , author=. International Conference on Machine Learning , pages=. 2020 , organization=

  50. [50]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Cafe: Learning to condense dataset by aligning features , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  51. [51]

    arXiv preprint arXiv:2310.05773 , year=

    Towards lossless dataset distillation via difficulty-aligned trajectory matching , author=. arXiv preprint arXiv:2310.05773 , year=

  52. [52]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Small scale data-free knowledge distillation , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  53. [53]

    Proceedings of the IEEE/CVF international conference on computer vision , pages=

    Data-free learning of student networks , author=. Proceedings of the IEEE/CVF international conference on computer vision , pages=

  54. [54]

    IEEE transactions on pattern analysis and machine intelligence , volume=

    Dataset distillation: A comprehensive review , author=. IEEE transactions on pattern analysis and machine intelligence , volume=. 2023 , publisher=

  55. [55]

    Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

    Arbitrary-steps image super-resolution via diffusion inversion , author=. Proceedings of the Computer Vision and Pattern Recognition Conference , pages=

  56. [56]

    Advances in Neural Information Processing Systems , volume=

    One-step effective diffusion network for real-world image super-resolution , author=. Advances in Neural Information Processing Systems , volume=

  57. [57]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    Unsupervised blind image deblurring based on self-enhancement , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  58. [58]

    Proceedings of the AAAI Conference on Artificial Intelligence , volume=

    Boosting image de-raining via central-surrounding synergistic convolution , author=. Proceedings of the AAAI Conference on Artificial Intelligence , volume=

  59. [59]

    Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

    Sinsr: diffusion-based image super-resolution in a single step , author=. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition , pages=

  60. [60]

    CBV: Clean-label Backdoor Attacks on Vision Language Models via Diffusion Models

    CBV: Clean-label Backdoor Attacks on Vision Language Models via Diffusion Models , author=. arXiv preprint arXiv:2605.02202 , year=

  61. [61]

    ICASSP 2026-2026 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) , pages=

    EDITS: Enhancing Dataset Distillation with Implicit Textual Semantics , author=. ICASSP 2026-2026 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) , pages=. 2026 , organization=

  62. [62]

    European Conference on Computer Vision , pages=

    Sit: Exploring flow and diffusion-based generative models with scalable interpolant transformers , author=. European Conference on Computer Vision , pages=. 2024 , organization=