Commutativity regularization on Jacobians reduces transient error amplification in neural simulators, enabling stable rollouts over thousands of steps on physical and climate data.
super hub
U-Net: Convolutional Networks for Biomedical Image Segmentation, pp.\ 234–241
25 Pith papers cite this work, alongside 56,482 external citations. Polarity classification is still indexing.
hub tools
authors
co-cited works
representative citing papers
GPROF-IR is a CNN-based retrieval that uses temporal context in geostationary IR observations to produce precipitation estimates with lower error than prior IR methods and climatological consistency with PMW retrievals for integration into IMERG V08.
AttentionBender applies 2D transforms to cross-attention maps in video diffusion transformers, producing distributed distortions and glitch aesthetics that reveal entangled attention mechanisms while serving as both an XAI probe and creative tool.
A Jacobian sensitivity curve computed at initialization identifies the narrowest U-Net configuration that avoids performance collapse, matching nnU-Net accuracy with 400-1600x fewer parameters on six medical datasets.
Spectral analysis of activations and gradients provides new diagnostics that link batch size to representation geometry, early covariance tails to token efficiency, and spectral shifts to learning dynamics in decoder-only LLMs, backed by a mechanistic model.
A recurrent Vision Transformer hypernetwork injects context into Flux Neural Operators to infer and solve unseen conservation laws while preserving robustness and long-time stability.
SIAM achieves state-of-the-art whole-head MRI segmentation of 16 structures including extra-cerebral tissues by training on synthetic data from just six manual templates, matching or exceeding prior methods on 301 scans across eight heterogeneous datasets.
Cross-domain transfer of remote-sensing HSI foundation models improves proximal sensing semantic segmentation over in-domain training and narrows the gap to cross-modality methods on the HS3-Bench benchmark.
VSLP infers dense segmentations from global label proportions via a pre-trained transformer for initial confidence maps followed by variational optimization using Wasserstein fidelity and a learned regularizer, outperforming prior weakly supervised methods on histopathology datasets.
The ICPR 2026 LRLPR competition on real low-quality license plate images drew 99 valid submissions, with the winning team reaching 82.13% recognition rate and four teams exceeding 80%.
A tornado outbreak with simultaneous tornadic supercells occurred in the Philippines within an easterly severe weather regime, documented as the first known instance there.
FlowForge predicts flow fields via staged local updates with a shared lightweight predictor, matching or exceeding baselines in accuracy while improving robustness to noise and reducing latency.
PSIRNet produces diagnostic-quality free-breathing PSIR LGE cardiac MRI from a single interleaved IR/PD acquisition over two heartbeats using a physics-guided deep learning network trained on over 800,000 slices.
CATMIL augments nnU-Net with component-adaptive Tversky and MIL-based lesion supervision to raise Dice scores, small-lesion recall, and error control on the MSLesSeg dataset.
RABC-Net achieves 86.58% DICE and 79.47% JAC on skin lesion segmentation across ISIC-2017, ISIC-2018, and PH2 using only pseudo-labels and no manual masks for training or adaptation.
ASTERIS, a self-supervised spatiotemporal denoising algorithm, improves astronomical detection limits by 1 magnitude at 90% completeness while identifying three times more redshift >9 galaxy candidates in JWST images.
Biased noise sampling for rectified flows combined with a bidirectional text-image transformer architecture yields state-of-the-art high-resolution text-to-image results that scale predictably with model size.
An EnsCGP coarse surrogate plus U-Net-ASPP corrector emulates LISFLOOD-FP flood depths on a 256x256 grid around one Chicago gauge, achieving R² ≈ 0.99 and MAE < 0.01 m on held-out events while matching the gauge depth at that single pixel.
A semi-supervised framework using weighted knowledge distillation and SinusCycle-GAN refinement achieves 96.35% Dice score for maxillary sinus segmentation in panoramic X-rays from 2,511 patients.
Training-inference input alignment outweighs framework choice for longitudinal retinal image prediction, with deterministic regression matching complex models when acquisition variability dominates disease progression.
MIGF improves multi-modal prostate MRI segmentation robustness via modality-isolated streams and dropout training, yielding ranking score gains of 2.8-13.4% across backbones and better tolerance to degraded diffusion sequences on PI-CAI and Prostate158.
MAML with auxiliary cavity tasks and boundary-aware loss achieves better few-shot 3D left atrial wall segmentation than standard fine-tuning, reaching DSC 0.64 at 5 shots versus 0.52.
μ-FlowNet applies an attention U-Net to map flow fields in irregular microchannels, reporting dice score 0.9317 and IoU 0.8731 on test data while outperforming standard U-Net and T-Net.
A detection-gated YOLOv8n-U-Net pipeline extracts glottal area waveforms from high-speed endoscopy videos, achieving cross-dataset DSC of 0.745 and using area coefficient of variation to distinguish healthy from pathological cases in a 40-subject study.
citing papers explorer
-
Controlling Transient Amplification Improves Long-horizon Rollouts
Commutativity regularization on Jacobians reduces transient error amplification in neural simulators, enabling stable rollouts over thousands of steps on physical and climate data.
-
GPROF-IR: An Improved Single-Channel Infrared Precipitation Retrieval for Merged Satellite Precipitation Products
GPROF-IR is a CNN-based retrieval that uses temporal context in geostationary IR observations to produce precipitation estimates with lower error than prior IR methods and climatological consistency with PMW retrievals for integration into IMERG V08.
-
AttentionBender: Manipulating Cross-Attention in Video Diffusion Transformers as a Creative Probe
AttentionBender applies 2D transforms to cross-attention maps in video diffusion transformers, producing distributed distortions and glitch aesthetics that reveal entangled attention mechanisms while serving as both an XAI probe and creative tool.
-
XTinyU-Net: Training-Free U-Net Scaling via Initialization-Time Sensitivity
A Jacobian sensitivity curve computed at initialization identifies the narrowest U-Net configuration that avoids performance collapse, matching nnU-Net accuracy with 400-1600x fewer parameters on six medical datasets.
-
Spectral Lens: Activation and Gradient Spectra as Diagnostics of LLM Optimization
Spectral analysis of activations and gradients provides new diagnostics that link batch size to representation geometry, early covariance tails to token efficiency, and spectral shifts to learning dynamics in decoder-only LLMs, backed by a mechanistic model.
-
A Robust Foundation Model for Conservation Laws: Injecting Context into Flux Neural Operators via Recurrent Vision Transformers
A recurrent Vision Transformer hypernetwork injects context into Flux Neural Operators to infer and solve unseen conservation laws while preserving robustness and long-time stability.
-
SIAM: Head and Brain MRI Segmentation from Few High-Quality Templates via Synthetic Training
SIAM achieves state-of-the-art whole-head MRI segmentation of 16 structures including extra-cerebral tissues by training on synthetic data from just six manual templates, matching or exceeding prior methods on 301 scans across eight heterogeneous datasets.
-
Cross-Domain Transfer of Hyperspectral Foundation Models
Cross-domain transfer of remote-sensing HSI foundation models improves proximal sensing semantic segmentation over in-domain training and narrows the gap to cross-modality methods on the HS3-Bench benchmark.
-
Semantic Segmentation for Histopathology using Learned Regularization based on Global Proportions
VSLP infers dense segmentations from global label proportions via a pre-trained transformer for initial confidence maps followed by variational optimization using Wasserstein fidelity and a learned regularizer, outperforming prior weakly supervised methods on histopathology datasets.
-
ICPR 2026 Competition on Low-Resolution License Plate Recognition
The ICPR 2026 LRLPR competition on real low-quality license plate images drew 99 valid submissions, with the winning team reaching 82.13% recognition rate and four teams exceeding 80%.
-
Localized Tornado Outbreak at the Upstream of a Tropical Easterly Wave in Camarines Norte, Philippines (13 September 2025)
A tornado outbreak with simultaneous tornadic supercells occurred in the Philippines within an easterly severe weather regime, documented as the first known instance there.
-
FlowForge: A Staged Local Rollout Engine for Flow-Field Prediction
FlowForge predicts flow fields via staged local updates with a shared lightweight predictor, matching or exceeding baselines in accuracy while improving robustness to noise and reducing latency.
-
PSIRNet: Deep Learning-based Free-breathing Rapid Acquisition Late Enhancement Imaging
PSIRNet produces diagnostic-quality free-breathing PSIR LGE cardiac MRI from a single interleaved IR/PD acquisition over two heartbeats using a physics-guided deep learning network trained on over 800,000 slices.
-
Component-Adaptive and Lesion-Level Supervision for Improved Small Structure Segmentation in Brain MRI
CATMIL augments nnU-Net with component-adaptive Tversky and MIL-based lesion supervision to raise Dice scores, small-lesion recall, and error control on the MSLesSeg dataset.
-
RABC-Net: Reliability-Aware Annotation-Free Skin Lesion Segmentation for Low-Resource Dermoscopy
RABC-Net achieves 86.58% DICE and 79.47% JAC on skin lesion segmentation across ISIC-2017, ISIC-2018, and PH2 using only pseudo-labels and no manual masks for training or adaptation.
-
Deeper detection limits in astronomical imaging using self-supervised spatiotemporal denoising
ASTERIS, a self-supervised spatiotemporal denoising algorithm, improves astronomical detection limits by 1 magnitude at 90% completeness while identifying three times more redshift >9 galaxy candidates in JWST images.
-
Scaling Rectified Flow Transformers for High-Resolution Image Synthesis
Biased noise sampling for rectified flows combined with a bidirectional text-image transformer architecture yields state-of-the-art high-resolution text-to-image results that scale predictably with model size.
-
Observation-Guided Neural Surrogate Learning for Scientific Simulation Emulation: A Single-Gauge Flood-Inundation Proof of Concept
An EnsCGP coarse surrogate plus U-Net-ASPP corrector emulates LISFLOOD-FP flood depths on a 256x256 grid around one Chicago gauge, achieving R² ≈ 0.99 and MAE < 0.01 m on held-out events while matching the gauge depth at that single pixel.
-
Weighted Knowledge Distillation for Semi-Supervised Segmentation of Maxillary Sinus in Panoramic X-ray Images
A semi-supervised framework using weighted knowledge distillation and SinusCycle-GAN refinement achieves 96.35% Dice score for maxillary sinus segmentation in panoramic X-rays from 2,511 patients.
-
Training-inference input alignment outweighs framework choice in longitudinal retinal image prediction
Training-inference input alignment outweighs framework choice for longitudinal retinal image prediction, with deterministic regression matching complex models when acquisition variability dominates disease progression.
-
Architecture-Agnostic Modality-Isolated Gated Fusion for Robust Multi-Modal Prostate MRI Segmentation
MIGF improves multi-modal prostate MRI segmentation robustness via modality-isolated streams and dropout training, yielding ranking score gains of 2.8-13.4% across backbones and better tolerance to degraded diffusion sequences on PI-CAI and Prostate158.
-
Few-Shot Left Atrial Wall Segmentation in 3D LGE MRI via Meta-Learning
MAML with auxiliary cavity tasks and boundary-aware loss achieves better few-shot 3D left atrial wall segmentation than standard fine-tuning, reaching DSC 0.64 at 5 shots versus 0.52.
-
$\mu$-FlowNet: A Deep Learning Approach for Mapping Flow Fields in Irregular Microchannels Using an Attention-based U-Net Encoder-Decoder Architecture
μ-FlowNet applies an attention U-Net to map flow fields in irregular microchannels, reporting dice score 0.9317 and IoU 0.8731 on test data while outperforming standard U-Net and T-Net.
-
A Detection-Gated Pipeline for Robust Glottal Area Waveform Extraction and Clinical Pathology Assessment
A detection-gated YOLOv8n-U-Net pipeline extracts glottal area waveforms from high-speed endoscopy videos, achieving cross-dataset DSC of 0.745 and using area coefficient of variation to distinguish healthy from pathological cases in a 40-subject study.
-
Topology-Driven Fusion of nnU-Net and MedNeXt for Accurate Brain Tumor Segmentation on Sub-Saharan Africa Dataset
Pre-training nnU-Net and MedNeXt on BraTS 2025 data then fine-tuning on BraTS-Africa with added topology refinement yields NSD scores of 0.810, 0.829, and 0.895 for SNFH, NETC, and ET.