pith. machine review for the scientific record. sign in

arxiv: 2604.11496 · v2 · submitted 2026-04-13 · 💻 cs.CV · cs.CL· cs.LG

Recognition: unknown

Revisiting Compositionality in Dual-Encoder Vision-Language Models: The Role of Inference

Authors on Pith no claims yet

Pith reviewed 2026-05-10 16:34 UTC · model grok-4.3

classification 💻 cs.CV cs.CLcs.LG
keywords dual-encoder vision-language modelscompositionalityinference protocollocalized alignmentCLIPcompositional generalizationdistribution shiftfine-grained alignment
0
0 comments X

The pith

The standard global cosine similarity inference is the main bottleneck for compositional generalization in dual-encoder vision-language models.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Dual-encoder VLMs such as CLIP perform poorly on compositional tasks because they match entire image and text embeddings rather than aligning their parts. Controlled experiments show that enforcing fine-grained region-to-segment matching at inference time, without any encoder updates, sharply raises scores on compositional benchmarks. A lightweight transformer trained to produce these alignments from frozen patch and token features matches the in-domain retrieval accuracy of full fine-tuning while delivering larger gains on out-of-domain compositional tests. The results indicate that the choice of inference protocol, not the quality of the pretrained representations, determines whether these models can handle novel combinations of objects and relations.

Core claim

Global embedding matching limits compositional ability; replacing it at inference with explicit or learned localized alignment between image regions and text tokens, using only frozen encoders, produces in-domain retrieval performance comparable to full fine-tuning and substantially better out-of-domain compositional generalization than either full fine-tuning or prior end-to-end compositional training methods.

What carries the argument

A lightweight transformer that learns localized region-segment alignments directly from the frozen patch and token embeddings of a dual-encoder VLM, replacing the standard global cosine similarity at inference time.

If this is right

  • Localized alignment over frozen representations matches full fine-tuning on in-domain retrieval tasks.
  • Localized alignment yields larger improvements than full fine-tuning on controlled out-of-domain compositional benchmarks.
  • Global embedding matching constitutes a key bottleneck preventing robust compositional generalization in dual-encoder VLMs.
  • Alignment mechanisms rather than end-to-end retraining are sufficient for strong compositional generalization under distribution shift.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • VLM architectures could add small inference-time alignment modules instead of retraining the entire model for each new domain.
  • Diagnostic protocols that separate representation quality from inference protocol could be applied to other reported VLM limitations such as bias or robustness failures.
  • If the lightweight transformer generalizes across base models, similar localized matching might improve compositionality in other multimodal dual-encoder systems.

Load-bearing premise

The controlled diagnostic experiments and out-of-domain benchmarks truly isolate the effect of the inference protocol without confounding factors such as dataset biases or unintended cues.

What would settle it

A controlled test in which full fine-tuning produces equal or larger gains than the localized-alignment method on the same out-of-domain compositional benchmarks would show that the inference protocol is not the primary bottleneck.

Figures

Figures reproduced from arXiv: 2604.11496 by Ander Salaberria, Eneko Agirre, Gorka Azkune, Imanol Miranda.

Figure 1
Figure 1. Figure 1: Vision-language compositional reasoning requires fine-grained alignment between textual segments describing [PITH_FULL_IMAGE:figures/full_fig_p002_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Examples from the BISCOR-CTRL dataset. From left to right: instances from the COLOR, SIZE, MATERIAL, and QUANTITY categories (the last containing 8 objects). Each instance consists of two image–caption pairs: a correct pair (top image and caption) and a hard negative pair (bottom image and caption). of image evidence [Lin et al., 2024b,a, Miranda et al., 2024, Udandarao et al., 2025]. Accordingly, BISCOR-C… view at source ↗
Figure 3
Figure 3. Figure 3: Example of a BISCOR instance after loading the dataset. Maintenance plan We are committed to maintaining the dataset to resolve any technical issues. We actively track issues in the HuggingFace or GitHub repositories. Licensing Our work is licensed under the MIT License4 for the code and a Creative Commons Attribution 4.0 International License (CC BY 4.0) for the data5 . Author statement We, the authors, a… view at source ↗
Figure 4
Figure 4. Figure 4: An example for our two text segmenting strategies. As can be seen, [PITH_FULL_IMAGE:figures/full_fig_p014_4.png] view at source ↗
read the original abstract

Dual-encoder Vision-Language Models (VLMs) such as CLIP are often characterized as bag-of-words systems due to their poor performance on compositional benchmarks. We argue that this limitation may stem less from deficient representations than from the standard inference protocol based on global cosine similarity. First, through controlled diagnostic experiments, we show that explicitly enforcing fine-grained region-segment alignment at inference dramatically improves compositional performance without updating pretrained encoders. We then introduce a lightweight transformer that learns such alignments directly from frozen patch and token embeddings. Comparing against full fine-tuning and prior end-to-end compositional training methods, we find that although these approaches improve in-domain retrieval, their gains do not consistently transfer under distribution shift. In contrast, learning localized alignment over frozen representations matches full fine-tuning on in-domain retrieval while yielding substantial improvements on controlled out-of-domain compositional benchmarks. These results identify global embedding matching as a key bottleneck in dual-encoder VLMs and highlight the importance of alignment mechanisms for robust compositional generalization.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 1 minor

Summary. The paper claims that compositional limitations in dual-encoder VLMs such as CLIP stem primarily from the global cosine similarity inference protocol rather than from the pretrained representations. Diagnostic experiments show that enforcing explicit region-segment alignment at inference improves compositional performance without updating encoders. A lightweight transformer is introduced to learn localized alignments from frozen patch and token embeddings; this matches full fine-tuning on in-domain retrieval while delivering substantial gains on controlled out-of-domain compositional benchmarks, identifying global embedding matching as the key bottleneck.

Significance. If the results hold, the work would meaningfully shift VLM research from representation-centric fine-tuning toward inference-time alignment mechanisms. The lightweight transformer offers an efficient path to compositional robustness that preserves in-domain performance, and the empirical contrast with end-to-end methods provides a practical demonstration that global matching is a removable limitation rather than an inherent representational deficit.

major comments (1)
  1. [Experimental evaluation and OOD benchmark definitions] The central claim that gains on controlled OOD compositional benchmarks reflect the inference protocol (rather than exploitation of benchmark-specific statistics) is load-bearing for the contrast with full fine-tuning. The manuscript must demonstrate that these OOD sets lack unintended correlations such as object co-occurrence frequencies, spatial priors, or attribute-visual shortcuts that localized alignment could exploit while global cosine similarity cannot; without explicit bias audits or controls in the experimental section, attribution to the alignment mechanism remains insecure.
minor comments (1)
  1. [Abstract] The abstract would be strengthened by briefly naming the specific OOD compositional benchmarks and reporting the magnitude of the observed improvements (e.g., percentage gains or absolute scores) to allow readers to assess the practical significance immediately.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for their constructive feedback, which identifies a key requirement for strengthening the attribution of our results to the inference protocol. We address the major comment below and outline the revisions we will make.

read point-by-point responses
  1. Referee: The central claim that gains on controlled OOD compositional benchmarks reflect the inference protocol (rather than exploitation of benchmark-specific statistics) is load-bearing for the contrast with full fine-tuning. The manuscript must demonstrate that these OOD sets lack unintended correlations such as object co-occurrence frequencies, spatial priors, or attribute-visual shortcuts that localized alignment could exploit while global cosine similarity cannot; without explicit bias audits or controls in the experimental section, attribution to the alignment mechanism remains insecure.

    Authors: We agree that explicit bias audits are necessary to secure the attribution of gains to the localized alignment mechanism. Section 4.2 of the manuscript details the construction of the OOD benchmarks, which are built from standard datasets using held-out combinations of objects, attributes, and relations to enforce compositional novelty. To directly address the concern, we will add quantitative bias audits in the revised experimental section. These will include: (i) co-occurrence frequency matrices for object pairs and attribute-object combinations, (ii) spatial prior statistics (e.g., bounding-box centroid distributions), and (iii) attribute-visual shortcut correlations, all compared between in-domain and OOD splits. We will also report whether localized alignment shows differential exploitation of any residual correlations relative to global cosine similarity. This addition will make the controls explicit and allow readers to evaluate the security of our claims. revision: yes

Circularity Check

0 steps flagged

No circularity in empirical evaluation of inference protocols

full rationale

The paper advances an empirical hypothesis that compositional failures in dual-encoder VLMs stem primarily from global cosine-similarity inference rather than encoder representations. This is tested via controlled diagnostic experiments that enforce region-segment alignment at inference time on frozen encoders, followed by introduction of a lightweight transformer trained on those frozen patch/token embeddings. Results are compared against full fine-tuning and prior end-to-end methods on both in-domain retrieval and out-of-domain compositional benchmarks. No equations, self-definitional loops, fitted-input predictions, or load-bearing self-citations appear in the derivation; all performance deltas are measured against external baselines and distribution-shift controls. The central claim therefore rests on falsifiable experimental contrasts rather than any reduction of outputs to the paper's own inputs by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Based on the abstract alone, the paper relies on standard machine-learning assumptions about benchmark validity and transformer capacity to learn alignments; no explicit free parameters, domain axioms, or invented entities are described.

pith-pipeline@v0.9.0 · 5482 in / 1105 out tokens · 35853 ms · 2026-05-10T16:34:13.336509+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

23 extracted references · 11 canonical work pages · 4 internal anchors

  1. [1]

    Image-text retrieval: A survey on recent research and development.arXiv preprint arXiv:2203.14713,

    Min Cao, Shiping Li, Juntao Li, Liqiang Nie, and Min Zhang. Image-text retrieval: A survey on recent research and development.arXiv preprint arXiv:2203.14713,

  2. [2]

    CLIPScore: A reference-free evaluation metric for image captioning

    Jack Hessel, Ari Holtzman, Maxwell Forbes, Ronan Le Bras, and Yejin Choi. CLIPScore: A reference-free evaluation metric for image captioning. In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors,Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7514– 7528, Online and Punta Cana, ...

  3. [3]

    Clipscore: A reference-free evaluation metric for image captioning

    Association for Computational Linguistics. doi:10.18653/v1/2021.emnlp-main.595. URLhttps://aclanthology.org/2021.emnlp-main.595/. Mert Yuksekgonul, Federico Bianchi, Pratyusha Kalluri, Dan Jurafsky, and James Zou. When and why vision-language models behave like bags-of-words, and what to do about it? InThe Eleventh International Conference on Learning Rep...

  4. [4]

    Why is winoground hard? investigating failures in visuolinguistic compositionality

    Anuj Diwan, Layne Berry, Eunsol Choi, David Harwath, and Kyle Mahowald. Why is winoground hard? investigating failures in visuolinguistic compositionality. InProceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2236–2250,

  5. [5]

    SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features

    Michael Tschannen, Alexey Gritsenko, Xiao Wang, Muhammad Ferjad Naeem, Ibrahim Alabdulmohsin, Nikhil Parthasarathy, Talfan Evans, Lucas Beyer, Ye Xia, Basil Mustafa, et al. Siglip 2: Multilingual vision-language encoders with improved semantic understanding, localization, and dense features.arXiv preprint arXiv:2502.14786,

  6. [6]

    Perception Encoder: The best visual embeddings are not at the output of the network

    Daniel Bolya, Po-Yao Huang, Peize Sun, Jang Hyun Cho, Andrea Madotto, Chen Wei, Tengyu Ma, Jiale Zhi, Jathushan Rajasegaran, Hanoona Rasheed, et al. Perception encoder: The best visual embeddings are not at the output of the network.arXiv preprint arXiv:2504.13181,

  7. [7]

    PaliGemma: A versatile 3B VLM for transfer

    Lucas Beyer, Andreas Steiner, André Susano Pinto, Alexander Kolesnikov, Xiao Wang, Daniel Salz, Maxim Neumann, Ibrahim Alabdulmohsin, Michael Tschannen, Emanuele Bugliarello, et al. Paligemma: A versatile 3b vlm for transfer. arXiv preprint arXiv:2407.07726,

  8. [8]

    Qwen3-VL Technical Report

    Shuai Bai, Yuxuan Cai, Ruizhe Chen, Keqin Chen, Xionghui Chen, Zesen Cheng, Lianghao Deng, Wei Ding, Chang Gao, Chunjiang Ge, et al. Qwen3-vl technical report.arXiv preprint arXiv:2511.21631,

  9. [9]

    Evaluating text-to-visual generation with image-to-text generation, 2024

    Zhiqiu Lin, Deepak Pathak, Baiqi Li, Jiayao Li, Xide Xia, Graham Neubig, Pengchuan Zhang, and Deva Ramanan. Evaluating text-to-visual generation with image-to-text generation.arXiv preprint arXiv:2404.01291, 2024a. Lewei Yao, Runhui Huang, Lu Hou, Guansong Lu, Minzhe Niu, Hang Xu, Xiaodan Liang, Zhenguo Li, Xin Jiang, and Chunjing Xu. FILIP: fine-grained ...

  10. [10]

    Filip: Fine-grained interactive language-image pre-training.ArXiv, abs/2111.07783, 2021

    URL https://arxiv.org/abs/2111.07783. Yan Zeng, Xinsong Zhang, and Hang Li. Multi-grained vision language pre-training: Aligning texts with visual concepts. InInternational Conference on Machine Learning, pages 25994–26009. PMLR,

  11. [11]

    Pyramidclip: Hierarchical feature alignment for vision-language model pretraining, 2022

    URLhttps://arxiv.org/abs/2204.14095. Youngtaek Oh, Jae Won Cho, Dong-Jin Kim, In So Kweon, and Junmo Kim. Preserving multi-modal capabilities of pre-trained vlms for improving vision-linguistic compositionality.arXiv preprint arXiv:2410.05210,

  12. [12]

    Ugur Sahin, Hang Li, Qadeer Khan, Daniel Cremers, and V olker Tresp

    URL https://proceedings.neurips.cc/paper_files/paper/ 2024/file/3122aaa22b2fe83f9cead1a696f65ceb-Paper-Conference.pdf. Ugur Sahin, Hang Li, Qadeer Khan, Daniel Cremers, and V olker Tresp. Enhancing multimodal compositional reasoning of visual language models with generative negative mining. InProceedings of the IEEE/CVF Winter Conference on Applications o...

  13. [13]

    Zhiqiu Lin, Xinyue Chen, Deepak Pathak, Pengchuan Zhang, and Deva Ramanan

    URL https://proceedings.neurips.cc/paper_files/paper/2024/file/ 39781da4b5d05bc2908ce08e43bc6404-Paper-Conference.pdf. Zhiqiu Lin, Xinyue Chen, Deepak Pathak, Pengchuan Zhang, and Deva Ramanan. Revisiting the role of language priors in vision-language models. InInternational Conference on Machine Learning, pages 29914–29934. PMLR, 2024b. Vishaal Udandarao...

  14. [14]

    spaCy: Industrial-strength natural language processing in python, 2020

    11 Revisiting Compositionality in Dual-Encoder Vision-Language Models: The Role of Inference Matthew Honnibal, Ines Montani, Sofie Van Landeghem, and Adriane Boyd. spaCy: Industrial-strength Natural Language Processing in Python.https://doi.org/10.5281/zenodo.1212303,

  15. [15]

    Microsoft coco: Common objects in context

    Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. InComputer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pages 740–755. Springer,

  16. [16]

    Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. Transformers: State-of-the-art...

  17. [17]

    URLhttps://www.aclweb.org/anthology/2020.emnlp-demos.6

    Association for Computational Linguistics. URLhttps://www.aclweb.org/anthology/2020.emnlp-demos.6. 12 Revisiting Compositionality in Dual-Encoder Vision-Language Models: The Role of Inference Appendix of Revisiting Compositionality in Dual-Encoder Vision-Language Models: The Role of Inference A BISCOR-CTRLdataset information We host BISCOR at HuggingFace

  18. [18]

    Dataset documentationBISCOR-CTRLis a benchmark of Bidirectional SWAPSfor Compositional Reasoning development

    We provide a summary below. Dataset documentationBISCOR-CTRLis a benchmark of Bidirectional SWAPSfor Compositional Reasoning development. Each instance consists of two images and two captions. Using each of the images and captions as a base, a model is asked to select the pair that correctly represents the base versus the hard negative distractor with min...

  19. [19]

    en_core_web_sm

    and (112, 56), combining different scales and aspect ratios. We resize all the crops to the input size of the model and we deploy those crops in two different ways: i)grid, avoiding any overlap of crops of the same size, and ii)overlap, using a stride of crop_size/2. This means that we process 86 crops per image with thegridconfiguration, and 270 crops wi...

  20. [20]

    •SUGARCREPE[Hsieh et al., 2024]: We obtain SUGARCREPEfrom the official GitHub repository

  21. [21]

    • CLIP:We obtain the pretrained baseline VIT-B-32 OpenAI’s CLIP model [Radford et al., 2021] from Hugging Face12

    E.4 Software information ModelsWe detail the sources of models we used. • CLIP:We obtain the pretrained baseline VIT-B-32 OpenAI’s CLIP model [Radford et al., 2021] from Hugging Face12. •SigLIP 2:We obtain all SigLIP 2 [Tschannen et al., 2025] models from Hugging Face collection

  22. [22]

    – Pe:We obtain PE-Core-B16-224 from the official Hugging Face repository

    • Perception Encoder:We obtain all Perception Encoder [Bolya et al., 2025] models from Hugging Face collection16. – Pe:We obtain PE-Core-B16-224 from the official Hugging Face repository

  23. [23]

    19 •TripletCLIP:We obtain the TripletCLIP model [Patel et al., 2024] from the official GitHub repository

    •NegCLIP:We obtain the NegCLIP model [Yuksekgonul et al., 2022] from the official GitHub repository. 19 •TripletCLIP:We obtain the TripletCLIP model [Patel et al., 2024] from the official GitHub repository. 20 •FSC-CLIP:We obtain the FSC-CLIP model [Oh et al., 2024] from the official GitHub repository. 21 •FineCLIP:We obtain the FineCLIP model [Jing et al...