pith. machine review for the scientific record. sign in

arxiv: 2510.02370 · v3 · submitted 2025-09-29 · 💻 cs.CL · cs.AI

Recognition: unknown

How Training Data Shapes the Use of Parametric and In-Context Knowledge in Language Models

Authors on Pith no claims yet
classification 💻 cs.CL cs.AI
keywords knowledgelanguagemodelsparametricdatain-contexttrainingintra-document
0
0 comments X
read the original abstract

Large language models leverage both parametric knowledge acquired during pretraining and in-context knowledge provided at inference time. Crucially, when these sources conflict, models arbitrate based on their internal confidence, preferring parametric knowledge for high-confidence facts while deferring to context for less familiar ones. However, the training conditions that give rise to these fundamental behaviors remain unclear. Here we conduct controlled experiments using synthetic corpora to identify the specific data properties that shape knowledge utilization. Our results reveal a counterintuitive finding: the robust, balanced use of both knowledge sources is an emergent property that requires the co-occurrence of three factors typically considered detrimental, including (i) intra-document repetition, (ii) a moderate degree of intra-document inconsistency, and (iii) a skewed knowledge distribution. We further show that these dynamics arise in real-world language model pretraining and analyze how post-training procedures reshape arbitration strategies. Together, our findings provide empirical guidance for designing training data that supports the reliable integration of parametric and in-context knowledge in language models.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Three Regimes of Context-Parametric Conflict: A Predictive Framework and Empirical Validation

    cs.CL 2026-05 conditional novelty 6.0

    A three-regime framework resolves contradictions in LLM context vs. parametric knowledge conflicts by distinguishing single-source updating, competitive integration, and task-appropriate selection, with empirical conf...