pith. machine review for the scientific record. sign in

arxiv: 2605.06211 · v1 · submitted 2026-05-07 · 💻 cs.LG · cs.AI· cs.CL· cs.DS

Recognition: unknown

Contrastive Identification and Generation in the Limit

Authors on Pith no claims yet

Pith reviewed 2026-05-08 13:32 UTC · model grok-4.3

classification 💻 cs.LG cs.AIcs.CLcs.DS
keywords contrastive identificationidentification in the limitgeneration in the limitnoisy learningbinary hypothesis classescommon crossing graphtell-tale sets
0
0 comments X

The pith

Some binary concept classes can be identified from contrastive pairs under any finite corruption by one algorithm, but cannot be identified from positive examples even with a single corruption.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces contrastive identification and generation in the limit, where a learner sees an infinite stream of unordered pairs of examples that differ in label under an unknown binary hypothesis, but without any information on which example carries which label. It gives an exact geometric characterization of which classes admit contrastive identification, defines a new combinatorial dimension that governs uniform generation, and establishes that contrastive generation and ordinary text identification are incomparable. Under adversarial noise the paper proves a sharp reversal: there exist classes for which a single, noise-budget-independent algorithm succeeds from contrastive pairs no matter how many pairs are corrupted, yet the same classes cannot be identified from positive examples once even one example is corrupted.

Core claim

A class of binary hypotheses is contrastive identifiable in the limit precisely when every hypothesis has a tell-tale set of pairs that is closed under the common crossing relation; the common crossing graph simultaneously encodes pairwise ambiguity, obstructions to generation, and the locations of possible corruptions, allowing a single algorithm to recover the target from any finite number of corrupted contrastive pairs.

What carries the argument

The common crossing graph on a hypothesis class, whose edges and incidences capture pairwise label disagreements and thereby serve as the single structure for identification, generation, and corruption tolerance.

If this is right

  • Every class satisfying the one-line geometric refinement of Angluin’s tell-tale condition is exactly contrastive identifiable.
  • Contrastive generation is uniformly possible for a class if and only if its contrastive closure dimension is finite, with sample complexity governed by that dimension.
  • Contrastive generation and ordinary text identification are strictly incomparable: each admits classes the other does not.
  • A single algorithm works for contrastive identification under any finite corruption budget, while positive-only identification can fail with one corruption on the same classes.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The robustness result suggests that relational supervision may remain useful in settings where label noise is unbounded yet the relation itself is preserved.
  • The common crossing graph may serve as a template for studying other relational or comparison-based learning models beyond binary hypotheses.
  • If the crossing-graph characterization extends to finite domains, it would yield concrete algorithms for noisy contrastive clustering or ranking tasks.

Load-bearing premise

The input consists of unordered pairs in which the two examples always receive opposite labels under the unknown target hypothesis.

What would settle it

Exhibit a concrete class C of binary functions on a countable domain such that one algorithm recovers the target from any finite number of corrupted contrastive pairs yet no algorithm recovers the target from positive examples once a single example is replaced by an arbitrary string.

Figures

Figures reproduced from arXiv: 2605.06211 by Andi Han, Jiaojiao Jiang, Junbin Gao, Xiaoyu Li.

Figure 1
Figure 1. Figure 1: Diamond hierarchy on countable UUS classes. Arrows de￾note strict containment (⊊), ori￾ented from smaller to larger; the dashed line denotes incomparabil￾ity (∥). (3) The clean diamond hierarchy. On countable classes with infinite supports, the four families CtrId,TxtId,CtrGen,TxtGen form a strict diamond rather than a chain ( view at source ↗
Figure 2
Figure 2. Figure 2: Common crossing graph on a four-vertex example with supp(h) = {u1 , u2 }, supp(g) = {u1 , u3 }. Dashed blue: ∆(h) only; dotted red: ∆(g) only; solid dark: Γ (h, g). Notation. For h ∈ H, the crossing-edge set is ∆(h) := {{x, y} ∈ [X ] 2 : h(x) ̸= h( y)}; a contrastive pair for h is precisely an edge in ∆(h). For two hypotheses h, g, the common crossing graph is Γ (h, g) := ∆(h) ∩ ∆(g), viewed as an undirect… view at source ↗
Figure 3
Figure 3. Figure 3: Four-region partition of X by membership in supp(h) and supp(g): A in both, B in supp(h) only, C in supp(g) only, D in neither. An A-vertex needs a partner in D to lie in Γ (h, g); a B-vertex needs a partner in C (Theorem 4.3). Lemma 4.4 (Pairwise shared presentation). h, g ∈ H admit a common contrastive presentation valid for both targets iff supp(h) ∪ supp(g) ⊆ V(Γ (h, g)). Mutual non-eliminability impli… view at source ↗
Figure 4
Figure 4. Figure 4: The honest contrastive presentation for target view at source ↗
read the original abstract

In the classical identification in the limit model of Gold [1967], a stream of positive examples is presented round by round, and the learner must eventually recover the target hypothesis. Recently, Kleinberg and Mullainathan [2024] introduced generation in the limit, where the learner instead must eventually output novel elements of the target's support. Both lines of work focus on positive-only or fully labeled data. Yet many natural supervision signals are inherently relational rather than singleton, which encode relationships between examples rather than labels of individual ones. We initiate the study of contrastive identification and generation in the limit, where the learner observes a contrastive presentation of data: a stream of unordered pairs $\{x,y\}$ satisfying $h(x)\ne h(y)$ for an unknown target binary hypothesis $h$, but which element is positive is hidden from the learner. We first present three results in the noiseless setting: an exact characterization of contrastive identifiable classes (a one-line geometric refinement of Angluin [1980]'s tell-tale condition), a combinatorial dimension called contrastive closure dimension (a contrasitive analogue of the closure dimension in Raman et al. [2025]) and exactly characterizing uniform contrastive generation with tight sample complexity, and a strict hierarchy in which contrastive generation and text identification are mutually incomparable. We then prove a sharp reversal under finite adversarial corruption: there exist classes identifiable from contrastive pairs under any finite corruption budget by a single budget-independent algorithm, yet not identifiable from positive examples under even one corrupted observation. The unifying technical object is the common crossing graph, which encodes pairwise ambiguity, family-level generation obstructions, and corruption defects in a single coverage-and-incidence language.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

0 major / 2 minor

Summary. The manuscript introduces the study of contrastive identification and generation in the limit, where learners receive streams of unordered pairs {x, y} with h(x) ≠ h(y) for an unknown binary hypothesis h, without knowing the sign. It offers an exact characterization of contrastively identifiable classes as a geometric refinement of Angluin's tell-tale condition, defines the contrastive closure dimension that exactly characterizes uniform contrastive generation with tight sample complexity, establishes a strict hierarchy showing mutual incomparability between contrastive generation and text identification, and proves a reversal theorem under finite adversarial corruption: certain classes are identifiable from contrastive pairs under any finite corruption budget via a budget-independent algorithm, but not from positive examples under even a single corruption. The common crossing graph is the key technical object unifying these results.

Significance. This paper makes a substantial contribution to learning theory by extending classical identification in the limit to relational, contrastive data and demonstrating robustness advantages in the presence of adversarial noise. The exact characterizations and the combinatorial dimension provide precise tools for analyzing learnability, while the reversal theorem highlights a qualitative difference between contrastive and positive presentations under corruption. The use of the common crossing graph to encode multiple aspects of the problem is elegant and likely to inspire further work. The constructive proofs for the hierarchy and reversal add rigor to the claims.

minor comments (2)
  1. The term 'text identification' in the hierarchy result should be briefly defined or referenced to prior work for clarity, as it may not be immediately familiar to all readers.
  2. The abstract states 'exactly characterizing uniform contrastive generation with tight sample complexity'; specifying the sample complexity bound (e.g., in terms of the dimension) would make the claim more concrete even in the abstract.

Simulated Author's Rebuttal

0 responses · 0 unresolved

We thank the referee for their positive assessment and accurate summary of the manuscript. We appreciate the recognition of the substantial contribution to learning theory, the value of the exact characterizations via the geometric refinement of Angluin's condition, the contrastive closure dimension, the mutual incomparability results, and especially the reversal theorem under finite corruption. The recommendation for minor revision is noted, and we will prepare a revised manuscript accordingly.

Circularity Check

0 steps flagged

No significant circularity; derivations are self-contained combinatorial refinements

full rationale

The paper's core results—an exact characterization refining Angluin's tell-tale condition, the contrastive closure dimension as an analogue to Raman et al., uniform generation bounds, incomparability hierarchy, and the reversal theorem under corruption—are built from independent combinatorial arguments using the common crossing graph. This graph encodes pairwise ambiguity and defects without reducing to self-definition, fitted parameters renamed as predictions, or load-bearing self-citations. All steps cite external priors (Angluin 1980, Raman 2025, Gold 1967) or derive new objects directly; no equation or theorem collapses to its own inputs by construction. The reversal claim is supported by explicit graph connectivity arguments that distinguish contrastive vs. positive cases under finite corruption, remaining falsifiable and non-circular.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 2 invented entities

The central claims rest on standard inductive-inference assumptions and introduce two new combinatorial objects whose only support is the characterizations themselves.

axioms (1)
  • standard math Angluin's tell-tale condition and the classical identification-in-the-limit framework of Gold 1967
    The contrastive characterization is described as a geometric refinement of this condition.
invented entities (2)
  • contrastive closure dimension no independent evidence
    purpose: Combinatorial dimension that exactly characterizes uniform contrastive generation and supplies tight sample complexity
    New analogue of the closure dimension from Raman et al. 2025; no independent evidence supplied beyond the claimed characterization.
  • common crossing graph no independent evidence
    purpose: Single graph that encodes pairwise ambiguity, family-level generation obstructions, and corruption defects
    Unifying technical object introduced to prove all three noiseless results and the corruption reversal; no independent evidence supplied.

pith-pipeline@v0.9.0 · 5616 in / 1465 out tokens · 140676 ms · 2026-05-08T13:32:24.231193+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

89 extracted references · 12 canonical work pages · 6 internal anchors

  1. [1]

    Proceedings of the 58th Annual ACM Symposium on Theory of Computing (STOC) , year =

    Language Generation and Identification From Partial Enumeration: Tight Density Bounds and Topological Characterizations , author =. Proceedings of the 58th Annual ACM Symposium on Theory of Computing (STOC) , year =

  2. [2]

    Information and control , volume=

    Approximate language identification , author=. Information and control , volume=. 1974 , publisher=

  3. [3]

    Calibrating Noise to Sensitivity in Private Data Analysis , year =

    Dwork, Cynthia and McSherry, Frank and Nissim, Kobbi and Smith, Adam , pages =. Calibrating Noise to Sensitivity in Private Data Analysis , year =. Theory of Cryptography , doi =

  4. [4]

    48th Annual IEEE Symposium on Foundations of Computer Science (FOCS'07) , pages=

    Mechanism design via differential privacy , author=. 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS'07) , pages=. 2007 , organization=

  5. [5]

    Foundations and trends

    The algorithmic foundations of differential privacy , author=. Foundations and trends. 2014 , publisher=

  6. [6]

    Annual International Conference on the Theory and Applications of Cryptographic Techniques , pages=

    Our data, ourselves: Privacy via distributed noise generation , author=. Annual International Conference on the Theory and Applications of Cryptographic Techniques , pages=. 2006 , organization=

  7. [7]

    Agnostic Language Identification and Generation

    Agnostic Language Identification and Generation , author=. arXiv preprint arXiv:2601.23258 , year=

  8. [8]

    2010 IEEE 51st annual symposium on foundations of computer science , pages=

    Boosting and differential privacy , author=. 2010 IEEE 51st annual symposium on foundations of computer science , pages=. 2010 , organization=

  9. [9]

    Differentially private

    Acharya, Jayadev and Sun, Ziteng and Zhang, Huanyu , booktitle=. Differentially private. 2021 , organization=

  10. [10]

    On the Opportunities and Risks of Foundation Models

    On the opportunities and risks of foundation models , author=. arXiv preprint arXiv:2108.07258 , year=

  11. [11]

    ACM Computing Surveys , volume=

    Security and privacy challenges of large language models: A survey , author=. ACM Computing Surveys , volume=. 2025 , publisher=

  12. [12]

    2024 IEEE Symposium on Security and Privacy (SP) , pages=

    Poisoning web-scale training datasets is practical , author=. 2024 IEEE Symposium on Security and Privacy (SP) , pages=. 2024 , organization=

  13. [13]

    2024 , month =

    Afonja, Gbola and Sim, Robert and Lin, Zinan and Inan, Huseyin Atahan and Yekhanin, Sergey , title =. 2024 , month =

  14. [14]

    2024 , month =

    Shah, Nisarg and Andrew, Galen and Amin, Kareem , title =. 2024 , month =

  15. [15]

    Forty-first International Conference on Machine Learning , year=

    Position: Will we run out of data? Limits of LLM scaling based on human-generated data , author=. Forty-first International Conference on Machine Learning , year=

  16. [16]

    Ponomareva, Natalia and Xu, Zheng and McMahan, H Brendan and Kairouz, Peter and Rosenblatt, Lucas and Cohen-Addad, Vincent and Guzm. How to. arXiv preprint arXiv:2512.03238 , year=

  17. [17]

    30th USENIX security symposium (USENIX Security 21) , pages=

    Extracting training data from large language models , author=. 30th USENIX security symposium (USENIX Security 21) , pages=

  18. [18]

    The Eleventh International Conference on Learning Representations , year=

    Quantifying memorization across neural language models , author=. The Eleventh International Conference on Learning Representations , year=

  19. [19]

    2023 IEEE Symposium on Security and Privacy (SP) , pages=

    Analyzing leakage of personally identifiable information in language models , author=. 2023 IEEE Symposium on Security and Privacy (SP) , pages=. 2023 , organization=

  20. [20]

    Advances in Neural Information Processing Systems , volume=

    Emergent and predictable memorization in large language models , author=. Advances in Neural Information Processing Systems , volume=

  21. [21]

    Information and Control , volume =

    Language Identification in the Limit , author =. Information and Control , volume =

  22. [22]

    Proceedings of the eleventh annual ACM Symposium on Theory of Computing (STOC) , pages=

    Finding patterns common to a set of strings , author=. Proceedings of the eleventh annual ACM Symposium on Theory of Computing (STOC) , pages=

  23. [23]

    Information and Control , volume =

    Inductive Inference of Formal Languages from Positive Data , author =. Information and Control , volume =

  24. [24]

    Learning Indexed Families of Recursive Languages from Positive Data:

    Lange, Steffen and Zeugmann, Thomas and Zilles, Sandra , journal =. Learning Indexed Families of Recursive Languages from Positive Data:

  25. [25]

    Advances in Neural Information Processing Systems (

    Language Generation in the Limit , author =. Advances in Neural Information Processing Systems (

  26. [26]

    Proceedings of the Thirty-Eighth Conference on Learning Theory (

    Generation through the Lens of Learning Theory , author =. Proceedings of the Thirty-Eighth Conference on Learning Theory (

  27. [27]

    Proceedings of the Thirty-Eighth Conference on Learning Theory (

    Exploring Facets of Language Generation in the Limit , author =. Proceedings of the Thirty-Eighth Conference on Learning Theory (

  28. [28]

    Proceedings of the 57th Annual

    On the Limits of Language Generation: Trade-Offs between Hallucination and Mode Collapse , author =. Proceedings of the 57th Annual

  29. [29]

    Proceedings of the International Conference on Algorithmic Learning Theory (

    On Characterizations for Language Generation: Interplay of Hallucinations, Breadth, and Stability , author =. Proceedings of the International Conference on Algorithmic Learning Theory (

  30. [30]

    Proceedings of the 42nd International Conference on Machine Learning (

    Representative Language Generation , author =. Proceedings of the 42nd International Conference on Machine Learning (

  31. [31]

    Proceedings of the 42nd International Conference on Machine Learning (

    Generation from Noisy Examples , author =. Proceedings of the 42nd International Conference on Machine Learning (

  32. [32]

    Proceedings of the 2026 Annual

    Language Generation in the Limit: Noise, Loss, and Feedback , author =. Proceedings of the 2026 Annual

  33. [33]

    Language generation with infinite contamination

    Language Generation with Infinite Contamination , author =. arXiv preprint arXiv:2511.07417 , year =

  34. [34]

    arXiv preprint arXiv:2601.21237 , year =

    Quantifying Noise in Language Generation , author =. arXiv preprint arXiv:2601.21237 , year =

  35. [35]

    arXiv preprint arXiv:2602.07710 , year =

    On Generation in Metric Spaces , author =. arXiv preprint arXiv:2602.07710 , year =

  36. [36]

    Safe language generation in the limit

    Safe Language Generation in the Limit , author =. arXiv preprint arXiv:2601.08648 , year =

  37. [37]

    Proceedings of the 66th

    Density Measures for Language Generation , author =. Proceedings of the 66th

  38. [38]

    Advances in Neural Information Processing Systems (

    On Union-Closedness of Language Generation , author =. Advances in Neural Information Processing Systems (

  39. [39]

    Karbasi, Amin and Montasser, Omar and Sous, John and Velegkas, Grigoris , booktitle =

  40. [40]

    Proceedings of the International Conference on Algorithmic Learning Theory (

    Pareto-optimal Non-uniform Language Generation , author =. Proceedings of the International Conference on Algorithmic Learning Theory (

  41. [41]

    Proceedings of the 43rd International Conference on Machine Learning (ICML) , publisher =

    Language Generation: Complexity Barriers and Implications for Learning , author =. Proceedings of the 43rd International Conference on Machine Learning (ICML) , publisher =

  42. [42]

    arXiv preprint arXiv:2511.04103 , year=

    A Characterization of List Language Identification in the Limit , author =. arXiv preprint arXiv:2511.04103 , year =

  43. [43]

    The Fourteenth International Conference on Learning Representations (

    Language Identification in the Limit with Computational Trace , author =. The Fourteenth International Conference on Learning Representations (

  44. [44]

    What Can We Learn Privately? , author =

  45. [45]

    Proceedings of the 56th

    Differentially Private Release and Learning of Threshold Functions , author =. Proceedings of the 56th

  46. [46]

    Advances in Neural Information Processing Systems (

    A Computational Separation between Private Learning and Online Learning , author =. Advances in Neural Information Processing Systems (

  47. [47]

    On the Sample Complexity of Privately Learning Unbounded High-Dimensional

    Aden-Ali, Ishaq and Ashtiani, Hassan and Kamath, Gautam , booktitle =. On the Sample Complexity of Privately Learning Unbounded High-Dimensional

  48. [48]

    Proceedings of the

    Deep Learning with Differential Privacy , author =. Proceedings of the

  49. [49]

    Large-Scale Differentially Private

    Anil, Rohan and Ghazi, Badih and Gupta, Vineet and Kumar, Ravi and Manurangsi, Pasin , booktitle =. Large-Scale Differentially Private. 2022 , doi =

  50. [50]

    International Conference on Learning Representations (

    Differentially Private Fine-Tuning of Language Models , author =. International Conference on Learning Representations (

  51. [51]

    International Conference on Learning Representations (

    Large Language Models Can Be Strong Differentially Private Learners , author =. International Conference on Learning Representations (

  52. [52]

    Sander, Tom and Stock, Pierre and Sablayrolles, Alexandre , booktitle =

  53. [53]

    Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (

    Synthetic Text Generation with Differential Privacy: A Simple and Practical Recipe , author =. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (

  54. [54]

    Advances in Neural Information Processing Systems (

    User-Level Differential Privacy with Few Examples per User , author =. Advances in Neural Information Processing Systems (

  55. [55]

    Advances in neural information processing systems (NeurIPS) , volume=

    Private estimation with public data , author=. Advances in neural information processing systems (NeurIPS) , volume=

  56. [56]

    Proceedings of the 29th International Conference on Machine Learning (

    Convergence rates for differentially private statistical estimation , author=. Proceedings of the 29th International Conference on Machine Learning (

  57. [57]

    Alon, Noga and Livni, Roi and Malliaris, Maryanthe and Moran, Shay , booktitle =. Private

  58. [58]

    Proceedings of the 61st

    An equivalence between private classification and online prediction , author =. Proceedings of the 61st

  59. [59]

    Proceedings of the 43rd Annual

    Privacy-preserving statistical estimation with optimal convergence rates , author =. Proceedings of the 43rd Annual

  60. [60]

    Sample complexity bounds on differentially private learning via communication complexity , author =

  61. [61]

    Conference on Learning Theory , pages=

    Locally private hypothesis selection , author=. Conference on Learning Theory , pages=. 2020 , organization=

  62. [62]

    Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing (STOC) , pages=

    The structure of optimal private tests for simple hypotheses , author=. Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing (STOC) , pages=

  63. [63]

    2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS) , pages=

    Max-information, differential privacy, and post-selection hypothesis testing , author=. 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS) , pages=. 2016 , organization=

  64. [64]

    Proceedings of the 43rd International Conference on Machine Learning (ICML) , publisher =

    Language Generation with Replay: A Learning-Theoretic View of Model Collapse , author =. Proceedings of the 43rd International Conference on Machine Learning (ICML) , publisher =

  65. [65]

    The Thirty Eighth Annual Conference on Learning Theory (COLT) , pages=

    Learning Algorithms in the Limit , author=. The Thirty Eighth Annual Conference on Learning Theory (COLT) , pages=. 2025 , organization=

  66. [66]

    Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining , pages=

    How to dp-fy ml: A practical tutorial to machine learning with differential privacy , author=. Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining , pages=

  67. [67]

    Proceedings of the 54th annual ACM SIGACT symposium on theory of computing , pages=

    Reproducibility in learning , author=. Proceedings of the 54th annual ACM SIGACT symposium on theory of computing , pages=

  68. [68]

    and Leike, Jan and Brown, Tom B

    Christiano, Paul F. and Leike, Jan and Brown, Tom B. and Martic, Miljan and Legg, Shane and Amodei, Dario , title =. Advances in Neural Information Processing Systems (NeurIPS) , year =

  69. [69]

    and Mishkin, Pamela and Zhang, Chong and Agarwal, Sandhini and Slama, Katarina and Ray, Alex and others , title =

    Ouyang, Long and Wu, Jeffrey and Jiang, Xu and Almeida, Diogo and Wainwright, Carroll L. and Mishkin, Pamela and Zhang, Chong and Agarwal, Sandhini and Slama, Katarina and Ray, Alex and others , title =. Advances in Neural Information Processing Systems (NeurIPS) , year =

  70. [70]

    and Finn, Chelsea , title =

    Rafailov, Rafael and Sharma, Archit and Mitchell, Eric and Ermon, Stefano and Manning, Christopher D. and Finn, Chelsea , title =. Advances in Neural Information Processing Systems (NeurIPS) , year =

  71. [71]

    Label ranking by learning pairwise preferences , journal =

    H. Label ranking by learning pairwise preferences , journal =

  72. [72]

    and Terry, Milton E

    Bradley, Ralph A. and Terry, Milton E. , title =. Biometrika , volume =

  73. [73]

    A survey of preference-based reinforcement learning methods , journal =

    Wirth, Christian and Akrour, Riad and Neumann, Gerhard and F. A survey of preference-based reinforcement learning methods , journal =

  74. [74]

    and Lowe, Ryan and Voss, Chelsea and Radford, Alec and Amodei, Dario and Christiano, Paul , title =

    Stiennon, Nisan and Ouyang, Long and Wu, Jeffrey and Ziegler, Daniel M. and Lowe, Ryan and Voss, Chelsea and Radford, Alec and Amodei, Dario and Christiano, Paul , title =. Advances in Neural Information Processing Systems (NeurIPS) , year =

  75. [75]

    Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback

    Bai, Yuntao and Jones, Andy and Ndousse, Kamal and Askell, Amanda and Chen, Anna and DasSarma, Nova and Drain, Dawn and Fort, Stanislav and Ganguli, Deep and Henighan, Tom and others , title =. arXiv preprint arXiv:2204.05862 , year =

  76. [76]

    A general theoretical paradigm to understand learning from human preferences , booktitle =

    Azar, Mohammad Gheshlaghi and Guo, Zhaohan Daniel and Piot, Bilal and Munos, R. A general theoretical paradigm to understand learning from human preferences , booktitle =

  77. [77]

    Nash learning from human feedback , booktitle =

    Munos, R. Nash learning from human feedback , booktitle =

  78. [78]

    On the Price of Privacy for Language Identification and Generation

    On the Price of Privacy for Language Identification and Generation , author =. arXiv preprint arXiv:2604.07238 , year =

  79. [79]

    Differentially Private Language Generation and Identification in the Limit

    Differentially Private Language Generation and Identification in the Limit , author =. arXiv preprint arXiv:2604.08504 , year =

  80. [80]

    Proceedings of the 35th International Conference on Machine Learning (ICML) , year =

    Classification from Pairwise Similarity and Unlabeled Data , author =. Proceedings of the 35th International Conference on Machine Learning (ICML) , year =

Showing first 80 references.