pith. machine review for the scientific record. sign in

arxiv: 2605.10196 · v1 · submitted 2026-05-11 · 💻 cs.LG

Recognition: no theorem link

Many Needles in a Haystack: Active Hit Discovery for Perturbation Experiments

Authors on Pith no claims yet

Pith reviewed 2026-05-12 03:35 UTC · model grok-4.3

classification 💻 cs.LG
keywords hit discoveryacquisition functionssequential experimental designperturbation experimentsthreshold exceedanceBayesian optimizationgene perturbation
0
0 comments X

The pith

Probability-of-Hit ranks perturbation candidates by their chance of exceeding a fixed phenotypic threshold to locate many effective interventions rather than one optimum.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

High-throughput gene perturbation experiments test many genetic changes at once but face tight experimental budgets, so the goal is to identify as many perturbations as possible whose effect crosses a known threshold. Pure random testing wastes effort on weak candidates, while standard Bayesian optimization chases a single global peak and overlooks other strong ones. This paper treats hit discovery as sequential design and introduces an acquisition function that picks the next test by computing the posterior probability a candidate is a hit. The approach comes with a proof of asymptotic optimality and delivers measurable gains on synthetic cases plus real immunology data.

Core claim

We formalize hit discovery as a sequential experimental design problem and propose Probability-of-Hit, an acquisition function that directly targets threshold exceedance by ranking candidates according to their posterior probability of being a hit. We prove asymptotic optimality of this approach and demonstrate strong empirical performance on both synthetic benchmarks and real biological immunology datasets, including up to 6.4% improvement over baselines on the Schmidt IL-2 dataset.

What carries the argument

Probability-of-Hit acquisition function that selects the next perturbation by maximizing the surrogate model's estimated probability that the candidate exceeds the predefined phenotypic threshold.

If this is right

  • Experimental budgets in perturbation screens are spent on candidates more likely to meet the effect threshold.
  • Asymptotic optimality ensures the fraction of discovered hits approaches the best possible rate as the number of tests grows.
  • Empirical gains appear on synthetic benchmarks and reach 6.4% higher hit recovery on the Schmidt IL-2 immunology dataset compared with baselines.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same ranking logic could transfer to other budgeted search tasks that seek many good solutions, such as screening chemical libraries or materials candidates.
  • Testing the method when the threshold itself must be learned from data would reveal whether the current fixed-threshold assumption limits practical use.
  • Comparing performance across different surrogate models on the same biological data would expose how sensitive the hit-ranking gains are to model choice.

Load-bearing premise

A surrogate model must supply reliable posterior probabilities that each candidate exceeds the fixed, known phenotypic threshold.

What would settle it

Running the method and standard Bayesian optimization on the same real perturbation dataset and finding that Probability-of-Hit recovers no more or fewer actual hits within the budget.

Figures

Figures reproduced from arXiv: 2605.10196 by Amir Akbarnejad, Andrea Rubbi, Arpit Merchant, Mo Lotfollahi, Pietro Li\`o, Samuel Ogden, Sattar Vakili.

Figure 1
Figure 1. Figure 1: (a) 2D UMAP gene embeddings of Achilles features with the Z-axis denoting phenotypic response to sensitivity of leukemia cells to cytotoxic activity in human NK cells and colors indicating threshold levels. ITGAV, CYB5B and EYA4 are representative genes from three separate modes. (b) Overview of our closed-loop experimental design system. In each cycle, the algorithm samples a batch of genes from a probabi… view at source ↗
Figure 2
Figure 2. Figure 2: Illustration of the learning process underlying Probability-of-Hit. The solid grey line represents the ground-truth distribution, the dotted grey line represents τ = 10%. The solid blue line and the shaded blue region around it represent the current surrogate and the model’s uncertainty with the red points denoting the samples chosen in each cycle [PITH_FULL_IMAGE:figures/full_fig_p007_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Impact of batch size (b) on CHR@10 for Perturbation￾Pathways (4D). Probability-of-Hit outperforms all baselines. Exploration-Exploitation Trade-off [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Trade-off at b = 5 between CHR@10 and SMAPE@10 on Perturbation-Pathways (4D) across acquisition functions. tions. Results are averaged over 5 seeds. Probability-of-Hit significantly outperforms Random (Cliff’ δ = 0.894, paired t-test p < 0.0001, Wilcoxon signed-rank p < 0.0001), re￾covering, on average, 107 more hits after 10 cycles. Probability-of-Hit displays the largest advantage on Schmidt IL-2. This d… view at source ↗
Figure 5
Figure 5. Figure 5: Method comparison with DiscoBax on 10 cycles, batches of 5 points, and threshold of 10% (over 3 random seeds). DiscoBax consistently performs better than Random, but still significantly worse than the other methods [PITH_FULL_IMAGE:figures/full_fig_p018_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Hits Recovered Ratio vs. Mean Runtime on 10 cycles with batches of 5 (over 3 random seeds). DiscoBax’s significantly longer runtimes, together with a sub-par performance with respect to the other methods, place it in the far right side of this plot, showcasing its general inefficiency. 18 [PITH_FULL_IMAGE:figures/full_fig_p018_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Impact of batch size on hit recovery for different thresholds across datasets for all acquisition functions. 19 [PITH_FULL_IMAGE:figures/full_fig_p019_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Trade-off between CHR@10 and SMAPE@10 on Perturbation-Pathways (4D) across acquisition functions, batch sizes, and thresholds. 20 [PITH_FULL_IMAGE:figures/full_fig_p020_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Learning curves for different acquisition functions across datasets and batch sizes for τ = 5%. 0.0 0.2 0.4 0.6 0.8 1.0 CHR@10 (batch=2) Sine (1D) Sine (2D) Branin-Hoo (2D) Perturbation-Pathways (4D) Perturbation-SEM (6D) 0.0 0.2 0.4 0.6 0.8 1.0 CHR@10 (batch=5) 0 2 4 6 8 Cycle 0.0 0.2 0.4 0.6 0.8 1.0 CHR@10 (batch=10) 0 2 4 6 8 Cycle 0 2 4 6 8 Cycle 0 2 4 6 8 Cycle 0 2 4 6 8 Cycle Probability-of-Hit Thomp… view at source ↗
Figure 10
Figure 10. Figure 10: Learning curves for different acquisition functions across datasets and batch sizes for τ = 20%. 21 [PITH_FULL_IMAGE:figures/full_fig_p021_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Learning curves for different acquisition functions across datasets and batch sizes for τ = 20%. E.2. Real-World Datasets The Schmidt IL-2 dataset with Achilles features shows the strongest Probability-Hit advantage: +58.8 hits (+11.4%). This dataset has the lowest local smoothness (S = 0.21) in our complexity analysis, suggesting Probability-Hit excels when the phenotype landscape is learnable [PITH_FUL… view at source ↗
read the original abstract

High-throughput gene perturbation experiments can test several genetic interventions in parallel, yet experimental budgets remain limited. A central goal is hit discovery: identifying as many perturbations as possible whose phenotypic effect exceeds a predefined threshold. Pure exploration strategies are statistically inefficient, wasting budget on low-value regions. Bayesian optimization methods offer a principled alternative but target a single global optimum, over-exploiting dominant modes while neglecting other high-value regions. We formalize hit discovery as a sequential experimental design problem and propose Probability-of-Hit, an acquisition function that directly targets threshold exceedance by ranking candidates according to their posterior probability of being a hit. We prove asymptotic optimality of this approach and demonstrate strong empirical performance on both synthetic benchmarks and real biological immunology datasets, including up to 6.4% improvement over baselines on the Schmidt IL-2 dataset.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript formalizes hit discovery in high-throughput gene perturbation experiments as a sequential experimental design problem. It introduces the Probability-of-Hit (PoH) acquisition function that ranks candidates by their posterior probability of exceeding a fixed phenotypic threshold τ, proves asymptotic optimality of this ranking strategy, and reports empirical gains (including up to 6.4% improvement over baselines) on synthetic benchmarks and real immunology datasets such as Schmidt IL-2.

Significance. If the asymptotic optimality holds under the paper's modeling assumptions and the empirical results prove robust to surrogate misspecification, the work offers a targeted alternative to standard Bayesian optimization for multi-hit discovery tasks. This could improve sample efficiency in biological screening by prioritizing threshold exceedance over single-mode exploitation, with potential impact on experimental design in perturbation biology.

major comments (3)
  1. [Section 4] Section 4 (Theoretical Analysis), Theorem 1: The asymptotic optimality proof for PoH appears to invoke posterior consistency of the surrogate (likely a GP) to guarantee that ranking by P(f(x) > τ | D) yields optimal hit discovery. In high-dimensional noisy perturbation data, such models are typically misspecified; please explicitly state the function class, noise, and consistency assumptions under which the proof holds and provide a robustness discussion or counterexample analysis when these fail, as this is load-bearing for the central theoretical claim.
  2. [Section 5.3] Section 5.3 (Real Data Experiments), Table 3 or equivalent results table: The 6.4% improvement on the Schmidt IL-2 dataset is reported, but it is unclear whether the threshold τ was fixed a priori (as required by the problem statement) or selected post-hoc, and whether hit definitions or data exclusion criteria were applied identically across all methods and runs. Clarify these choices and include sensitivity analysis to τ, since arbitrary threshold selection directly affects whether the reported gains validate the method.
  3. [Section 3.2] Section 3.2 (Acquisition Function Definition): PoH is defined as the posterior probability of threshold exceedance. Specify the exact surrogate model, how the probability is computed (closed-form, MC sampling, etc.), and any approximations used in high dimensions, as these details are necessary to assess both the proof and reproducibility of the empirical results.
minor comments (2)
  1. [Notation] Notation: Ensure the posterior probability symbol (e.g., P(hit|x,D)) is used consistently and defined precisely in the problem formulation section to avoid ambiguity with related quantities like expected improvement.
  2. [Figures] Figures: In regret or hit-discovery curves (e.g., Figure 4 or 5), include standard error bars across multiple random seeds to convey variability, and label axes with explicit units or normalized scales.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their constructive comments on our manuscript. We have carefully considered each point and provide detailed responses below. Where appropriate, we have revised the manuscript to address the concerns raised.

read point-by-point responses
  1. Referee: [Section 4] Section 4 (Theoretical Analysis), Theorem 1: The asymptotic optimality proof for PoH appears to invoke posterior consistency of the surrogate (likely a GP) to guarantee that ranking by P(f(x) > τ | D) yields optimal hit discovery. In high-dimensional noisy perturbation data, such models are typically misspecified; please explicitly state the function class, noise, and consistency assumptions under which the proof holds and provide a robustness discussion or counterexample analysis when these fail, as this is load-bearing for the central theoretical claim.

    Authors: We agree that the assumptions should be stated explicitly. In the revised manuscript, we will add a paragraph in Section 4 specifying the function class (Gaussian Process with continuous kernel such as Matérn), noise model (homoscedastic Gaussian), and consistency conditions drawn from standard GP regression theory. We will also add a robustness discussion noting that empirical gains persist on real immunology data despite likely misspecification, supporting practical utility. A full counterexample analysis lies outside the paper's scope, but the added discussion addresses the load-bearing aspect of the claim. revision: yes

  2. Referee: [Section 5.3] Section 5.3 (Real Data Experiments), Table 3 or equivalent results table: The 6.4% improvement on the Schmidt IL-2 dataset is reported, but it is unclear whether the threshold τ was fixed a priori (as required by the problem statement) or selected post-hoc, and whether hit definitions or data exclusion criteria were applied identically across all methods and runs. Clarify these choices and include sensitivity analysis to τ, since arbitrary threshold selection directly affects whether the reported gains validate the method.

    Authors: The threshold τ was fixed a priori using domain-specific biological criteria from the Schmidt IL-2 data source. Hit definitions and data exclusion criteria were applied identically to all methods and runs. In the revision we will state this explicitly in Section 5.3 and add a sensitivity analysis over a range of τ values, confirming that relative gains remain stable. revision: yes

  3. Referee: [Section 3.2] Section 3.2 (Acquisition Function Definition): PoH is defined as the posterior probability of threshold exceedance. Specify the exact surrogate model, how the probability is computed (closed-form, MC sampling, etc.), and any approximations used in high dimensions, as these details are necessary to assess both the proof and reproducibility of the empirical results.

    Authors: We use a Gaussian Process surrogate (kernel and hyperparameter details in Section 3.1). The probability is obtained via Monte Carlo sampling from the posterior predictive distribution. In high dimensions we apply standard sparse GP approximations for tractability. We will expand Section 3.2 with these exact computational details and pseudocode to improve reproducibility. revision: yes

Circularity Check

0 steps flagged

No circularity: Probability-of-Hit is a direct definition, not a reduction to fitted inputs

full rationale

The paper defines hit discovery as sequential design and introduces Probability-of-Hit as an acquisition function that ranks points by their posterior probability of exceeding a fixed threshold τ. This is a straightforward application of standard posterior inference rather than a quantity fitted to the target metric and then relabeled as a prediction. The claimed asymptotic optimality is presented as a separate proof step whose details are not visible in the abstract, but nothing in the provided text indicates that the proof or the function reduces by construction to the data, to a self-citation chain, or to an ansatz imported from the authors' prior work. No equations are shown that equate the acquisition value to a fitted parameter or that rename an existing empirical pattern. The derivation therefore remains self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Review based on abstract only; no explicit free parameters, axioms, or invented entities are stated in the provided text.

pith-pipeline@v0.9.0 · 5457 in / 1166 out tokens · 50585 ms · 2026-05-12T03:35:40.384489+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

69 extracted references · 69 canonical work pages

  1. [1]

    Langley , title =

    P. Langley , title =. Proceedings of the 17th International Conference on Machine Learning (ICML 2000) , address =. 2000 , pages =

  2. [2]

    T. M. Mitchell. The Need for Biases in Learning Generalizations. 1980

  3. [3]

    M. J. Kearns , title =

  4. [4]

    Machine Learning: An Artificial Intelligence Approach, Vol. I. 1983

  5. [5]

    R. O. Duda and P. E. Hart and D. G. Stork. Pattern Classification. 2000

  6. [6]

    Suppressed for Anonymity , author=

  7. [7]

    Newell and P

    A. Newell and P. S. Rosenbloom. Mechanisms of Skill Acquisition and the Law of Practice. Cognitive Skills and Their Acquisition. 1981

  8. [8]

    A. L. Samuel. Some Studies in Machine Learning Using the Game of Checkers. IBM Journal of Research and Development. 1959

  9. [9]

    Stem cell reviews and reports , volume=

    Genome editing in induced pluripotent stem cells using CRISPR/Cas9 , author=. Stem cell reviews and reports , volume=. 2018 , publisher=

  10. [10]

    Current opinion in systems biology , volume=

    Exploring intermediate cell states through the lens of single cells , author=. Current opinion in systems biology , volume=. 2018 , publisher=

  11. [11]

    Genome biology , volume=

    CHESS: a new human gene catalog curated from thousands of large-scale RNA sequencing experiments reveals extensive transcriptional noise , author=. Genome biology , volume=. 2018 , publisher=

  12. [12]

    Discovery medicine , volume=

    The cost of new drug discovery and development , author=. Discovery medicine , volume=

  13. [13]

    Journal of health economics , volume=

    Innovation in the pharmaceutical industry: new estimates of R&D costs , author=. Journal of health economics , volume=. 2016 , publisher=

  14. [14]

    Genome research , volume=

    Defining cell types and states with single-cell genomics , author=. Genome research , volume=. 2015 , publisher=

  15. [15]

    Biostatistics , volume=

    Estimation of clinical trial success rates and related parameters , author=. Biostatistics , volume=. 2019 , publisher=

  16. [16]

    Annual review of genomics and human genetics , volume=

    Single-cell (multi) omics technologies , author=. Annual review of genomics and human genetics , volume=. 2018 , publisher=

  17. [17]

    Frontiers in oncology , volume=

    The unique molecular and cellular microenvironment of ovarian cancer , author=. Frontiers in oncology , volume=. 2017 , publisher=

  18. [18]

    Nature Reviews Genetics , volume=

    Using next-generation sequencing to isolate mutant genes from forward genetic screens , author=. Nature Reviews Genetics , volume=. 2014 , publisher=

  19. [19]

    Science , volume=

    CRISPR technology: A decade of genome editing is only the beginning , author=. Science , volume=. 2023 , publisher=

  20. [20]

    British journal of pharmacology , volume=

    Principles of early drug discovery , author=. British journal of pharmacology , volume=. 2011 , publisher=

  21. [21]

    Nature Reviews Genetics , volume=

    A new era in functional genomics screens , author=. Nature Reviews Genetics , volume=. 2022 , publisher=

  22. [22]

    Science , volume=

    The emerging era of cell engineering: Harnessing the modularity of cells to program complex biological function , author=. Science , volume=. 2022 , publisher=

  23. [23]

    International Conference on Research in Computational Molecular Biology , pages=

    Sequential optimal experimental design of perturbation screens guided by multi-modal priors , author=. International Conference on Research in Computational Molecular Biology , pages=. 2024 , organization=

  24. [24]

    Nature Biotechnology , volume=

    Predicting transcriptional outcomes of novel multigene perturbations with GEARS , author=. Nature Biotechnology , volume=. 2024 , publisher=

  25. [25]

    Gaussian process optimization in the bandit setting: No regret and experimental design,

    Gaussian process optimization in the bandit setting: No regret and experimental design , author=. arXiv preprint arXiv:0912.3995 , year=

  26. [26]

    Journal of Global optimization , volume=

    Efficient global optimization of expensive black-box functions , author=. Journal of Global optimization , volume=. 1998 , publisher=

  27. [27]

    Advances in neural information processing systems , volume=

    Batchbald: Efficient and diverse batch acquisition for deep bayesian active learning , author=. Advances in neural information processing systems , volume=

  28. [28]

    International Conference on Machine Learning , pages=

    DiscoBAX: Discovery of optimal intervention sets in genomic experiment design , author=. International Conference on Machine Learning , pages=. 2023 , organization=

  29. [29]

    International Conference on Machine Learning , pages=

    Bayesian algorithm execution: Estimating computable properties of black-box functions using mutual information , author=. International Conference on Machine Learning , pages=. 2021 , organization=

  30. [30]

    international conference on machine learning , pages=

    Dropout as a bayesian approximation: Representing model uncertainty in deep learning , author=. international conference on machine learning , pages=. 2016 , organization=

  31. [31]

    GeneDisco: A Benchmark for Experimental Design in Drug Discovery , year=

    Mehrjou, Arash and Soleymani, Ashkan and Jesson, Andrew and Notin, Pascal and Gal, Yarin and Bauer, Stefan and Schwab, Patrick , booktitle=. GeneDisco: A Benchmark for Experimental Design in Drug Discovery , year=

  32. [32]

    Science , volume=

    CRISPR activation and interference screens decode stimulation responses in primary human T cells , author=. Science , volume=

  33. [33]

    Communications Biology , volume=

    Genome-wide CRISPR screen identifies protein pathways modulating tau protein levels in neurons , author=. Communications Biology , volume=

  34. [34]

    Nature Communications , volume=

    A genome-wide CRISPR screen identifies host factors that regulate SARS-CoV-2 entry , author=. Nature Communications , volume=

  35. [35]

    Frontiers in Immunology , volume=

    Genome-wide CRISPR screen reveals cancer cell resistance to NK cells induced by NK-derived IFN-γ , author=. Frontiers in Immunology , volume=

  36. [36]

    Cell , volume=

    Defining a cancer dependency map , author=. Cell , volume=

  37. [37]

    Nucleic Acids Research , volume=

    STRING v11: protein--protein association networks with increased coverage , author=. Nucleic Acids Research , volume=

  38. [38]

    BioRxiv , pages=

    Extracting biological insights from the project achilles genome-scale CRISPR screens in cancer cell lines , author=. BioRxiv , pages=. 2019 , publisher=

  39. [39]

    Proceedings of the 17th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval , pages=

    A sequential algorithm for training text classifiers , author=. Proceedings of the 17th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval , pages=

  40. [40]

    Neural Computation , volume=

    Information-based objective functions for active data selection , author=. Neural Computation , volume=

  41. [41]

    2006 , publisher=

    Gaussian Processes for Machine Learning , author=. 2006 , publisher=

  42. [42]

    1996 , school=

    Bayesian Learning for Neural Networks , author=. 1996 , school=

  43. [43]

    International Conference on Machine Learning , pages=

    Weight uncertainty in neural networks , author=. International Conference on Machine Learning , pages=

  44. [44]

    Advances in Neural Information Processing Systems , volume=

    Simple and scalable predictive uncertainty estimation using deep ensembles , author=. Advances in Neural Information Processing Systems , volume=

  45. [45]

    International Conference on Machine Learning , pages=

    On kernelized multi-armed bandits , author=. International Conference on Machine Learning , pages=. 2017 , organization=

  46. [46]

    Proceedings of the 24th International Conference on Artificial Intelligence and Statistics (AISTATS) , year =

    On Information Gain and Regret Bounds in Gaussian Process Bandits , author =. Proceedings of the 24th International Conference on Artificial Intelligence and Statistics (AISTATS) , year =

  47. [47]

    Nature Cell Biology , volume =

    Decoding heterogeneous single-cell perturbation responses , author =. Nature Cell Biology , volume =. 2025 , doi =

  48. [48]

    Nature Biotechnology , year =

    Systema: a framework for evaluating genetic perturbation response prediction beyond systematic variation , author =. Nature Biotechnology , year =

  49. [49]

    Computational and Structural Biotechnology Journal , volume =

    A mini-review on perturbation modelling across single-cell omic modalities , author =. Computational and Structural Biotechnology Journal , volume =

  50. [50]

    Nature Machine Intelligence , volume =

    Active learning for optimal intervention design in causal models , author =. Nature Machine Intelligence , volume =. 2023 , doi =

  51. [51]

    Nature Communications , volume =

    Large scale active-learning-guided exploration for in vitro protein production optimization , author =. Nature Communications , volume =. 2020 , doi =

  52. [52]

    Proceedings of the National Academy of Sciences , volume =

    Machine learning-assisted directed evolution with combinatorial libraries , author =. Proceedings of the National Academy of Sciences , volume =. 2019 , doi =

  53. [53]

    arXiv preprint arXiv:2509.19988 , year =

    BioBO: Biology-informed Bayesian Optimization for Perturbation Design , author =. arXiv preprint arXiv:2509.19988 , year =

  54. [54]

    Advances in Neural Information Processing Systems , year =

    Amortized Bayesian Experimental Design for Decision-Making , author =. Advances in Neural Information Processing Systems , year =

  55. [55]

    Scientific Reports , volume =

    Knowledge graph-aided Bayesian active learning for top- K genetic interaction discovery , author =. Scientific Reports , volume =. 2025 , doi =

  56. [56]

    Nature Communications , volume =

    A Bayesian active learning platform for scalable combination drug screens , author =. Nature Communications , volume =. 2025 , doi =

  57. [57]

    Nature Genetics , volume =

    A genome-wide CRISPR screen identifies CALCOCO2 as a regulator of beta cell function influencing type 2 diabetes risk , author =. Nature Genetics , volume =. 2023 , doi =

  58. [58]

    Cell , volume =

    High-Resolution CRISPR Screens Reveal Fitness Genes and Genotype-Specific Cancer Liabilities , author =. Cell , volume =. 2015 , doi =

  59. [59]

    Proceedings of the IEEE , volume =

    Taking the Human Out of the Loop: A Review of Bayesian Optimization , author =. Proceedings of the IEEE , volume =. 2016 , doi =

  60. [60]

    Proceedings of the 27th International Conference on Machine Learning , year =

    Gaussian Process Optimization in the Bandit Setting: No Regret and Experimental Design , author =. Proceedings of the 27th International Conference on Machine Learning , year =

  61. [61]

    Proceedings of the 23rd Conference on Learning Theory , year =

    Best Arm Identification in Multi-Armed Bandits , author =. Proceedings of the 23rd Conference on Learning Theory , year =

  62. [62]

    arXiv preprint arXiv:0802.2655 , year =

    Pure Exploration in Multi-Armed Bandits Problems , author =. arXiv preprint arXiv:0802.2655 , year =

  63. [63]

    Proceedings of the 27th Conference on Learning Theory , year =

    lil’UCB: An Optimal Exploration Algorithm for Multi-Armed Bandits , author =. Proceedings of the 27th Conference on Learning Theory , year =

  64. [64]

    Proceedings of the 23rd International Joint Conference on Artificial Intelligence , year =

    Active Learning for Level Set Estimation , author =. Proceedings of the 23rd International Joint Conference on Artificial Intelligence , year =

  65. [65]

    Neural Computation , year =

    Active Learning for Distributionally Robust Level-Set Estimation , author =. Neural Computation , year =

  66. [66]

    SIAM/ASA Journal on Uncertainty Quantification , volume =

    Quantifying Uncertainties on Excursion Sets Under a Gaussian Random Field Prior , author =. SIAM/ASA Journal on Uncertainty Quantification , volume =

  67. [67]

    Foundations and Trends in Machine Learning , volume =

    A Tutorial on Thompson Sampling , author =. Foundations and Trends in Machine Learning , volume =

  68. [68]

    Proceedings of the 20th Machine Learning in Computational Biology Meeting , series =

    PerTurboAgent: An LLM-based Agent for Designing Iterative Perturb-Seq Experiments , author =. Proceedings of the 20th Machine Learning in Computational Biology Meeting , series =

  69. [69]

    Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD) , pages =

    A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise , author =. Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD) , pages =. 1996 , address =