Paper List
-
Macroscopic Dominance from Microscopic Extremes: Symmetry Breaking in Spatial Competition
This paper addresses the fundamental question of how microscopic stochastic advantages in spatial exploration translate into macroscopic resource domi...
-
Linear Readout of Neural Manifolds with Continuous Variables
This paper addresses the core challenge of quantifying how the geometric structure of high-dimensional neural population activity (neural manifolds) d...
-
Theory of Cell Body Lensing and Phototaxis Sign Reversal in “Eyeless” Mutants of Chlamydomonas
This paper solves the core puzzle of how eyeless mutants of Chlamydomonas exhibit reversed phototaxis by quantitatively modeling the competition betwe...
-
Cross-Species Transfer Learning for Electrophysiology-to-Transcriptomics Mapping in Cortical GABAergic Interneurons
This paper addresses the challenge of predicting transcriptomic identity from electrophysiological recordings in human cortical interneurons, where li...
-
Uncovering statistical structure in large-scale neural activity with Restricted Boltzmann Machines
This paper addresses the core challenge of modeling large-scale neural population activity (1500-2000 neurons) with interpretable higher-order interac...
-
Realizing Common Random Numbers: Event-Keyed Hashing for Causally Valid Stochastic Models
This paper addresses the critical problem that standard stateful PRNG implementations in agent-based models violate causal validity by making random d...
-
A Standardized Framework for Evaluating Gene Expression Generative Models
This paper addresses the critical lack of standardized evaluation protocols for single-cell gene expression generative models, where inconsistent metr...
-
Single Molecule Localization Microscopy Challenge: A Biologically Inspired Benchmark for Long-Sequence Modeling
This paper addresses the core challenge of evaluating state-space models on biologically realistic, sparse, and stochastic temporal processes, which a...
Budget-Sensitive Discovery Scoring: A Formally Verified Framework for Evaluating AI-Guided Scientific Selection
Indian Institute of Information Technology Allahabad (IIITA) | National Institute of Electronics & Information Technology (NIELIT)
30秒速读
IN SHORT: This paper addresses the critical gap in evaluating AI-guided scientific selection strategies under realistic budget constraints, where existing metrics fail to jointly model budget limitations, asymmetric error costs, and the option to abstain.
核心创新
- Methodology Introduces the Budget-Sensitive Discovery Score (BSDS), a formally verified metric with 20 machine-checked theorems that jointly penalizes false discoveries (λ-weighted FDR) and excessive abstention (γ-weighted coverage gap) at each budget level.
- Methodology Proposes the Discovery Quality Score (DQS) as a budget-averaged summary statistic that prevents proposers from inflating scores by performing well at cherry-picked budgets.
- Biology Provides the first comprehensive evaluation showing that LLMs add no marginal value to existing ML pipelines for drug discovery candidate selection, with the simple RF-based Greedy-ML proposer achieving the best DQS (-0.046).
主要结论
- The RF-based Greedy-ML proposer achieves the best DQS (-0.046), outperforming all 39 proposers including 28 LLM configurations, demonstrating that simple ML baselines remain superior for drug discovery candidate selection.
- No LLM configuration (zero-shot or few-shot) surpasses the Greedy-ML baseline on HIV or Tox21 datasets, establishing that LLMs provide no marginal value over existing trained classifiers in realistic deployment scenarios.
- The proposer hierarchy generalizes robustly across five MoleculeNet benchmarks spanning extreme prevalence ranges (0.18%–46.2%) and a non-drug AV safety domain, with parameter robustness demonstrated across a 9×7 grid (τ≥0.636, mean τ=0.863).
摘要: Scientific discovery increasingly relies on AI systems to select candidates for expensive experimental validation, yet no principled, budget-aware evaluation framework exists for comparing selection strategies—a gap intensified by large language models (LLMs), which generate plausible scientific proposals without reliable downstream evaluation. We introduce the Budget-Sensitive Discovery Score (BSDS), a formally verified metric—20 theorems machine-checked by the Lean 4 proof assistant—that jointly penalizes false discoveries (λ-weighted FDR) and excessive abstention (γ-weighted coverage gap) at each budget level. Its budget-averaged form, the Discovery Quality Score (DQS), provides a single summary statistic that no proposer can inflate by performing well at a cherry-picked budget. As a case study, we apply BSDS/DQS to a question of broad interest: do LLMs add marginal value to an existing ML pipeline for drug discovery candidate selection? We evaluate 39 proposers—11 mechanistic variants, 14 zero-shot LLM configurations, and 14 few-shot LLM configurations—using SMILES (Simplified Molecular Input Line Entry System) representations on MoleculeNet HIV (41,127 compounds, 3.5% active, 1,000 bootstrap replicates) under both random and scaffold splits. Three findings emerge. First, the simple RF-based Greedy-ML proposer achieves the best DQS (−0.046), outperforming all MLP variants and LLM configurations; additional MLP reranking layers degrade rather than improve the RF’s discriminative ranking. Second, no LLM surpasses the Greedy-ML baseline under either zero-shot or few-shot evaluation on HIV or Tox21—establishing that LLMs provide no marginal value over an existing trained classifier, the realistic deployment scenario. Third, the proposer hierarchy generalizes across five MoleculeNet benchmarks spanning 0.18%–46.2% prevalence, a non-drug AV safety domain, and a 9×7 grid of penalty parameters (τ≥0.636, mean τ=0.863). The framework applies in principle to any setting where candidates are selected under budget constraints and asymmetric error costs, as demonstrated here across pharmaceutical screening and autonomous vehicle safety triage.