Enhancing Agentic Autonomous Scientific Discovery with Vision-Language Model Capabilities (2511.14631v1)
Abstract: We show that multi-agent systems guided by vision-LLMs (VLMs) improve end-to-end autonomous scientific discovery. By treating plots as verifiable checkpoints, a VLM-as-a-judge evaluates figures against dynamically generated domain-specific rubrics, enabling agents to correct their own errors and steer exploratory data analysis in real-time. Case studies in cosmology and astrochemistry demonstrate recovery from faulty reasoning paths and adaptation to new datasets without human intervention. On a 10-task benchmark for data-driven discovery, VLM-augmented systems achieve pass at 1 scores of 0.7-0.8, compared to 0.2-0.3 for code-only and 0.4-0.5 for code-and-text baselines, while also providing auditable reasoning traces that improve interpretability. Code available here: https://github.com/CMBAgents/cmbagent
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days freePaper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
Enhancing Agentic Autonomous Scientific Discovery with Vision–LLM Capabilities — Explained Simply
Overview
This paper shows how a team of AI assistants can do better science on their own by “looking” at graphs, not just reading text or writing code. The key idea is to use a special kind of AI, called a vision–LLM (VLM), that understands both pictures and words. The system treats plots (graphs) like checkpoints. Each time it makes a plot, a VLM acts like a teacher with a checklist and decides if the plot makes scientific sense. If not, the AI fixes its mistakes. If the plot shows something new and unexpected, the AI launches more experiments to explore it.
Key Questions
The paper asks:
- Can AI scientists use plots, the same way human scientists do, to catch mistakes and make discoveries?
- Does adding a VLM “judge” to a multi-agent AI system improve accuracy and reliability?
- Can this approach work across different scientific fields without needing constant human help?
How the System Works (Methods)
Think of the system as a research lab where each AI agent has a role (planner, coder, reviewer, scientist). They follow a plan, write code, run experiments, and make plots. Now add a VLM that looks at the plots and gives feedback.
Here’s the simple cycle:
- The coding agents make a plot from data.
- The VLM acts as a judge, using a “rubric” (a checklist like teachers use) built from the task’s scientific goals.
- The VLM says either “continue” (looks good) or “retry” (something’s off), and explains why.
The system can run in two modes:
- Correction mode: If the plot looks wrong, a “Plot Debugger” agent reads the VLM’s feedback and suggests specific code fixes (for example, “Change this line; you scaled the data twice”). The coders try again and re-check the plot.
- Discovery mode: If the plot shows something interesting, a “Plot Scientist” agent suggests running a mini-experiment series to test different explanations. An “Experiment Proposer” agent picks a few model ideas and a fair scoring rule (like a balanced score that rewards good fit without over-complicating the model). The coders run them, make comparison plots, and the VLM helps pick the winner.
Key terms in plain language:
- Vision–LLM (VLM): An AI that understands both images (like plots) and text. Here, it judges whether a graph matches what science would expect.
- Rubric: A customized checklist that describes what a correct plot should look like and what scientific features to check (for example, “Is the main peak in the right place?”).
- Residuals: The difference between the data and the model’s predictions. Random-looking residuals are good; patterns often mean the model is missing something.
- Model comparison score (like BIC): A number that helps pick the best model by balancing accuracy (fits the data well) and simplicity (doesn’t add unnecessary parts).
Main Results and Why They Matter
Case Study 1: Fixing a Cosmology Plot
- Task: Make a standard plot of tiny ripples in the early universe (from the Cosmic Microwave Background). Scientists know what this plot should roughly look like.
- What happened: The first AI-made plot was wrong: peaks were in the wrong place and too large.
- How it was fixed: The VLM noticed these issues and said “retry.” The debugger found a simple coding mistake—the data was scaled twice. After one fix, the plot matched expectations. This shows the VLM-as-judge can catch subtle, domain-specific errors quickly.
Case Study 2: Discovering the Right Spectral Line Model
- Task: Fit a curve to a spectrum (a graph scientists use to paper chemicals in space). The starting guess was “one simple bump” (a single Gaussian).
- What happened: The VLM saw a dip in the middle and non-random residuals—signs that the simple model was wrong.
- What the system did: It proposed several alternative models (two bumps, absorption in the middle, etc.), ran them all, scored them, and picked the best. The winner was a “self-absorption” model, which correctly explained the data. This shows the system can explore, compare, and choose, like a scientist would.
Benchmark Performance
To test the system more broadly, the authors created 10 scientific tasks across areas like physics, chemistry, epidemiology, and cosmology. They measured “pass@1,” which is simply: If you run the system once, what’s the chance it gets the right discovery?
- Code-only systems: pass@1 about 0.2–0.3 (poor).
- Text-based multi-agent systems without vision: pass@1 about 0.4–0.5 (better but still misses subtle visual clues).
- With VLM (vision) checkpoints: pass@1 about 0.7–0.8 (much better).
Why this matters:
- Many scientific insights live in plots. Humans naturally spot patterns and anomalies in figures. VLMs help AI do the same.
- The visual checkpoints also create an “audit trail”: you can see why a decision was made by looking at the plots and the VLM’s comments, not just a final number.
What This Means for the Future (Implications)
- More reliable AI scientists: By adding “eyes” (vision) and a clear checklist (rubrics), AI agents make fewer mistakes and recover faster when they do.
- Faster exploration: The discovery mode guides the AI to run the right follow-up experiments when something interesting pops up in a plot.
- Better trust and transparency: Because the system explains its decisions with plots and structured reasoning, human researchers can check the logic step by step.
- Broad usefulness: The approach is “domain-agnostic,” meaning it can be applied to many scientific fields that rely on data and figures.
In short, this paper shows a practical way to make AI better at doing real science: treat plots as checkpoints, let a vision-enabled AI judge them, fix errors when needed, and launch smart follow-up experiments when the data suggests something new. This makes AI research systems both more accurate and easier to understand.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
Below is a single, consolidated list of what remains missing, uncertain, or unexplored in the paper, framed as concrete items that future researchers can act on.
- Benchmark scale and statistical rigor:
- Pass@1 is reported on only 10 tasks without confidence intervals; expand the benchmark size, run multiple seeds per task, and report statistical significance and variability (e.g., CIs, bootstrapped error bars).
- Quantify sensitivity of results to dataset difficulty and heterogeneity across domains (oscillators, spectral lines, SEIR, cosmology).
- Dataset provenance and realism:
- The benchmark relies on author-generated synthetic datasets; evaluate on real, public datasets with independent ground truth and expert annotations.
- Measure robustness to realistic perturbations (non-Gaussian noise, missing data, calibration errors, instrument systematics) and domain-specific confounders.
- Failure-mode analysis:
- Provide a detailed taxonomy of failures (e.g., Q5, Q7, Q9 where systems still failed) and root-cause analyses; identify whether failures stem from rubric generation, VLM visual recognition limits, experiment proposal design, or code execution issues.
- Investigate how often VLM feedback misclassifies scientifically relevant features versus noise-driven artifacts and quantify false-positive/false-negative rates.
- Ablation and component contributions:
- Isolate the performance impact of each element (visual checkpoints, Pydantic schemas, deterministic routing, separation of Plot Scientist vs Experiment Proposer, black-box judging) via controlled ablation studies.
- Compare multimodal judging to text-only judging under identical workflows to quantify the marginal benefit of vision versus structured orchestration.
- Generalization across plot types and styles:
- Test VLM judging across diverse visualization modalities (scatter plots with error bars, multi-panel figures, heatmaps, contour maps, images, 3D surfaces, log-scaled axes, unconventional styles) to assess style sensitivity and coverage.
- Evaluate robustness to aesthetic variations (fonts, colors, line styles, annotation density) and the risk that visual polish biases VLM judgments.
- Rubric generation reliability:
- The GPT-4o-generated, “domain-specific” rubrics may embed incorrect priors; quantify rubric accuracy, detect contradictions, and introduce validation (e.g., cross-rubric consistency checks, external knowledge-base verification).
- Study the impact of rubric errors on downstream decisions (continue vs retry/explore) and design mitigation (rubric ensemble, calibration, confidence scoring).
- Black-box plot judging risks:
- The judge does not inspect code; assess vulnerability to “rubric gaming” (plots that visually satisfy criteria while underlying computations are wrong).
- Explore code–plot cross-validation (e.g., symbolic checks, unit tests, physics-based invariants) to prevent deceptive or misleading figures.
- Exploration policy and statistical safeguards:
- Formalize criteria and thresholds for “explore” vs “continue” decisions; measure trade-offs between exploration depth, computational cost, and success rates.
- Incorporate multiple-hypothesis testing controls (e.g., false discovery rate, pre-registration of experiments), out-of-sample validation, and model comparison robustness (beyond BIC/χ²) to avoid chasing noise.
- Interpretability claims:
- Claims of improved interpretability via visual checkpoints are qualitative; develop quantitative interpretability metrics (trace completeness, decision auditability scores, alignment with expert ratings) and evaluate them.
- Compare visual-audit logs to text-only rationales in controlled user studies with domain experts.
- Cost, latency, and scaling:
- Report compute, token usage, wall-clock latency, and cost across workflows; characterize scalability and efficiency bottlenecks in multi-agent/VLM loops.
- Investigate caching, incremental rendering, and early-stopping policies to reduce overhead in repeated judge–debug loops.
- Model diversity and openness:
- Evaluate broader VLM/LLM families (including open-source VLMs) to assess cross-model generalization, sensitivity to training corpus biases, and reproducibility.
- Study inter-model disagreement and propose ensemble or arbitration strategies for more reliable judgments.
- Domain transfer and breadth:
- Extend beyond the tested domains to high-dimensional data (e.g., genomics, materials, climate), spatial-temporal models, and tasks with complex physical constraints to evaluate transferability.
- Test on figure types typical in each field (e.g., sky maps, phase-space diagrams, confusion matrices for epidemiology) with domain-specific criteria.
- Loop stability and convergence:
- Analyze the dynamics of the self-correction loop (maximum retries, convergence rates, oscillations) and propose adaptive policies (e.g., confidence-weighted stopping, diminishing returns criteria).
- Characterize conditions under which repeated VLM feedback leads to improvement versus instability or overfitting to rubric specifics.
- Human-in-the-loop baselines:
- Include controlled comparisons to expert-in-the-loop workflows (e.g., semi-automated systems with targeted human interventions) to contextualize autonomy claims and performance ceilings.
- Establish criteria for when escalation to human review is warranted (e.g., low rubric confidence, conflicting model rankings).
- Security and safety in execution:
- The system runs code locally; assess sandboxing, dependency management, and prevention of unsafe code generation.
- Develop policies for handling untrusted datasets, malicious inputs, or adversarial figures intended to mislead VLM judges.
- Reproducibility and release:
- Provide complete benchmark specifications (dataset generation scripts, seeds, prompts, agent configurations, rubric schemas) and end-to-end pipelines to enable exact replication.
- Document versioning and configuration management for AG2 orchestration, agent prompts, and VLM/LLM APIs.
- Scientific validity checks:
- Integrate physics or domain constraints as hard checks (e.g., conservation laws, parameter bounds, dimensional analysis) to complement VLM visual judgments.
- Evaluate how domain-theory constraints interact with VLM feedback to reduce erroneous exploration and improve correction accuracy.
- Ethical and bias considerations:
- Assess whether VLMs preferentially detect features common in their training corpus, potentially biasing discovery toward familiar patterns; design debiasing strategies or calibration methods.
- Examine the risk of reinforcing conventional expectations via rubrics, thereby missing novel or atypical phenomena.
Practical Applications
Immediate Applications
Below are actionable use cases that can be deployed now, drawing directly from the paper’s VLM-as-a-judge checkpoints, correction and discovery workflows, and multi-agent orchestration.
- Scientific research (academia; physics, astronomy, chemistry, ecology)
- Plot quality assurance in analysis pipelines: add “visual checkpoints” that auto-validate figures (axes, scaling, expected physical features) and block downstream steps when issues are detected.
- Tools/workflows: Jupyter/VS Code extension for “plot linting”; GitHub Actions that run the Plot Judge and fail CI on “retry.”
- Assumptions/dependencies: access to a VLM (e.g., GPT‑4o, Gemini 2.5 Pro); clear plotting conventions; Python runtime and plotting libraries.
- Automated self-debugging for analysis code: route faulty plots to a Plot Debugger that pinpoints root causes (e.g., incorrect rescaling) and proposes minimal code diffs.
- Tools/workflows: AG2-based multi-agent “engineering team” with structured Pydantic schemas; git-diff style patching.
- Assumptions/dependencies: deterministic routing with schemas; permissioned code execution; developer acceptance of auto-applied patches.
- Guided exploratory data analysis (EDA) via visual anomalies: when plots deviate from priors, automatically propose and run alternative models, compare with BIC/χ², and select winners with auditable reasoning.
- Tools/workflows: Experiment Proposer + Plot Scientist agents; standardized metrics and overlay plots.
- Assumptions/dependencies: suitable evaluation metrics; compute for running model variants; curated priors.
- Industrial R&D and QA (materials, spectroscopy, semiconductors, pharma/biotech)
- Spectral and signal analysis copilots: VLM-guided detection of misfits (e.g., self-absorption dips, multimodal peaks) and automated testing of line-shape alternatives.
- Tools/products: “SpectraGuard” plugin for lab software (Matplotlib/Plotly, instrument vendors) that flags plots and suggests candidate models.
- Assumptions/dependencies: domain-appropriate rubrics; integration with instrument data formats; on-prem or private VLM access.
- Process monitoring and anomaly triage: apply visual checkpoints on QC dashboards; escalate to exploratory branches that quantify and explain deviations.
- Tools/workflows: MLOps/DevOps pipeline hooks tied to Plot Judge verdicts; runbooks that map anomalies to predefined experiments.
- Assumptions/dependencies: reliable labeling/units on dashboards; tolerance for false positives; policy for automated reruns.
- Healthcare analytics (operations, labs, non-diagnostic research)
- QC for lab assay curves and device time-series plots: detect axis errors, mis-calibration signatures, or suspicious residual patterns; trigger comparative fits.
- Tools/products: “AssayPlot QA” module embedded in LIMS/ELN systems.
- Assumptions/dependencies: strict non-diagnostic use unless validated; data privacy; regulated change control for automation.
- Patient-operations dashboards: VLM checks on trends/seasonality artifacts (e.g., in hospital census) with automatic “what-if” fits.
- Tools/workflows: BI integrations with Plot Scientist for descriptive analytics.
- Assumptions/dependencies: governance controls; human in the loop for decisions.
- Finance and econometrics
- Backtest and risk-report validation: VLM-checked plots (P&L curves, drawdowns, factor loadings) to flag suspicious artifacts and recommend stress tests or out-of-sample variants.
- Tools/products: “Visual Guardrails for Backtests” CI step in quant research repos.
- Assumptions/dependencies: private deployment; domain-tailored rubrics to avoid fragile heuristics; oversight to prevent gaming.
- Software/data platforms (data science teams, SaaS analytics)
- Visual unit tests in CI/CD: treat new plots as test artifacts and enforce pass/fail based on schema-driven rubrics.
- Tools/workflows: pytest plugins for plotting; metadata tagging of figures; artifact stores with rubric verdicts and traces.
- Assumptions/dependencies: plot determinism for reproducibility; storage for images and logs.
- Notebook assistants that propose experiments: one-click generation of alternative analyses and comparison plots, with structured, auditable reasoning.
- Tools/products: Jupyter/Colab add-ons leveraging AG2 orchestration.
- Assumptions/dependencies: compute budget; user acceptance of agent-inserted cells.
- Education and citizen science
- Auto-feedback on lab assignments and research projects: evaluate student plots against domain rubrics and suggest targeted fixes.
- Tools/products: LMS plugins; grading assistants with visual criteria.
- Assumptions/dependencies: rubric templates per course; guard against overfitting to rubric wording.
- Reproducible narratives from analyses: generate human-readable summaries that cite plots and metrics, improving scientific communication skills.
- Tools/workflows: Experiment Proposer’s narrative compression used as a teaching aid.
- Assumptions/dependencies: alignment with course policies on AI assistance.
- Publishing and peer review support
- Pre-submission figure checks: ensure axes, units, and expected trends match domain norms; attach visual audit trails.
- Tools/products: “Figure QA” submission plugin; journal-side triage assistant for reviewers.
- Assumptions/dependencies: policy acceptance; discipline-specific guideline libraries.
Long-Term Applications
These use cases require further validation, scaling, domain adaptation, or regulatory alignment before routine deployment.
- Autonomous discovery assistants in high-stakes science (drug discovery, materials, climate, high-energy physics)
- Closed-loop experiment planning: VLM-driven detection of emergent features in plots from instruments/robots that trigger new experiments and model updates without human intervention.
- Potential products: lab automation stacks that integrate Plot Scientist with robotic platforms; “self-driving lab” extensions.
- Dependencies: robust lab integration; safety interlocks; formal verification of agent decisions; domain-specific VLMs trained on scientific plots.
- Real-time monitoring of critical infrastructure (energy grids, manufacturing lines, aerospace)
- VLM-as-judge for streaming visual analytics: detect subtle anomalies (e.g., spectral signatures, oscillations) and propose diagnostic tests on the fly.
- Potential workflows: digital twins that accept visual checkpoints as state validation and launch targeted simulations.
- Dependencies: latency constraints; high reliability; adversarial robustness; well-defined escalation paths to human operators.
- Regulatory-grade audit trails for AI-driven research and analytics
- Standardized “visual checkpoints” and schema-logged rubrics as part of compliance packages (GLP/GMP in pharma, model risk management in finance).
- Potential products: audit-log services that store figures, rubrics, verdicts, and code diffs for traceability.
- Dependencies: accepted standards; secure storage; third-party certification; model governance frameworks.
- Domain-specialized, on-prem or open VLMs for scientific visualization
- Training/finetuning VLMs on discipline-specific figures and conventions to improve reliability and reduce reliance on proprietary APIs.
- Potential tools: astro-VisVLM, chem-VisVLM, bio-VisVLM; adapters for Matplotlib/Plotly grammar.
- Dependencies: curated corpora with image-text pairs; compute for finetuning; continuous evaluation suites.
- Robustness, safety, and interpretability standards for agentic systems
- Benchmarks and protocols for evaluating multi-agent, vision-guided discovery (beyond pass@1), including stress tests for spurious correlations and adversarial plots.
- Potential products: standardized evaluation toolkits and datasets; certification services.
- Dependencies: community consensus; reproducible, compute-efficient benchmarks; cross-institution collaboration.
- Human–AI co-authorship pipelines
- Integrated planning–execution–visual checkpoint workflows that draft full papers with visual audit trails and structured, domain-informed narratives.
- Potential tools: paper-writing suites where figures, rubrics, and experiment logs are embedded as supplementary materials.
- Dependencies: publisher policies; provenance tracking; plagiarism and data fabrication safeguards.
- Cross-modal “visual grammar” standards for scientific agents
- A common schema for expressing expected visual features and scientific priors, portable across plotting libraries and sectors.
- Potential products: open specification for visual rubrics; validators and linters for figures.
- Dependencies: standards bodies or community working groups; wide toolchain support.
- Widespread adoption in education and public data portals
- Visual-checkpointed reports for citizen-facing dashboards (public health, environment) with clear narratives and uncertainty communication.
- Potential tools: CMS integrations that auto-generate accessible explanations when anomalies are detected.
- Dependencies: careful UX; bias mitigation; governance over automated messaging.
- Integration with digital labs and observatories
- Agents that schedule telescope/lab time based on plot-derived signals, optimizing observation or experiment portfolios.
- Potential workflows: multi-agent scheduling informed by VLM verdicts and expected value of information.
- Dependencies: operational constraints; fairness and priority policies; simulation-backed decision policies.
Notes on feasibility across applications
- Performance and cost depend on access to capable VLMs (e.g., GPT‑4o, Gemini 2.5 Pro) or strong open alternatives; on-prem options may be needed for sensitive data.
- Reliability hinges on rubric quality and domain priors; mis-specified rubrics or out-of-distribution plots can cause false positives/negatives.
- Deterministic orchestration and structured schemas (Pydantic, AG2) reduce brittleness but require engineering effort and maintenance.
- In regulated domains (healthcare, finance, pharma), automated decisions should remain decision-support with human oversight until validated.
- Plot reproducibility (fixed seeds, deterministic rendering) and artifact storage are necessary for CI and auditing.
- Data privacy, IP protection, and latency constraints may limit cloud VLM usage; consider hybrid or on-prem deployments.
Glossary
- Acoustic peak: A peak in the CMB power spectrum caused by acoustic oscillations in the early universe. "For the first acoustic peak, it noted that theory predicts a maximum near with amplitude ."
- AG2: An agent orchestration framework used for routing, handoffs, and message management in multi-agent systems. "The orchestration, handoffs, and message routing are implemented with AG2."
- Bayesian Information Criterion (BIC): A model selection criterion that balances fit quality with penalization for model complexity. "The agent specifies the Bayesian Information Criterion (BIC) as the tiebreaking evaluation metric."
- Behavioral feedback: An extension in epidemiological models where population behavior influences disease dynamics. "an SEIR model with behavioral feedback."
- CAMB: Cosmology code (Code for Anisotropies in the Microwave Background) for computing theoretical CMB spectra. "Using CAMB, compute the lensed CMB temperature–temperature (TT) power spectrum..."
- Chi-squared (): A goodness-of-fit statistic measuring the discrepancy between observed data and a model. "Each set is accompanied by a comparison metric (e.g., , WAIC, or likelihood score)"
- Chirped harmonic oscillator: An oscillator whose frequency changes over time, producing a chirp signal. "a damped harmonic oscillator and a chirped harmonic oscillator, respectively."
- Cosmic Microwave Background (CMB): The relic radiation from the early universe, used to infer cosmological parameters. "a CMB EE power spectrum"
- Damped harmonic oscillator: An oscillator with energy loss causing amplitude decay over time. "a damped harmonic oscillator and a chirped harmonic oscillator, respectively."
- Damping tail: The high- suppression in the CMB power spectrum due to diffusion damping (Silk damping). "and for the high- damping tail."
- Dℓ (): The rescaled angular CMB power spectrum defined by . "plot for multipoles on linear axes with units of ."
- EE power spectrum: The CMB polarization power spectrum for E-mode polarization. "a CMB EE power spectrum"
- Experiment Proposer: A specialized agent that designs alternative experiments and specifies evaluation metrics. "The Experiment Proposer agent generates a set of three to five candidate experiments, including the baseline, that test alternative models, parameter values, or analysis methods."
- Hubble constant: The parameter describing the current rate of cosmic expansion. "a CDM spectrum with an incorrect Hubble constant."
- Hyperfine structure: Spectral line splitting due to interactions between nuclear spin and electronic magnetic fields. "emission with hyperfine structure."
- Likelihood score: A measure of how probable the observed data are under a given model, used for model comparison. "Each set is accompanied by a comparison metric (e.g., , WAIC, or likelihood score)"
- ΛCDM (Lambda-CDM): The standard cosmological model with dark energy (Λ) and cold dark matter (CDM). "with the Planck 2018 best-fit CDM parameters"
- Lensed: Refers to gravitational lensing effects that modify observed CMB spectra. "compute the lensed CMB temperature–temperature (TT) power spectrum"
- Monte Carlo Graph Search: A search algorithm that explores graph-structured decision spaces using stochastic sampling. "applies a Monte Carlo Graph Search–based generation algorithm with reinforcement learning and human feedback to efficiently explore the visualization space."
- Multipoles: The spherical harmonic indices () that index angular scales in CMB power spectra. "for multipoles "
- pass@1: A benchmark metric indicating the probability that a single run succeeds. "pass@1 scores of 0.7â0.8"
- Planck 2018: A widely used set of cosmological parameters derived from the Planck satellite’s 2018 data release. "with the Planck 2018 best-fit CDM parameters"
- Posterior parameter estimates: Parameter values inferred from the posterior distribution in Bayesian analysis. "returning posterior parameter estimates, fit statistics (, reduced , and BIC), and residual summary statistics."
- Pydantic schema: A Python data modeling approach for defining and validating structured outputs. "evaluates it against a structured rubric defined by a Pydantic schema."
- Residuals: The differences between observed data and model predictions, often analyzed to diagnose model fit. "the residuals form a strong âWâ pattern with deviations up to "
- Scalar spectral index: The parameter () describing the tilt of the primordial scalar power spectrum. "a CDM spectrum with an incorrect scalar spectral index"
- SEIR model: An epidemiological compartmental model with Susceptible, Exposed, Infectious, and Recovered compartments. "a standard SEIR model"
- Self-absorption: When emitted radiation is absorbed by the same medium, producing a central dip in spectral lines. "emission with self-absorption"
- TT power spectrum: The temperature–temperature angular power spectrum of the CMB. "temperature–temperature (TT) power spectrum"
- Vision-LLM (VLM): A model that jointly processes images and text to reason about visual content. "vision-LLMs (VLMs) improve end-to-end autonomous scientific discovery."
- WAIC: Widely Applicable Information Criterion, a Bayesian model comparison metric. "Each set is accompanied by a comparison metric (e.g., , WAIC, or likelihood score)"
Collections
Sign up for free to add this paper to one or more collections.