Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Can LLMs replace Neil deGrasse Tyson? Evaluating the Reliability of LLMs as Science Communicators (2409.14037v1)

Published 21 Sep 2024 in cs.CL and cs.AI
Can LLMs replace Neil deGrasse Tyson? Evaluating the Reliability of LLMs as Science Communicators

Abstract: LLMs and AI assistants driven by these models are experiencing exponential growth in usage among both expert and amateur users. In this work, we focus on evaluating the reliability of current LLMs as science communicators. Unlike existing benchmarks, our approach emphasizes assessing these models on scientific questionanswering tasks that require a nuanced understanding and awareness of answerability. We introduce a novel dataset, SCiPS-QA, comprising 742 Yes/No queries embedded in complex scientific concepts, along with a benchmarking suite that evaluates LLMs for correctness and consistency across various criteria. We benchmark three proprietary LLMs from the OpenAI GPT family and 13 open-access LLMs from the Meta Llama-2, Llama-3, and Mistral families. While most open-access models significantly underperform compared to GPT-4 Turbo, our experiments identify Llama-3-70B as a strong competitor, often surpassing GPT-4 Turbo in various evaluation aspects. We also find that even the GPT models exhibit a general incompetence in reliably verifying LLM responses. Moreover, we observe an alarming trend where human evaluators are deceived by incorrect responses from GPT-4 Turbo.

Evaluating LLMs as Reliable Science Communicators

The paper "Can LLMs replace Neil deGrasse Tyson? Evaluating the Reliability of LLMs as Science Communicators" addresses a timely and critical question concerning the effectiveness of LLMs in scientific communication. Given the rise in LLM usage among both experts and laypersons, the authors focus on determining the reliability of these models when tasked with scientific question-answering that requires nuanced understanding and self-awareness.

Core Contributions and Methodology

The authors introduce a new dataset, SCiPS-QA, composed of 742 scientifically complex Yes/No questions. This dataset is designed to be substantially challenging, demanding extensive reasoning and sophisticated knowledge retrieval. SCiPS-QA includes closed questions that have definitive answers and open questions that pertain to unresolved scientific problems. The dataset spans multiple scientific disciplines, including Physics, Chemistry, Mathematics, Theoretical Computer Science, Astronomy, Economics, and Biology.

Three proprietary models from the OpenAI GPT family and thirteen open-access models from Meta's Llama-2, Llama-3, and the Mistral family were benchmarked using SCiPS-QA. The evaluation metrics were meticulously defined to assess correctness, consistency, and hallucination propensity. Additionally, the ability of these models to abstain from answering scientifically open queries was scrutinized.

The paper aimed to answer four primary research questions:

  1. Can existing LLMs answer scientific reasoning questions successfully and faithfully?
  2. Are LLMs effective at abstaining from assertively answering scientific open problems?
  3. Can LLMs successfully verify LLM-generated responses?
  4. Can human evaluators be misled by incorrect yet confident LLM responses?

Key Findings

Reliability in Answering Scientific Questions:

  • Llama-3-70B emerged as the top-performing open-access model, often outperforming GPT-4 Turbo.
  • Despite parameter scaling, many models—including the proprietary GPT models—failed to reliably verify their responses.

Propensity to Abstain from Answering Open Questions:

  • The paper found that most models, including GPT-4 Turbo, struggled to handle open scientific queries, often failing to abstain from answering when they should.

Human Evaluator Reliability:

  • Alarmingly, humans were often deceived by incorrect yet confident responses generated by GPT-4 Turbo. This suggests a high potential for these models to propagate scientific misconceptions.

Hallucination Quantification and Evaluation:

  • Various hallucination detection techniques such as SelfCheckGPT were employed. They generally failed to flag incorrect answers reliably, with GPT-3.5 Turbo ironically performing better at detecting its own hallucinations compared to GPT-4 Turbo.

Self-Awareness and Consistency:

  • The evaluated models showed limited self-awareness and consistency in their responses. The instruction-finetuned Llama variants performed better than their non-finetuned counterparts, indicating the importance of fine-tuning in enhancing model performance.

Implications and Speculations for Future Developments

Practical Implications:

  • The evident unreliability of current LLMs in scientific communication signals a cautious adoption in academic and research settings. Given that these models can convincingly deliver incorrect information, there are substantial risks associated with their unregulated use in conveying scientific concepts and findings.

Theoretical Implications:

  • The findings support the notion that mere parameter scaling is insufficient for achieving true comprehension and accuracy in AI models. Attention to model architecture and training methodologies, potentially informed by human cognitive processes, may be necessary.

Future Developments:

  • Development of more sophisticated verification algorithms within LLMs can be a plausible avenue of research. Integrating mechanisms to better assess uncertainties in model responses might mitigate the risk of misleading confident assertions.
  • Exploration into mixed-method approaches where human oversight and automated checks are combined might help in refining the output quality from these models.
  • Advanced datasets with even more nuanced and interdisciplinary queries could be curated to push the boundaries of current LLM's reasoning capabilities.

Conclusion

The paper provides an insightful and rigorous assessment of the current capabilities and limitations of LLMs as science communicators. By introducing SCiPS-QA and employing thorough evaluation frameworks, the authors deliver compelling evidence that while LLMs hold promise, there are significant hurdles to overcome before they can be deemed reliable for scientific communication. These findings urge the AI research community to address these limitations through innovative approaches to model training, evaluation, and application.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Prasoon Bajpai (5 papers)
  2. Niladri Chatterjee (14 papers)
  3. Subhabrata Dutta (24 papers)
  4. Tanmoy Chakraborty (224 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com