< Explain other AI papers

ResearchBench: Benchmarking LLMs in Scientific Discovery via Inspiration-Based Task Decomposition

Yujie Liu, Zonglin Yang, Tong Xie, Jinjie Ni, Ben Gao, Yuqiang Li, Shixiang Tang, Wanli Ouyang, Erik Cambria, Dongzhan Zhou

2025-03-28

ResearchBench: Benchmarking LLMs in Scientific Discovery via
  Inspiration-Based Task Decomposition

Summary

This paper is about creating a new way to test how well AI can help scientists discover new ideas and research hypotheses.

What's the problem?

There isn't a good way to measure if AI can truly help scientists come up with new ideas for research.

What's the solution?

The researchers developed a test called ResearchBench that measures how well AI can find inspiration, create hypotheses, and rank them.

Why it matters?

This work matters because it can help scientists use AI to generate new research ideas and speed up the process of scientific discovery.

Abstract

Large language models (LLMs) have demonstrated potential in assisting scientific research, yet their ability to discover high-quality research hypotheses remains unexamined due to the lack of a dedicated benchmark. To address this gap, we introduce the first large-scale benchmark for evaluating LLMs with a near-sufficient set of sub-tasks of scientific discovery: inspiration retrieval, hypothesis composition, and hypothesis ranking. We develop an automated framework that extracts critical components - research questions, background surveys, inspirations, and hypotheses - from scientific papers across 12 disciplines, with expert validation confirming its accuracy. To prevent data contamination, we focus exclusively on papers published in 2024, ensuring minimal overlap with LLM pretraining data. Our evaluation reveals that LLMs perform well in retrieving inspirations, an out-of-distribution task, suggesting their ability to surface novel knowledge associations. This positions LLMs as "research hypothesis mines", capable of facilitating automated scientific discovery by generating innovative hypotheses at scale with minimal human intervention.