While Large Language Models (LLMs) demonstrate remarkable capabilities in scientific tasks such as literature analysis and experimental design (e.g., accurately extracting key findings from papers or generating coherent experimental procedures), existing evaluation benchmarks primarily assess performance using rich contextual inputs. We introduce LiveIdeaBench, a comprehensive benchmark evaluating LLMs' scientific idea generation by assessing divergent thinking capabilities using single-keyword prompts. Drawing from Guilford's creativity theory, our benchmark employs a dynamic panel of state-of-the-art LLMs to assess generated ideas across five key dimensions: originality, feasibility, fluency, flexibility, and clarity. Through extensive experimentation with over 40 leading models across 1,180 keywords spanning 22 scientific domains, we reveal that the scientific idea generation capabilities measured by our benchmark, are poorly predicted by standard metrics of general intelligence. Our results demonstrate that models like QwQ-32B-preview achieve creative performance comparable to top-tier models such as claude-3.7-sonnet:thinking, despite significant gaps in their general intelligence scores. These findings highlight the need for specialized evaluation benchmarks for scientific idea generation and suggest that enhancing these idea generation capabilities in LLMs may require different training strategies than those used for improving general problem-solving abilities, potentially enabling a wider range of AI tools tailored for different stages of the scientific process.
We are pleased to announce that, based on the invaluable feedback from reviewers, we have enhanced our benchmark by upgrading it to version 2. This update introduces a new dimension—Clarity—and improves the prompts, evaluation process (including the rejection handling mechanism), making our benchmark more comprehensive and objective.
This v2 version of the benchmark incorporates the latest models, including: claude-3.7-sonnet:thinking, o3-mini-high, gpt-4.5-preview, qwq-32b, deepseek-r1, gemini-2.0-flash-thinking, and a total of 41 state-of-the-art models.
Check it out here: https://huggingface.co/datasets/6cf/liveideabench-v2
We are excited to announce that the latest dataset, including supplementary tests for models like deepseek-R1, deepseek-V3, minimax-01, phi-4, and Opus, has been uploaded to Hugging Face! 🚀
Check it out here: https://huggingface.co/datasets/6cf/liveideabench-DLC-250127
Model | Organization | Fluency | Feasibility | Originality | Flexibility | Average | Open |
---|---|---|---|---|---|---|---|
google/gemini-pro-1.5 🥇 | 8.88 | 6.84 | 7.31 | 7.67 | 7.67 | 🔒 | |
o1-preview 🥈 | OpenAI | 9.07 | 6.58 | 7.34 | 7.67 | 7.66 | 🔒 |
qwen/qwq-32b-preview 🥉 | Alibaba | 9.12 | 6.94 | 6.73 | 7.33 | 7.53 | ✅ |
anthropic/claude-3.5-sonnet | Anthropic | 8.93 | 5.95 | 7.86 | 7.22 | 7.49 | 🔒 |
google/gemini-2.0-flash-exp | 8.72 | 6.78 | 7.07 | 7.33 | 7.48 | 🔒 | |
openai/gpt-4o-2024-11-20 | OpenAI | 8.37 | 6.34 | 7.59 | 7.00 | 7.33 | 🔒 |
mistralai/mistral-large-2411 | Mistral AI | 8.52 | 6.82 | 6.92 | 7.00 | 7.31 | ✅ |
amazon/nova-pro-v1 | Amazon | 8.50 | 7.05 | 6.57 | 7.00 | 7.28 | 🔒 |
nvidia/llama-3.1-nemotron-70b-instruct | NVIDIA | 8.21 | 6.34 | 7.54 | 6.89 | 7.24 | ✅ |
qwen/qwen-2.5-coder-32b-instruct | Alibaba | 8.43 | 6.65 | 6.90 | 6.78 | 7.19 | ✅ |
meta-llama/llama-3.1-405b-instruct | Meta | 8.28 | 6.31 | 7.04 | 6.67 | 7.07 | ✅ |
sammcj/qwen2.5-dracarys2-72b:Q4_K_M | Abacus.AI | 7.98 | 6.91 | 6.64 | 6.56 | 7.02 | ✅ |
openai/o1-mini | OpenAI | 7.55 | 6.88 | 7.15 | 6.44 | 7.00 | 🔒 |
qwen/qwen-2.5-72b-instruct | Alibaba | 7.90 | 6.75 | 6.74 | 6.56 | 6.99 | ✅ |
step-2-16k | Other | 7.97 | 6.67 | 6.28 | 6.33 | 6.81 | 🔒 |
anthropic/claude-3.5-haiku | Anthropic | 7.58 | 5.64 | 7.74 | 6.22 | 6.80 | 🔒 |
x-ai/grok-2-1212 | xAI | 7.56 | 6.60 | 6.83 | 6.11 | 6.78 | 🔒 |
openai/gpt-4o-mini | OpenAI | 7.10 | 6.87 | 6.76 | 6.11 | 6.71 | 🔒 |
deepseek/deepseek-chat | DeepSeek | 7.02 | 6.37 | 7.19 | 6.11 | 6.67 | ✅ |
meta-llama/llama-3.3-70b-instruct | Meta | 7.25 | 6.70 | 6.35 | 6.11 | 6.60 | ✅ |
@article{ruan2024liveideabench,
title={LiveIdeaBench: Evaluating LLMs' Scientific Creativity and Idea Generation with Minimal Context},
author={Ruan, Kai and Wang, Xuan and Hong, Jixiang and Sun, Hao},
journal={arXiv preprint arXiv:2412.17596},
year={2024}
}