Skip to content

Code for - Hypothesizing Missing Causal Variables with LLMs: Benchmark and Evaluation

Notifications You must be signed in to change notification settings

ivaxi0s/hypothesizing-causal-variable-llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 

Repository files navigation

Hypothesizing Missing Causal Variables with LLMs: Benchmark and Evaluation

Ivaxi Sheth, Sahar Abdelnabi, Mario Fritz

Abstract

Scientific discovery is a catalyst for human intellectual advances, driven by the cycle of hypothesis generation, experimental design, data evaluation, and iterative assumption refinement. This process, while crucial, is expensive and heavily dependent on the domain knowledge of scientists to generate hypotheses and navigate the scientific cycle. Central to this is causality, the ability to establish the relationship between the cause and the effect. Motivated by the scientific discovery process, in this work, we formulate a novel task where the input is a partial causal graph with missing variables, and the output is a hypothesis about the missing variables to complete the partial graph. We design a benchmark with varying difficulty levels and knowledge assumptions about the causal graph. With the growing interest in using Large Language Models (LLMs) to assist in scientific discovery, we benchmark open-source and closed models on our testbed. We show the strong ability of LLMs to hypothesize the mediation variables between a cause and its effect. In contrast, they underperform in hypothesizing the cause and effect variables themselves. We also observe surprising results where some of the open-source models outperform the closed GPT-4 model. %Lastly, we find that the LLMs underperform in cause and effect node type variables.

About

Code for - Hypothesizing Missing Causal Variables with LLMs: Benchmark and Evaluation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published