A suite of model evaluation for epistemic competence. We test knowledge a language model has about contexts using a reinforcement learning environment
Relevant Research
- Epistemic Neural Networks A library for neural networks that know what they don't know(this is also called Negative Expertise). For background information, please see the paper
This project provides interactive environments and metrics for evaluating the epistemic capabilities of artificial agents.
The benchmark currently includes two types of text-based interactive games and one experimental spatial game:
Choose Your Own Adventure - Branching narrative games where the agent makes choices that affect the story trajectory. Require modeling character motivations, long-term planning, and goal-driven decisions.
Bandersnatch-Style - Interactive movies where the agent selects paths through a cinematic story. Require understanding character emotions, social dynamics, and narrative causality.
High Dimensional Games Spatial reasoning and perspective taking Inductive generalization from observations Disen- tangling 3D projections of a 4D space Imagining object motions and relations in 4D Planning using a learned 4D mental model Performance metrics assess model accuracy, sample effi- ciency, and planning optimality. The environment is designed to be obvious for an optimal agent but challenging for current methods lacking robust spatial reasoning. Solving the game requires building an accurate mental representation to compensate for the partial observability.
![image](https://private-user-images.githubusercontent.com/25654848/260972949-ecb0282b-c658-4f0f-9105-c957017ece22.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzkyNzU2OTQsIm5iZiI6MTczOTI3NTM5NCwicGF0aCI6Ii8yNTY1NDg0OC8yNjA5NzI5NDktZWNiMDI4MmItYzY1OC00ZjBmLTkxMDUtYzk1NzAxN2VjZTIyLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTElMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjExVDEyMDMxNFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTBiYTA3NTU0NDdmNWNjOTliYWIyOWUxNzI3MzJiNzg3M2VhZjkyZjRiNTU2MmMwOTI2MWMyODcyNWIwNjBhY2ImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.B4Wng3LgVUj7s514xcvjbJ0JVxDCuQkGjAet-0yos04)
Both game formats test the agent's ability to build accurate mental models from complex sequential observations and make optimal decisions through planning.
This is based on the Mechavellian Benchmark
The benchmark evaluates agents along the following epistemic dimensions:
Environment modeling - Accuracy of learned dynamics model, ability to explain environment behavior
Adaptability - Speed and accuracy of model updates in response to new observations
Social intelligence - Capacity for theory of mind and modeling other agents
Causal reasoning - Effectiveness at inferring causal relationships from events
Transfer learning - Leveraging knowledge on new games with similar dynamics
Imagination - Ability to predict hypotheticals and potential outcomes
Strategy optimization - Rational decision making given internal environment model