Skip to content

sepermetric/seper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SePer: Official Implementation

Arxiv ICLR 2025 Spotlight python 3.11 transformer 4.40 license Apache-2.0


TL;DR: SePer is an accurate / fast / API-free metric to measure retrieval utility via information gain.

This repository contains the official implementation of the ICLR 2025 Spotlight paper:

SePer: Measure Retrieval Utility Through The Lens Of Semantic Perplexity Reduction

Authors: Lu Dai, Yijie Xu, Jinhui Ye, Hao Liu, Hui Xiong

Quick Links

Overview

SePer introduces a framework to evaluate retrieval utility by analyzing semantic perplexity and semantic perplexity reduction. This provides a more fine-grained utility signal than relying on ranking metrics or downstream answer quality alone.

Below is an illustration of SePer's fine-grained evaluation ability:

illustration

When to use SePer

SePer is especially useful when:

  1. Two retrievers have similar ranking metrics, but downstream quality differs.
  2. You want to measure whether retrieved evidence truly reduces model uncertainty.
  3. You need a utility-centric signal to complement framework-level RAG evaluation.

Installation

conda create -n seper python=3.11
conda activate seper
pip install torch
pip install -r requirements.txt

Quick Start

A minimal walkthrough is provided in example.ipynb.

Benchmark for Retrievers

The retriever benchmark is available at: https://sepermetric.github.io/

Citation

If you find our work useful, please cite:

@inproceedings{dai2025seper,
  title={SePer: Measure Retrieval Utility Through The Lens Of Semantic Perplexity Reduction},
  author={Dai, Lu and Xu, Yijie and Ye, Jinhui and Liu, Hao and Xiong, Hui},
  booktitle={International Conference on Learning Representations (ICLR)},
  year={2025},
  doi={10.48550/arXiv.2503.01478},
  url={https://openreview.net/forum?id=ixMBnOhFGd}
}

Metadata for scholarly indexing and LLM tools

This repo includes machine-readable metadata for discoverability:

  • CITATION.cff (GitHub citation support)
  • codemeta.json (CodeMeta metadata)
  • llms.txt and docs/llms-full.txt (LLM-oriented summaries)

About

SePer is an accurate / fast / free-of-API metric to measure document quality via information gain

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors