Skip to content

prs-eth/v3

Repository files navigation

Continuous Space-Time Video Super-Resolution with 3D Fourier Fields

ICLR 2026 (accepted)

Page Paper License

This is the official implementation of the paper

Continuous Space-Time Video Super-Resolution with 3D Fourier Fields
Alexander Becker, Julius Erbach, Dominik Narnhofer, Konrad Schindler
ETH Zurich

News

2026-01-26: Paper is accepted to ICLR 2026. 🎉
2025-09-30: Paper is now on arXiv.

Setup

This code is tested on Linux with Python 3.11 and an NVIDIA GPU with > 20GB VRAM (e.g., {3,4,5}090 or H100). Install the required packages with

pip install -r requirements.txt

Then build deformable attention CUDA extensions for JAX:

cd models/rvrt_jax/deform_attn && ./build.sh && cd -

Evaluate pre-trained models

Download pre-trained V3 checkpoint from Google Drive and place them in the checkpoints/ folder.

An example command for running inference on the Vid4 dataset (x2 temporal upsampling) is:

python run_inference.py --data-dir data/ --checkpoint-path checkpoints/v3.pkl --eval-sets Vid4 --space-scale 4 --time-scale 2

Make sure that --data-dir points to a folder containing the dataset (i.e., in this case with a subfolder Vid4), and that --checkpoint-path points to a downloaded pre-trained model. For the default Adobe240 and GoPro benchmarks you likely want to use --space-scale 4 --time-scale 8.

We have included some logic for quickly evaluating the relevant metrics of the inference results against the ground truth, e.g.,

python eval_imgs.py data/Vid4 results/Vid4/x4 --time-scale 2

Train your own model

To train on Adobe240, download the dataset from the official sources and then place videos under ./data with the following structure:

data/
├── Adobe240
│   ├── train
│   │   ├── ...
│   ├── valid
│   │   ├── ...
│   ├── test
│   │   ├── ...

Then use the convenience training script:

bash run_train.sh

Make sure to check run_train.sh and args.py for the availabe training options.

Reproducibility

This branch contains the current reference implementation with the recommended initialization and training settings. Results remain consistent with those reported in the paper. For exact reproduction of the paper results, please use branch v0.1-paper.

Citation

@inproceedings{
  becker2026continuous,
  title={Continuous Space-Time Video Super-Resolution with 3D Fourier Fields},
  author={Alexander Becker and Julius Erbach and Dominik Narnhofer and Konrad Schindler},
  booktitle={The Fourteenth International Conference on Learning Representations},
  year={2026}
}

About

[ICLR 2026] Continuous Space-Time Video Super-Resolution with 3D Fourier Fields

Resources

Stars

Watchers

Forks