Skip to content

Arbitrary-timestep Transformer for Enhanced Multi-frame Interpolation and Synthesis

License

Notifications You must be signed in to change notification settings

starboi-63/ArTEMIS

Repository files navigation

ArTEMIS

Arbitrary timestep Transformer for Enhanced Multi-frame Interpolation and Synthesis.

If you are interested in the model architecture and results, please visit ArTEMIS's project website.

Getting Started

ArTEMIS is a deep learning model that interleaves VFIT and EDSC together in order to enable the synthesis of intermediate video frames at arbitrary timesteps while using multiple frames on either side of the target timestep as context. The model is trained on the Vimeo-90K Septuplet dataset.

Prerequisites

ArTEMIS requires CUDA to execute. If your GPU does not support CUDA, then we also provide a notebook file which can be uploaded to Google Colab and executed there. If executing locally, ensure that you have Python installed on your system. To use ArTEMIS, you need to set up a Python environment with the necessary packages installed. You can do this by running the following commands in your terminal.

First, clone the repository to your local machine.

git clone https://github.com/starboi-63/ArTEMIS.git

Next, navigate to the project directory.

cd ArTEMIS

Then, install virtualenv if you don't already have it.

pip install virtualenv

Create a new virtual environment named artemis-env in the project directory.

python -m venv artemis-env

Activate the virtual environment.

source artemis-env/bin/activate

Install the required packages, which are listed in the requirements.txt file.

pip install -r requirements.txt

Finally, depending on your hardware, you will need to install the appropriate versions of the CUDA Toolkit, cuDNN, and cupy. To see what the maximum version of CUDA your GPU supports, you can check the output of the following command:

nvidia-smi

Vimeo-90K Septuplet Dataset

To train or test the model with default settings, you will need to download the "The original training + test set (82GB)" version of the Vimeo-90K Septuplet dataset. This data can be found on the official website at http://toflow.csail.mit.edu/.

You can also run the following command in your terminal to download the dataset. This will take some time, as the dataset is quite large.

wget http://data.csail.mit.edu/tofu/dataset/vimeo_septuplet.zip

Then, unzip the downloaded file. This will also take a few minutes.

unzip vimeo_septuplet.zip

Usage

To use ArTEMIS, you can run main.py in your terminal with the appropriate command line arguments. For a full list of command line arguments, execute:

python main.py --help

There are three modes in which you can run the model: train, test, and interpolate. The train and test modes are used to train/test the model on the Vimeo-90K Septuplet dataset respectively. Finally, the interpolate mode is used to generate interpolated frames between a single window of four context frames.

For the train and test modes, the following command line arguments will be critical.

  • --model: The model to use. Right now, we have only implemented the ArTEMIS model.
  • --mode: The mode in which to run the model. This can be either train or test.
  • --dataset: The dataset to use. Right now, we have only implemented the vimeo90K_septuplet dataset.
  • --data_dir: The directory containing the Vimeo-90K Septuplet dataset.
  • --output_dir: The directory to periodically save some output frames to while training or testing.
  • --use_checkpoint: Whether to use a checkpoint to initialize the model.
  • --checkpoint_dir: The directory containing the checkpoint file.
  • --log_dir: The directory to save logs to while training/testing.
  • --log_iter: The frequency at which to log training information and save outputs (default = 100 steps).
  • --batch_size: The batch size to use while training or testing.

For the interpolate mode, the following command line arguments must be used.

  • --model: The model to use. Right now, we have only implemented the ArTEMIS model.
  • --mode: The mode in which to run the model. Should be set to interpolate.
  • --model_path: The path to the pre-trained model checkpoint (model.ckpt available on Google Drive).
  • --frame1_path: The path to the first context frame (before the interpolated frame in time).
  • --frame2_path: The path to the second context frame (before the interpolated frame in time).
  • --frame3_path: The path to the third context frame (after the interpolated frame in time).
  • --frame4_path: The path to the fourth context frame (after the interpolated frame in time).
  • --timesteps: A comma-separated list of timesteps in the range (0,1) to interpolate frames for (e.g. '0.25, 0.5, 0.75').
  • --save_path: The directory to save the interpolated frames to.

For example, to train the model, you can run the following command:

python main.py --model ArTEMIS --mode train --data_dir <data_dir> --output_dir <output_dir> --log_dir <log_dir> --use_checkpoint --checkpoint_dir <checkpoint_dir> --batch_size <batch_size>

Alternatively, to generate intermediate frames for a single window of context frames, you can run:

python main.py --model ArTEMIS --mode interpolate --model_path <model_path> --frame1_path <frame1_path> --frame2_path <frame2_path> --frame3_path <frame3_path> --frame4_path <frame4_path> --timesteps <timesteps> --save_path <save_path>

About

Arbitrary-timestep Transformer for Enhanced Multi-frame Interpolation and Synthesis

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •