Skip to content

CVRL/GrainsOfSaliency

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Grains of Saliency

Official repository for the paper: Colton R. Crum, Samuel Webster, Adam Czajka, "Grains of Saliency: Optimizing Saliency-based Training of Biometric Attack Detection Models," IEEE/IAPR International Joint Conference on Biometrics (IJCB), September 15-18, 2024, Buffalo, NY, USA

Paper: IEEEXplore | ArXiv pre-print

Abstract

Incorporating human-perceptual intelligence into model training has shown to increase the generalization capability of models in several difficult biometric tasks, such as presentation attack detection (PAD) and detection of synthetic samples. After the initial collection phase, human visual saliency (e.g., eye-tracking data, or handwritten annotations) can be integrated into model training through attention mechanisms, augmented training samples, or through human perception-related components of loss functions. Despite their successes, a vital, but seemingly neglected, aspect of any saliency-based training is the level of salience granularity (e.g., bounding boxes, single saliency maps, or saliency aggregated from multiple subjects) necessary to find a balance between reaping the full benefits of human saliency and the cost of its collection. In this paper, we explore several different levels of salience granularity and demonstrate that increased generalization capabilities of PAD and synthetic face detection can be achieved by using simple yet effective saliency post-processing techniques across several different CNNs.

Installation and Usage

How to run & requirements

  • Build the conda environment using grains.yml
  • Train or test CNNs (DenseNet, ResNet, and Inception) using ./training_code/train.py and ./testing_code/test.py

Model weights

Model weights and salience granularity for all architectures and biometric tasks can be downloaded from this Box folder.

The best performing iris PAD model (reported in the IJCB 2024 paper) was Inception AOI, which achieved AUC=0.975 on the [LivDet-2020 test set]. That specific model's weights are stored in:

./Iris-PAD/inception/AOI/inception_segmentation_mse_20_0.50_2/Logs/final_model.pth

(after unzipping "Model weights.zip" located in the Box folder linked above).

Citations

If you find this work useful in your research, please cite the following paper:

@InProceedings{Crum_IJCB_2024,
  author    = {Colton R. Crum and Samuel Webster and Adam Czajka},
  booktitle = {The IEEE/IAPR International Joint Conference on Biometrics (IJCB)},
  title     = {{Grains of Saliency: Optimizing Saliency-based Training of Biometric Attack Detection Models}},
  year      = {2024},
  address   = {Buffalo, NY, USA},
  month     = {September 15-18},
  pages     = {1-8},
  publisher = {IEEE}
}

This GitHub repository:

@Misc{ND_GrainsOfSaliency_GitHub,
  howpublished = {\url{https://github.com/CVRL/GrainsOfSaliency/}},
  note         = {Accessed: X},
  title        = {{Grains of Saliency: Optimizing Saliency-based Training of Biometric Attack Detection Models (IJCB 2024 paper repository)}},
  author       = {Colton R. Crum and Samuel Webster and Adam Czajka},
}

Acknowledgment

This material is based upon work partially supported by the National Science Foundation under Grant No. 2237880. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages