Skip to content

[ICML'24] Code for the loss functions and metrics presented in "Bridging mini-batch and asymptotic analysis in contrastive learning: From InfoNCE to Kernel-based losses"

Notifications You must be signed in to change notification settings

pakoromilas/DHEL-KCL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Bridging Mini-Batch and Asymptotic Analysis in Contrastive Learning: From InfoNCE to Kernel-Based Losses

InfoNCE variants demonstrate direct and indirect coupling between the alignment and uniformity terms thus hurting optimisation. We introduce the Decoupled Hyperspherical Energy Loss (DHEL) that completly decouples alignment from uniformity. We also revisit Kernel Contrastive Losses (KCL) that also decouple these terms.

DHEL and KCL:

  • outperform other InfoNCE variants, such as SimCLR and DCL, even with smaller batch sizes
  • demonstrate robustness against hyperparameters
  • effectively utilize more dimensions, mitigating the dimensionality collapse problem

Also, KCL possesses several intriguing properties:

  • the expected loss remains unaffected by the number of negative samples
  • its minima can be identified non-asymptotically.

This repository provides implementations for DHEL and KCL, as presented in the paper available here.

Additionally, it includes the metrics applied on the learned representations, such as the introduced Wasserstein distance, which measures uniformity and effective rank. The introduced metric measures the Wasserstein distance between learned and optimal similarity distributions. Unlike the conventional uniformity metric, it accurately estimates uniformity without underestimation.

Our paper's experiments were conducted using the codebase provided in this repository.

Citation


@InProceedings{pmlr-v235-koromilas24a,
  title = 	 {Bridging Mini-Batch and Asymptotic Analysis in Contrastive Learning: From {I}nfo{NCE} to Kernel-Based Losses},
  author =       {Koromilas, Panagiotis and Bouritsas, Giorgos and Giannakopoulos, Theodoros and Nicolaou, Mihalis and Panagakis, Yannis},
  booktitle = 	 {Proceedings of the 41st International Conference on Machine Learning},
  pages = 	 {25276--25301},
  year = 	 {2024},
  editor = 	 {Salakhutdinov, Ruslan and Kolter, Zico and Heller, Katherine and Weller, Adrian and Oliver, Nuria and Scarlett, Jonathan and Berkenkamp, Felix},
  volume = 	 {235},
  series = 	 {Proceedings of Machine Learning Research},
  month = 	 {21--27 Jul},
  publisher =    {PMLR},
  pdf = 	 {https://raw.githubusercontent.com/mlresearch/v235/main/assets/koromilas24a/koromilas24a.pdf},
  url = 	 {https://proceedings.mlr.press/v235/koromilas24a.html}
}

About

[ICML'24] Code for the loss functions and metrics presented in "Bridging mini-batch and asymptotic analysis in contrastive learning: From InfoNCE to Kernel-based losses"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages