Influenciae is a Python toolkit dedicated to computing influence values for the discovery of potentially problematic samples in a dataset and the generation of data-centric explanations for deep learning models. In this library based on Tensorflow, we gather state-of-the-art methods for estimating the importance of training samples and their influence on test data-points for validating the quality of datasets and of the models trained on them.
We propose some hands-on tutorials to get familiar with the library and it's API:
- Getting Started
- Benchmarking with Mislabeled sample detection
- Using the first order influence calculator
- Using the second order influence calculator
- Using Arnoldi Influence Calculator
- Using TracIn
- Using Representer Point Selection - L2 (RPS_L2)
- Using Representer Point Selection - Local Jacobian Expansion (RPS_LJE)
- Using Boundary-based Influence
Influenciae requires a version of python 3.7 or higher and several libraries, including Tensorflow and Numpy. Installation can be done using Pypi:
pip install influenciae
Once Influenciae is installed, there are two major applications for the different modules (that all follow the same API).
So, except for group-specific functions that are only available on the influence
module, all the classes are able to compute self-influence values, the influence with one point wrt another, as well as find the top-k samples for both of these situations.
Particularly useful when validating datasets, influence functions (and related notions) allow for gaining an insight into what samples the models thinks to be "important". For this, the training dataset and a trained model are needed.
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import FirstOrderInfluenceCalculator
from deel.influenciae.utils import ORDER
# load the model, the training loss (without reduction) and the training data (with the labels and in a batched TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = FirstOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)
data_and_influence_dataset = influence_calculator.compute_influence_values(train_dataset)
# or influence_calculator.compute_top_k_from_training_dataset(train_dataset, k_samples, ORDER.DESCENDING) when the
# dataset is too large
This is also explained more in depth in the Getting Started tutotial
Another application is to explain some model's predictions by looking on which training samples they are based on. Again, the training dataset, the model and the samples we wish to explain are needed.
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import FirstOrderInfluenceCalculator
from deel.influenciae.utils import ORDER
# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in batched a TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = FirstOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)
data_and_influence_dataset = influence_calculator.estimate_influence_values_in_batches(samples_to_explain, train_dataset)
# or influence_calculator.top_k(samples_to_explain, train_dataset, k_samples, ORDER.DESCENDING) when the
# dataset is too large
This is also explained more in depth in the Getting Started tutorial
The previous examples use notions of influence that are applied individually to each data-point, but it is possible to extend this to groups. That is, answer the question of what would a model look like if it hadn't seen a whole group of data-points during training, for example. This can be computed namely using the FirstOrderInfluenceCalculator
and SecondOrderInfluenceCalculator
, for implementations where pairwise interactions between each of the data-points are not taken into account and do, respectively.
For obtaining the groups' influence:
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import SecondOrderInfluenceCalculator
# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in a batched TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = SecondOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator) # or FirstOrderInfluenceCalculator
data_and_influence_dataset = influence_calculator.estimate_influence_values_group(groups_train, groups_to_explain)
For the data-centric explanations:
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import SecondOrderInfluenceCalculator
# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in a batched TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = SecondOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator) # or FirstOrderInfluenceCalculator
data_and_influence_dataset = influence_calculator.estimate_influence_values_group(groups_train)
All the influence calculation methods work on Tensorflow models trained for any sort of task and on any type of data. Visualization functionality is implemented for image datasets only (for the moment).
Influence Method | Source | Tutorial |
---|---|---|
Influence Functions | Paper | |
RelatIF | Paper | |
Influence Functions (first order, groups) | Paper | |
Influence Functions (second order, groups) | Paper | |
Arnoldi iteration (Scaling Up Influence Functions) | Paper | |
Trac-In | Paper | |
Representer Point Selection (L2) | Paper | |
Representer Point Selection (Local Jacobian Expansion) | Paper | |
Boundary-based influence | -- |
This library proposes implementations of some of the different popular ways of calculating the influence of data-points on TF, but there are also other ones using other frameworks.
Some other tools for efficiently computing influence functions.
- Scaling Up Influence Functions a Python library using JAX implementing a scalable algorithm for computing influence functions.
- FastIF: Scalable Influence Functions for Efficient Model Interpretation and Debugging a Python library using PyTorch implementing another scalable algorithm for computing influence functions.
More from the DEEL project:
- Xplique a Python library exclusively dedicated to explaining neural networks.
- deel-lip a Python library for training k-Lipschitz neural networks on TF.
- deel-torchlip a Python library for training k-Lipschitz neural networks on PyTorch.
- DEEL White paper a summary of the DEEL team on the challenges of certifiable AI and the role of data quality, representativity and explainability for this purpose.
This library was first created as a research tool by Agustin Martin PICARD in the context of the DEEL project with the help of David Vigouroux and Thomas FEL. Later on, Lucas Hervier joined the team to transform the code base as a practical user-(almost)-friendly and efficient tool.
If you use Influenciae as part of your workflow in a scientific publication, please consider citing the 🗞️ official paper:
@unpublished{picard:hal-04284178,
TITLE = {Influenci{\ae}: A library for tracing the influence back to the data-points},
AUTHOR = {Picard, Agustin Martin and Hervier, Lucas and Fel, Thomas and Vigouroux, David},
URL = {https://hal.science/hal-04284178},
NOTE = {working paper or preprint},
YEAR = {2023},
MONTH = Nov,
KEYWORDS = {Data-centric ai ; XAI ; Explainability ; Influence Functions ; Open-source toolbox},
PDF = {https://hal.science/hal-04284178/file/ms.pdf},
HAL_ID = {hal-04284178},
HAL_VERSION = {v1},
}
The package is released under MIT license.