From 77cc2a1a4f3c4d0a155039761ca2f4d77a081a49 Mon Sep 17 00:00:00 2001 From: Riccardo Orlando Date: Fri, 19 May 2023 10:22:36 +0200 Subject: [PATCH] Expose transformer model `config` --- setup.py | 2 +- transformers_embedder/embedder.py | 6 +++--- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/setup.py b/setup.py index beee333..28fa401 100644 --- a/setup.py +++ b/setup.py @@ -12,7 +12,7 @@ setuptools.setup( name="transformers_embedder", - version="3.0.9", + version="3.0.10", author="Riccardo Orlando", author_email="orlandoricc@gmail.com", description="Word level transformer based embeddings", diff --git a/transformers_embedder/embedder.py b/transformers_embedder/embedder.py index d96e6b1..beb4bc6 100644 --- a/transformers_embedder/embedder.py +++ b/transformers_embedder/embedder.py @@ -75,7 +75,7 @@ def __init__( ) -> None: super().__init__() if isinstance(model, str): - config = tr.AutoConfig.from_pretrained( + self.config = tr.AutoConfig.from_pretrained( model, output_hidden_states=True, output_attentions=True, @@ -84,11 +84,11 @@ def __init__( ) if from_pretrained: self.transformer_model = tr.AutoModel.from_pretrained( - model, config=config, *args, **kwargs + model, config=self.config, *args, **kwargs ) else: self.transformer_model = tr.AutoModel.from_config( - config, *args, **kwargs + self.config, *args, **kwargs ) else: self.transformer_model = model