From 032895f1954d01c4f902b07e960630e62f6dac24 Mon Sep 17 00:00:00 2001 From: Evs91 <8013869+Evs91@users.noreply.github.com> Date: Tue, 6 Jan 2026 14:49:04 -0500 Subject: [PATCH 1/3] Update nodes.py imports for LTX1 embedding module ComfyUI updated location of import to new Embeddings1DConnector module. From 4a6b705db5ba0766030441baeb6291548efca636 Mon Sep 17 00:00:00 2001 From: Evs91 <8013869+Evs91@users.noreply.github.com> Date: Wed, 7 Jan 2026 18:22:22 -0500 Subject: [PATCH 2/3] update nodes.py for new embeddings module. --- nodes.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/nodes.py b/nodes.py index 04bff4a..18f39d0 100644 --- a/nodes.py +++ b/nodes.py @@ -9,7 +9,7 @@ from unittest.mock import patch from comfy.ldm.flux.layers import timestep_embedding, apply_mod -from comfy.ldm.lightricks.model import precompute_freqs_cis +from comfy.ldm.lightricks.model.embeddings_connector import Embeddings1DConnector from comfy.ldm.lightricks.symmetric_patchifier import latent_to_pixel_coords from comfy.ldm.wan.model import sinusoidal_embedding_1d @@ -628,7 +628,7 @@ def teacache_ltxvmodel_forward( if attention_mask is not None and not torch.is_floating_point(attention_mask): attention_mask = (attention_mask - 1).to(x.dtype).reshape((attention_mask.shape[0], 1, -1, attention_mask.shape[-1])) * torch.finfo(x.dtype).max - pe = precompute_freqs_cis(fractional_coords, dim=self.inner_dim, out_dtype=x.dtype) + pe = Embeddings1DConnector.precompute_freqs_cis(fractional_coords, dim=self.inner_dim, out_dtype=x.dtype) batch_size = x.shape[0] timestep, embedded_timestep = self.adaln_single( From 44e2e26bb746836bf77ef3386057409033ab5ee9 Mon Sep 17 00:00:00 2001 From: Evs91 <8013869+Evs91@users.noreply.github.com> Date: Wed, 7 Jan 2026 19:26:01 -0500 Subject: [PATCH 3/3] Update nodes.py --- nodes.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/nodes.py b/nodes.py index 18f39d0..be1c23b 100644 --- a/nodes.py +++ b/nodes.py @@ -9,7 +9,7 @@ from unittest.mock import patch from comfy.ldm.flux.layers import timestep_embedding, apply_mod -from comfy.ldm.lightricks.model.embeddings_connector import Embeddings1DConnector +from comfy.ldm.lightricks.embeddings_connector import Embeddings1DConnector from comfy.ldm.lightricks.symmetric_patchifier import latent_to_pixel_coords from comfy.ldm.wan.model import sinusoidal_embedding_1d