Skip to content
This repository was archived by the owner on May 11, 2025. It is now read-only.

Commit 133dd7a

Browse files
committed
Fix usage of engine
1 parent fbeea40 commit 133dd7a

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

awq/modules/fused/attn.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@ def forward(
219219
xv = xv.view((bsz,) + self.attention_shapes["single_xv_view"])
220220

221221
past_key_value = (xk, xv) if use_cache else None
222-
attention_weight = awq_inference_engine.single_query_attention(
222+
attention_weight = ft_inference_engine.single_query_attention(
223223
xq, # query
224224
xk, # key
225225
xv, # value

0 commit comments

Comments
 (0)