We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 608e9e6 commit 224ec40Copy full SHA for 224ec40
vllm/model_executor/models/llama_eagle3.py
@@ -249,8 +249,7 @@ def get_input_embeddings(
249
multimodal_embeddings: Optional[NestedTensors] = None,
250
is_multimodal: Optional[torch.Tensor] = None,
251
) -> torch.Tensor:
252
- # The llama3 drafter only processes text embeddings
253
- return self.model.embed_tokens(input_ids)
+ return self.model.get_input_embeddings(input_ids)
254
255
def forward(
256
self,
0 commit comments