Skip to content

Conversation

honghua
Copy link

@honghua honghua commented Sep 27, 2024

No description provided.

Harry Yang added 2 commits September 27, 2024 20:36
1. on my VM running inference with CPU is only using one core, instead of all 16.
delete `torch.set_default_tensor_type(torch.BFloat16Tensor)` make it using
all 16 cores

2. reduce default max_seq_len to 128
* `device` is undefined
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant