You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First of all, thank you for the amazing work on this framework! I’m currently using the Sentence Transformer with the BGE Small EN model for sentence encoding, but I’ve encountered an issue on my server.
My server has 8 CPUs, and the transformer seems to always utilize all of them. However, there are multiple tasks running simultaneously on the server, so I would like to limit the CPU usage to just 2 cores to avoid impacting other tasks.
I’ve attempted the following settings, but they don’t seem to have the desired effect:
Could anyone provide guidance on how to effectively limit the model to use only 2 CPUs? Additionally, I would appreciate any advice on optimizing offline inference performance using Sentence Transformers on CPUs—what’s the fastest way to achieve this with just 2 cores?
Thanks in advance for your help!
The text was updated successfully, but these errors were encountered:
Hi Everyone,
First of all, thank you for the amazing work on this framework! I’m currently using the Sentence Transformer with the BGE Small EN model for sentence encoding, but I’ve encountered an issue on my server.
My server has 8 CPUs, and the transformer seems to always utilize all of them. However, there are multiple tasks running simultaneously on the server, so I would like to limit the CPU usage to just 2 cores to avoid impacting other tasks.
I’ve attempted the following settings, but they don’t seem to have the desired effect:
Could anyone provide guidance on how to effectively limit the model to use only 2 CPUs? Additionally, I would appreciate any advice on optimizing offline inference performance using Sentence Transformers on CPUs—what’s the fastest way to achieve this with just 2 cores?
Thanks in advance for your help!
The text was updated successfully, but these errors were encountered: