Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CPU limits #48

Closed
lystrata opened this issue Dec 7, 2023 · 1 comment
Closed

CPU limits #48

lystrata opened this issue Dec 7, 2023 · 1 comment
Labels
question Further information is requested

Comments

@lystrata
Copy link

lystrata commented Dec 7, 2023

On an 8 Core system, I see that onprem pegs 4 coress @100% while the 0ther 4 cores are relatively inactive. Is this an intentional design limit? Or can we enable the code to make use of more cores?

@amaiya amaiya added the question Further information is requested label Dec 8, 2023
@amaiya
Copy link
Owner

amaiya commented Dec 8, 2023

The default is to use half of your CPUs and is set by llama-cpp-python. You can change it by supplying the n_threads parameter to LLM:

from onprem import LLM
llm = LLM(n_threads=desired_number_of_cores)

@amaiya amaiya closed this as completed Dec 8, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants