Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Include max_threads into cluster_profile. #351

Closed
rajwanir opened this issue Nov 7, 2024 · 1 comment · May be fixed by #361
Closed

Include max_threads into cluster_profile. #351

rajwanir opened this issue Nov 7, 2024 · 1 comment · May be fixed by #361
Assignees

Comments

@rajwanir
Copy link

rajwanir commented Nov 7, 2024

The current method for most rules is set with attempts. For example, threads: lambda wildcards, attempt: attempt * 2. This is not ideal for rules that are known to be compute intensive and can use max threads efficiently when provided in the first attempt.

Currently, there is no way to determine what the max_threads is for each HPC system so at the rule level threads could either be assigned by attempts or hardcoded to a maximum. Ideally, there should be a way to encode the max threads in cluster profiles at cluster_profiles/<system>/config.yaml . So that at the rule level threads can be assigned to system max for select compute intensive rules to run most efficiently.

@rajwanir rajwanir self-assigned this Nov 7, 2024
@rajwanir
Copy link
Author

rajwanir commented Nov 7, 2024

Turns out simpler than I had imagined. The known cluster_profiles now sets max-threads via config. I have currently set to what is known: 44 for CCAD2 and 56 for biowulf. The defualt slurm is set to 8 based on what some rules are already requesting in the pipeline. If a rule requests more than max-threads , it will autoscale down. If the cluster_profile do not exist for a system, a user can also set this at invocation with cgr submit --max-threads <max_threads>

rajwanir pushed a commit that referenced this issue Nov 25, 2024
@rajwanir rajwanir linked a pull request Nov 25, 2024 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant