-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Robust qv_scope_split_at implementation #9
Comments
You can reuse the guided mode implementation from Hsplit if you wish because I think the functionality seems rather similar to me (this does not answer @eleon's comment). |
Please re-test to see if 2703e88 fixes this issue. |
Thank you, @samuelkgutierrez . Unfortunately, this is still an issue. Here's an example.
Perhaps, this will be solved by the affinity preserving CPU/GPU algorithms for split. |
Thank you for testing, @eleon. Yes, an affinity preserving algorithm should fix this issue. |
Greetings @samuelkgutierrez There are still issues with the latest build. Same test machine as above, same command as above:
|
Can you please try again by modifying the test to use |
No, you will have to modify the test code to use |
Thank you, @samuelkgutierrez! Using
The issues are with the following calls:
|
Thank you, @eleon. Can you push the changes you made so I can see what's going on? Regarding the second issue, are both GPUs attached to the package containing cores 0-17? |
Greetings, @samuelkgutierrez --I pushed the changes to |
Yikes! I didn't notice. My apologies. If both GPUs are attached to the same socket, then |
Sounds good, @samuelkgutierrez. Thank you! |
Adding another thought before I forget. The way I see the regular |
Maybe |
Possibly, @samuelkgutierrez. It's just that I think about it the opposite way: |
Good morning, @samuelkgutierrez . Progress! but some issues too. Case 1: Not using
The GPUs are split correctly among the MPI workers. However, the assigned CPUs are not local to the assigned GPUs. Case 2: Using
Two main issues here:
Thanks! |
Another subtlety about the
Let's say we split at GPUs. What I'm looking to answer is can the resulting subscopes have more than one GPU? Here's my desired behavior. Let's say we have a node with 3 GPUs and a job with 2 tasks.
|
Let’s say we have a compute node with 3 GPUs and 3 NUMA domains. The 3 GPUs hang off the first NUMA domain. When I use split_at(..., HW_OBJ_GPUs, ...) I would expect the subscopes to be derived from NUMA 0, but most likely this implementation will derive one subscope per NUMA. f2d5cee
The text was updated successfully, but these errors were encountered: