Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update ginkgo interface #669

Open
wants to merge 3 commits into
base: develop
Choose a base branch
from

Conversation

fritzgoebel
Copy link

@fritzgoebel fritzgoebel commented Oct 17, 2023

This PR updates the Ginkgo interface in order to be able to handle data coming from the GPU.
Currently this means:

  • in the first solver call: Copying matrices to the CPU, transform to CSR and start factorization preprocessing on the host before moving to the GPU
  • if the matrix is updated: copying the matrix values to the CPU, move them to the CSR format and then refactorize on the GPU (this should be all moved to the GPU)
  • in the solve: handle vectors on the GPU if possible, if they are on the CPU move them to the GPU for the solve

I would be grateful for some instructions on how to test the mem_space == device @pelesh @nkoukpaizan

Note: I will update the interface further to be based on the most current Ginkgo release (1.7.0)

@fritzgoebel fritzgoebel changed the base branch from update_ginkgo_interface to develop October 17, 2023 19:26
@pelesh
Copy link
Collaborator

pelesh commented Oct 17, 2023

@fritzgoebel, all you need to do is to set GPU examples to use gpu mode. In sparse HiOp examples that would look something like this:

      if (use_ginkgo_cuda) {
          nlp.options->SetStringValue("compute_mode", "gpu");
          nlp.options->SetStringValue("ginkgo_exec", "cuda");
      } else if (use_ginkgo_hip) {
          nlp.options->SetStringValue("compute_mode", "gpu");
          nlp.options->SetStringValue("ginkgo_exec", "hip");
      } else {
          nlp.options->SetStringValue("ginkgo_exec", "reference");
      }

I think all the other options you set earlier should stay the same.

When you set compute_mode to gpu, HiOp will hand data to the linear solver on the device.

@fritzgoebel
Copy link
Author

I added ginkgo as an option for the NlpSparseRajaEx2 example and tested that this works with "mem_space" = "device" both on Summit and on Frontier. @pelesh

@nkoukpaizan
Copy link
Collaborator

I successfully tested this on Frontier with ginkgo@glu_experimental built with rocm/5.2. I haven't been able to build ginkgo@glu_experimental with more recent versions of ROCm. @fritzgoebel correct me if I'm wrong, but the interface will need to be changed again to use [email protected] (assuming that version has everything we need for HiOp).

@nychiang
Copy link
Collaborator

@nkoukpaizan what is the error you observed when using a more recent version of ROCm?

@cameronrutherford
Copy link
Collaborator

If LLNL devs / HiOp devs are happy with this PR, can we please get a PR into develop from a local branch (instead of a fork) so we can get CI running? CI should be failing here since we need a [email protected] module on CI platforms, and so testing with that would be great.

This also would make merging updates into ExaGO easier...

I can create the PR myself as well

@nychiang
Copy link
Collaborator

If LLNL devs / HiOp devs are happy with this PR, can we please get a PR into develop from a local branch (instead of a fork) so we can get CI running? CI should be failing here since we need a [email protected] module on CI platforms, and so testing with that would be great.

This also would make merging updates into ExaGO easier...

I can create the PR myself as well

@cameronrutherford @fritzgoebel Thanks! Please create a PR from a local branch (instead of a fork), in order to use the CI features. Otherwise this PR looks good to me.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants