-
Notifications
You must be signed in to change notification settings - Fork 23
Open
Description
Hi developers,
Thanks for this awesome project again! Recently I am porting my cryptographic application to a machine with four GPUs. Specifically, I have two possible scenarios in mind:
- Operations (e.g. addition, multiplication, rotation) are sent one at a time, and each operation is processed by all four GPUs before moving to the next operation.
- There are a lot of (independent and parallelizable) operations to process, so each operation is processed by a single GPU and all four GPUs work simultaneously on different operations.
Since the current examples begin with cudaSetDevice(0)
which means the computation is limited to a single GPU, I would like to ask whether it is currently supported (or planned to be supported) to utilize multiple GPUs in parallel, according to either of the two scenarios above? If so, could you please give me some hints on how to implement it?
Thanks a lot in advance!
Metadata
Metadata
Assignees
Labels
No labels