This refers to one application's using multiple graphics-processing units, either in traditional (graphical) or general-purpose (GPGPU) applications.
Questions tagged [multi-gpu]
387 questions
9
votes
2 answers
OpenCL/OpenGL Interop with Multiple GPUs
I'm having trouble using multiple GPUs with OpenCL/OpenGL interop. I'm trying to write an application which renders the result of an intensive computation. In the end it will run an optimization problem, and then, based on the result, render…

matth
- 563
- 7
- 22
8
votes
1 answer
How to solve dist.init_process_group from hanging (or deadlocks)?
I was to set up DDP (distributed data parallel) on a DGX A100 but it doesn't work. Whenever I try to run it simply hangs. My code is super simple just spawning 4 processes for 4 gpus (for the sake of debugging I simply destroy the group immediately…

Charlie Parker
- 5,884
- 57
- 198
- 323
8
votes
2 answers
How to do parallel GPU inferencing in Tensorflow 2.0 + Keras?
Let's begin with the premise that I'm newly approaching to TensorFlow and deep learning in general.
I have TF 2.0 Keras-style model trained using tf.Model.train(), two available GPUs and I'm looking to scale down inference times.
I trained the model…

Filippo Castelli
- 81
- 1
- 3
7
votes
1 answer
Multi GPU training slower than single GPU on Tensorflow
I have created 3 virtual GPU's (have 1 GPU) and try to speedup vectorization on images. However, using provided below code with manual placement from off docs (here) I got strange results: training on all GPU two times slower than on a single one.…

Dmitriy Kisil
- 2,858
- 2
- 16
- 35
7
votes
3 answers
Multiple monitors in .NET
Are all displays returned from .NET's Screen.AllScreens regardless of hardware configuration? For example, on a single PC you can have:
Video card out to two display = total 2 displays
Video cards each out to 1 display = total 2 displays,
Video…

Nicros
- 5,031
- 12
- 57
- 101
7
votes
1 answer
Tensorflow can't detect GPU when invoked by Ray worker
When I try the following code sample for using Tensorflow with Ray, Tensorflow fails to detect the GPU's on my machine when invoked by the "remote" worker but it does find the GPU's when invoked "locally". I put "remote" and "locally" in scare…

2daaa
- 2,788
- 7
- 33
- 44
7
votes
3 answers
Can not save model using model.save following multi_gpu_model in Keras
Following the upgrade to Keras 2.0.9, I have been using the multi_gpu_model utility but I can't save my models or best weights using
model.save('path')
The error I get is
TypeError: can’t pickle module objects
I suspect there is some problem…

GhostRider
- 2,109
- 7
- 35
- 53
7
votes
1 answer
Multi GPU/Tower setup Tensorflow 1.2 Estimator
I want to turn my _model_fn for Estimator into a multi GPU solution.
Is there a way to do it within the Esitmator API or do I have to explicitly code device placement and synchronization.
I know I can use tf.device('gpu:X') to place my model on GPU…

dparted
- 395
- 1
- 2
- 15
7
votes
0 answers
nnGraph multi-GPU Torch
This question is about making any nnGraph network run on multiple GPUs and not specific to the following network instance
I am trying to train a network which is constructed with nnGraph. The backward diagram is attached. I am trying to run the…

Bharat
- 2,139
- 2
- 16
- 35
7
votes
1 answer
CUDA SDK examples throw various errors in multi-gpu system
I have a Dell Precision Rack running Ubuntu Precise and featuring two Tesla C2075 plus a Quadro 600 which is the display device. I have recently finished some tests on my desktop-computer and now tried to port stuff to the workstation.
Since CUDA…

pdresselhaus
- 679
- 14
- 32
6
votes
2 answers
How to use keras.utils.Sequence data generator with tf.distribute.MirroredStrategy for multi-gpu model training in tensorflow?
I want to train a model on several GPUs using tensorflow 2.0. In the tensorflow tutorial for distributed training (https://www.tensorflow.org/guide/distributed_training), the tf.data datagenerator is converted into a distributed dataset as follows:…

surya1495
- 65
- 1
- 4
6
votes
0 answers
Allocate two separate GPU's to different threads in python?
I want to use two python source codes, the first one is developed with tensorflow and the other is developed using pythorch. I want to run each of these codes in a thread with a separate GPU. the input of both codes are the same and there are some…

R.y.z
- 213
- 2
- 15
6
votes
2 answers
ffmpeg - cuda encode - OpenEncodeSessionEx failed: out of memory
I'm having a problem with ffmpeg video encoding using GPU (CUDA).
I have 2x nVidia GTX 1050 Ti
The problem comes when i try to do multiple parallel encodings. More than 2 processes and ffmpeg dies like this:
[h264_nvenc @ 0xcc1cc0]…

VelDev
- 160
- 2
- 7
6
votes
2 answers
Poor performance when calling cudaMalloc with 2 GPUs simultaneously
I have an application where I split the processing load among the GPUs on a user's system. Basically, there is CPU thread per GPU that initiates a GPU processing interval when triggered periodically by the main application thread.
Consider the…

rmccabe3701
- 1,418
- 13
- 31
6
votes
1 answer
How can I override the CUDA kernel execution time limit on Windows with a secondary GPUs?
From Nvidia's website, it explain the time-out problem:
Q: What is the maximum kernel execution time? On Windows, individual
GPU program launches have a maximum run time of around 5 seconds.
Exceeding this time limit usually will cause a…

user2003564
- 213
- 2
- 9