Questions tagged [multi-gpu]

This refers to one application's using multiple graphics-processing units, either in traditional (graphical) or general-purpose (GPGPU) applications.

387 questions
9
votes
2 answers

OpenCL/OpenGL Interop with Multiple GPUs

I'm having trouble using multiple GPUs with OpenCL/OpenGL interop. I'm trying to write an application which renders the result of an intensive computation. In the end it will run an optimization problem, and then, based on the result, render…
matth
  • 563
  • 7
  • 22
8
votes
1 answer

How to solve dist.init_process_group from hanging (or deadlocks)?

I was to set up DDP (distributed data parallel) on a DGX A100 but it doesn't work. Whenever I try to run it simply hangs. My code is super simple just spawning 4 processes for 4 gpus (for the sake of debugging I simply destroy the group immediately…
Charlie Parker
  • 5,884
  • 57
  • 198
  • 323
8
votes
2 answers

How to do parallel GPU inferencing in Tensorflow 2.0 + Keras?

Let's begin with the premise that I'm newly approaching to TensorFlow and deep learning in general. I have TF 2.0 Keras-style model trained using tf.Model.train(), two available GPUs and I'm looking to scale down inference times. I trained the model…
7
votes
1 answer

Multi GPU training slower than single GPU on Tensorflow

I have created 3 virtual GPU's (have 1 GPU) and try to speedup vectorization on images. However, using provided below code with manual placement from off docs (here) I got strange results: training on all GPU two times slower than on a single one.…
Dmitriy Kisil
  • 2,858
  • 2
  • 16
  • 35
7
votes
3 answers

Multiple monitors in .NET

Are all displays returned from .NET's Screen.AllScreens regardless of hardware configuration? For example, on a single PC you can have: Video card out to two display = total 2 displays Video cards each out to 1 display = total 2 displays, Video…
Nicros
  • 5,031
  • 12
  • 57
  • 101
7
votes
1 answer

Tensorflow can't detect GPU when invoked by Ray worker

When I try the following code sample for using Tensorflow with Ray, Tensorflow fails to detect the GPU's on my machine when invoked by the "remote" worker but it does find the GPU's when invoked "locally". I put "remote" and "locally" in scare…
7
votes
3 answers

Can not save model using model.save following multi_gpu_model in Keras

Following the upgrade to Keras 2.0.9, I have been using the multi_gpu_model utility but I can't save my models or best weights using model.save('path') The error I get is TypeError: can’t pickle module objects I suspect there is some problem…
GhostRider
  • 2,109
  • 7
  • 35
  • 53
7
votes
1 answer

Multi GPU/Tower setup Tensorflow 1.2 Estimator

I want to turn my _model_fn for Estimator into a multi GPU solution. Is there a way to do it within the Esitmator API or do I have to explicitly code device placement and synchronization. I know I can use tf.device('gpu:X') to place my model on GPU…
dparted
  • 395
  • 1
  • 2
  • 15
7
votes
0 answers

nnGraph multi-GPU Torch

This question is about making any nnGraph network run on multiple GPUs and not specific to the following network instance I am trying to train a network which is constructed with nnGraph. The backward diagram is attached. I am trying to run the…
Bharat
  • 2,139
  • 2
  • 16
  • 35
7
votes
1 answer

CUDA SDK examples throw various errors in multi-gpu system

I have a Dell Precision Rack running Ubuntu Precise and featuring two Tesla C2075 plus a Quadro 600 which is the display device. I have recently finished some tests on my desktop-computer and now tried to port stuff to the workstation. Since CUDA…
pdresselhaus
  • 679
  • 14
  • 32
6
votes
2 answers

How to use keras.utils.Sequence data generator with tf.distribute.MirroredStrategy for multi-gpu model training in tensorflow?

I want to train a model on several GPUs using tensorflow 2.0. In the tensorflow tutorial for distributed training (https://www.tensorflow.org/guide/distributed_training), the tf.data datagenerator is converted into a distributed dataset as follows:…
surya1495
  • 65
  • 1
  • 4
6
votes
0 answers

Allocate two separate GPU's to different threads in python?

I want to use two python source codes, the first one is developed with tensorflow and the other is developed using pythorch. I want to run each of these codes in a thread with a separate GPU. the input of both codes are the same and there are some…
R.y.z
  • 213
  • 2
  • 15
6
votes
2 answers

ffmpeg - cuda encode - OpenEncodeSessionEx failed: out of memory

I'm having a problem with ffmpeg video encoding using GPU (CUDA). I have 2x nVidia GTX 1050 Ti The problem comes when i try to do multiple parallel encodings. More than 2 processes and ffmpeg dies like this: [h264_nvenc @ 0xcc1cc0]…
VelDev
  • 160
  • 2
  • 7
6
votes
2 answers

Poor performance when calling cudaMalloc with 2 GPUs simultaneously

I have an application where I split the processing load among the GPUs on a user's system. Basically, there is CPU thread per GPU that initiates a GPU processing interval when triggered periodically by the main application thread. Consider the…
rmccabe3701
  • 1,418
  • 13
  • 31
6
votes
1 answer

How can I override the CUDA kernel execution time limit on Windows with a secondary GPUs?

From Nvidia's website, it explain the time-out problem: Q: What is the maximum kernel execution time? On Windows, individual GPU program launches have a maximum run time of around 5 seconds. Exceeding this time limit usually will cause a…
user2003564
  • 213
  • 2
  • 9
1
2
3
25 26