Questions tagged [gpu]

Acronym for "Graphics Processing Unit". For programming traditional graphical applications, see the tag entry for "graphics programming". For general-purpose programming using GPUs, see the tag entry for "gpgpu". For specific GPU programming technologies, see the popular tag entries for "opencl", "cuda" and "thrust".

Acronym for "Graphics Processing Unit". For programming traditional graphical applications, see the tag entry for . For general-purpose programming using GPUs, see the tag entry for . For specific GPU programming technologies, see the popular tag entries for , and .

More information on GPU at http://en.wikipedia.org/wiki/Graphics_processing_unit

8854 questions
3
votes
1 answer

Loading ResNet50 on RTX2070 - Out of Memory

I'm trying to load ResNext50, and on top of it CenterNet, I'm able to do it with Google Colab or Kaggle's GPU. But, Would love to know how much GPU Memory (VRAM) does this network need? When using RTX 2070 with free 5.5GB VRAM left on it (out of…
Ilan Aizelman
  • 49
  • 1
  • 7
3
votes
3 answers

What's the speed of texture upload?

I would like to upload two images to the GPU memory, and I'm interested how fast I can do this? In fact - will it be faster to compare two bitmaps in RAM with CPU, or upload them to GPU and use GPU parallelism to do it?
Daniel Mošmondor
  • 19,718
  • 12
  • 58
  • 99
3
votes
1 answer

How is the size of a surface determined in Vulkan?

I'm following the Vulkan Tutorial and the section Window Surface says that on Windows a VkSurfaceKHR object is created using the following code: VkWin32SurfaceCreateInfoKHR createInfo = {}; createInfo.sType =…
David DiGioia
  • 181
  • 15
3
votes
1 answer

Is there anyway to run GPT2 without GPU and TensorFlow

GPT2 is an excellent OpenAI project for NLP. The developer requirement stated we need to use tensor Flow and GPU. I only want to use (not to train) the existing trained parameters. Is there any way to use GPT2 without the expensive hardware with GPU…
Kardi Teknomo
  • 1,375
  • 16
  • 24
3
votes
1 answer

Pytorch on Google VM (Linux) does not recognize GPU

I created a Google VM instance using this available image: c1-deeplearning-common-cu100-20191226 Description Google, Deep Learning Image: Base, m39 (with CUDA 10.0), A Debian based image with CUDA 10.0 I then installed Anaconda onto this VM, then…
Bill
  • 653
  • 2
  • 7
  • 20
3
votes
2 answers

GPU with rootless Docker

I do not want my users to have the opportunity to run their containers in privileged mode. Recently, in Docker 19.03 they introduced a rootless experimental mode. It works great, however I do not know how to combine it with a nvidia-docker…
piko
  • 132
  • 1
  • 9
3
votes
2 answers

OOm - cannot run StyleGAN2 despite reducing batch size

I am trying to run StyleGAN2 using a cluster equipped with eight GPUs (NVIDIA GeForce RTX 2080). At present, I am using the following configuration in training_loop.py: minibatch_size_dict = {4: 512, 8: 256, 16: 128, 32: 64, 64: 32}, #…
albus_c
  • 6,292
  • 14
  • 36
  • 77
3
votes
1 answer

How to solve ""RuntimeError: CUDA out of memory."? Is there a way to free more memory?

In this case, I'm using jupyter notebook on a VM for trainning some CNN models. the VM has 16v CPU with 60GB memory. And I just attched a NVIDIA TESLA P4 for better performance. But it always gives error like "RuntimeError: CUDA out of memory.…
Amyano
  • 41
  • 1
  • 4
3
votes
4 answers

How to make a PyTorch Distribution on GPU

Is it possible to make the PyTorch distributions create their samples directly on GPU. If I do from torch.distributions import Uniform, Normal normal = Normal(3, 1) sample = normal.sample() Then sample will be on CPU. Of course it is possible to…
patapouf_ai
  • 17,605
  • 13
  • 92
  • 132
3
votes
1 answer

Compile CUDA code in Qt5.13 using msvc 15

I have to build a cuda code along with other c++ files in Qt . But that is getting failed. I have tried with the samples available in Stackoverflow, but that also got failed. I have attached the sample code . Please provide an advise .pro file QT …
Masthan
  • 727
  • 1
  • 8
  • 29
3
votes
1 answer

How do I create a CuArray populated with strings in Julia?

I have a lot of String operations to perform and I'd like to implement code in Julia 1.2.0 using GPU processing. Right off the bat I ran into issues defining a CuArray, an example is below. I essentially want to be able to hand a 1D Array of strings…
3
votes
0 answers

Convert PyCuda code to PyOpenCL code : How to perform a high number of 3x3 matrix inversion?

I try to convert a PyCuda code to PyOpenCL. This is the following of a previous version of working NVIDIA GPU code. This code aims to invert a high number of 3x3 matrixes. Here's the PyCuda working version : $ cat t14.py import numpy as np # import…
user1773603
3
votes
1 answer

What is the meaning of CUDA return value error 35?

My source code of simple C++ cuda code #include #include using namespace std; __global__ void AddIntsCUDA(int *a, int *b, int *c) { *c = *a + *b; } int main() { int a, b, c; int *d_a, *d_b, *d_c; int size =…
r00tk1ll3r
  • 63
  • 1
  • 10
3
votes
1 answer

Can multiple processes share one CUDA context?

This question is a followup on Jason R's comment to Robert Crovellas answer on this original question ("Multiple CUDA contexts for one device - any sense?"): When you say that multiple contexts cannot run concurrently, is this limited to kernel…
alex
  • 10,900
  • 15
  • 70
  • 100
3
votes
1 answer

Techniques for optimising GPU utilisation processing discrete images

I have a server which is applying filters (implemented as OpenGL shaders) to images. They are mostly direct colour mappings but also occasionally blurs and other convolutions. The source images are PNGs and JPGs in a variety of sizes from e.g.…
Dave Durbin
  • 3,562
  • 23
  • 33
1 2 3
99
100