nvidia Questions

3

I read in slurm docs that we could use (after setting up the accounting) sacct --format="JobID,AllocCPUS,**ReqGRES** to get the statistics of requests for GRES. I have also configured my GPUs (the...
Tomblin asked 6/6, 2016 at 14:49

2

Recently, I have come across a solution of the triton serving config file disable flag "--strict-model-config=false" while running the inferencing server. This would enable to create its ...

2

Solved

I want to put in my system info about machine available GPU and status for real time monitoring. I'm outputting as XML the GPU info using $ nvidia-smi -x -q This command will print output curren...
Lounging asked 13/3, 2017 at 17:0

3

I saw several Q&As on this topic and tried both approaches. Any advice on how to proceed with either route are appreciated: Running nvidia-docker from within WSL2 I followed NVIDIA docs and thi...

1

Solved

According to https://pytorch.org/blog/accelerating-training-on-nvidia-gpus-with-pytorch-automatic-mixed-precision/ We can use: with torch.cuda.amp.autocast(): loss = model(data) In order to cast...
Roley asked 16/11, 2021 at 19:1

6

Solved

I have used atomicMax() to find the maximum value in the CUDA kernel: __global__ void global_max(float* values, float* gl_max) { int i=threadIdx.x + blockDim.x * blockIdx.x; float val=values[i];...
Preoccupied asked 1/7, 2013 at 7:11

3

Solved

I use the cuFFT library calls in a host code they work fine, but I want to call the cuFFT library from a kernel. Earlier versions of the CUDA didn't have this kind of support but with the dynamic p...
Titanic asked 24/6, 2013 at 7:49

1

Solved

I'm tending to collect my GPU status during my python code is running. I need to import nvidia_smi in my code to do this. but even by installing it by pip install nvidia_smi hit this error: No modu...
Templet asked 14/5, 2022 at 7:4

2

If I run the following: import torch import sys print('A', sys.version) print('B', torch.__version__) print('C', torch.cuda.is_available()) print('D', torch.backends.cudnn.enabled) device = torch.d...
Geminian asked 15/1, 2021 at 16:21

2

I'm trying to install Nvidia drivers on Ubuntu 13.10 by running NVIDIA-Linux-x86_64-331.49.run. I created a blacklist file under /etc/modprobe.d/blacklist-file-drivers.conf and added nouveau...
Bijugate asked 14/7, 2014 at 10:56

0

Problem description When I run vulkaninfo in docker, it complains: Cannot create Vulkan instance. This problem is often caused by a faulty installation of the Vulkan driver or attempting to use a G...
Court asked 17/4, 2022 at 3:33

4

Solved

I just managed to install my cuda SDK under Linux Ubuntu 10.04. My graphic card is an NVIDIA geForce GT 425M, and I'd like to use it for some heavy computational problem. What I wonder is: is there...
Milford asked 28/5, 2011 at 14:10

3

I want to create some neural network in tensorflow 2.x that trains on a GPU and I want to set up all the necessary infrastructure inside a docker-compose network (assuming that this is actually pos...
Baggy asked 26/2, 2020 at 16:16

1

Solved

I am using Nvidia's HPC compiler nvc++. Is there a way to detect that the program is being compile with this specific compiler and the version? I couldn't find anything in the manual https://docs.n...
Unaccomplished asked 16/3, 2022 at 0:54

1

When I use nvidia-smi, I found nearly 20GB GPU Memory is missing somewhere (total listed processes took 17745MB, meanwhile Memory-Usage is 37739MB): Then I use nvitop, you can see No Such Process ...
Vitrescence asked 11/3, 2022 at 3:33

1

I would like to train keras models using multiple GPUs. My understanding is that you cannot currently train multiple gpus using XLA. The issue is I can't figure out how to turn off XLA. Every GPU i...
Ponderous asked 9/8, 2019 at 23:39

2

Here is a very simple test program. When vsync is disabled this program runs at 100FPS and uses up virtually 0% of the CPU. When I enable vsync, I get 60FPS and 25% (100% of one core on a 4 core sy...
Microanalysis asked 21/2, 2014 at 4:30

0

My server do not response to nvidia-smi after I use ctrl+c kill the process running my GPU-training code. Before today, when I tap ctrl+c, the process first shows keyboard interrupt and then kille...
Shrine asked 12/2, 2022 at 3:59

4

Solved

Has anyone successfully installed Tensorflow-GPU on WSL2 with NVIDIA GPUs? I have Ubuntu 18.04 on WSL2, but am struggling to get NVIDIA drivers installed. Any help would be appreciated as I'm lost....
Bedcover asked 1/9, 2020 at 0:41

1

NVIDIA GeForce RTX 3070 with CUDA capability sm_86 is not compatible with the current PyTorch installation. The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_70. So I'm ...
Butterfield asked 26/7, 2021 at 11:44

1

Solved

On a Linux server, nvidia-smi can show GPU name, as well as other useful information. However, the full name is not displayed properly. How can I know the full name of GPUs?
Tachometer asked 9/1, 2022 at 10:56

3

I am running a model written with TensorFlow 1.x on 4x RTX 3090 and it is taking a long time to start up the training than as in 1x RTX 3090. Although, as training starts, it gets finished up earli...
Webbed asked 21/10, 2020 at 11:13

1

Solved

I have an electron repo (https://github.com/MartinBarker/RenderTune) which used to work on windows 10 fine when ran with command prompt. After a couple months I come back on a new fresh windows 10 ...
Drumfire asked 7/12, 2021 at 23:1

4

I have been working on getting an application that relies on TensorFlow to work as a docker container with nvidia-docker. I have compiled my application on top of the tensorflow/tensorflow:latest-g...
Rollo asked 16/5, 2017 at 3:56

5

Solved

I'm new to the CUDA paradigm. My question is in determining the number of threads per block, and blocks per grid. Does a bit of art and trial play into this? What I've found is that many examples h...
Maren asked 8/12, 2010 at 18:58

© 2022 - 2024 — McMap. All rights reserved.