Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

nvtop --No GPU to monitor. #318

Open
lransona opened this issue Oct 4, 2024 · 11 comments
Open

nvtop --No GPU to monitor. #318

lransona opened this issue Oct 4, 2024 · 11 comments

Comments

@lransona
Copy link

lransona commented Oct 4, 2024

Hi after reviewing most of the suggestions from MS and NV, and surfing the web, I can not get nvtop to see my GPU. Ollama saw the GPU and used but nvtop failed to see it?

Here is some information about the setup of my machine:

wsl --version
Versión de WSL: 2.3.24.0
Versión de kernel: 5.15.153.1-2
Versión de WSLg: 1.0.65
Versión de MSRDC: 1.2.5620
Versión de Direct3D: 1.611.1-81528511
Versión DXCore: 10.0.26100.1-240331-1435.ge-release
Versión de Windows: 10.0.22631.4169

$ nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2021 NVIDIA Corporation
Built on Thu_Nov_18_09:45:30_PST_2021
Cuda compilation tools, release 11.5, V11.5.119
Build cuda_11.5.r11.5/compiler.30672275_0

$ nvidia-smi
Fri Oct 4 13:57:04 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.51.01 Driver Version: 565.90 CUDA Version: 12.7 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=====================================+========================+======================|
| 0 NVIDIA GeForce RTX 4090 On | 00000000:05:00.0 On | Off |
| 0% 49C P8 10W / 450W | 1536MiB / 24564MiB | 5% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+----------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=====================================================================================|
| 0 N/A N/A 26 G /Xwayland N/A |
+-----------------------------------------------------------------------------------------+

$ curl -fsSL https://ollama.com/install.sh | sh

Installing ollama to /usr/local
Downloading Linux amd64 bundle
######################################################################## 100.0%##O=# #
######################################################################## 100.0%
Adding ollama user to render group...
Adding ollama user to video group...
Adding current user to ollama group...
Creating ollama systemd service...
Enabling and starting ollama service...
Nvidia GPU detected.
The Ollama API is now available at 127.0.0.1:11434.
Install complete. Run "ollama" from the command line.
mywsl@Galileo:~$ ollama run llama3.2:3b
Good morning
Good morning! How can I assist you today?

$ nvtop -v -> nvtop version 1.2.2
$ nvtop -> No GPU to monitor.

Do you have any suggestions about how to fix the problem? Have I missed something?
Many thanks for considering my request.

@Xiaomin-HUANG
Copy link

Your Kernel might need to be upgraded. My system is Ubuntu, and this works for me :
$ sudo apt remove nvtop
$ sudo reboot to update to a new kernel
then reboot the system by power off so that the new kernel can be loaded
$ sudo apt install nvtop

@lransona
Copy link
Author

lransona commented Oct 16, 2024

Thanks for your answer.
I did what you suggested but still the same error.

~$ nvtop
No GPU to monitor.
~$ nvtop -v
nvtop version 3.0.0

@najibveto
Copy link

Thanks for your answer. I did what you suggested but still the same error.

~$ nvtop No GPU to monitor. ~$ nvtop -v nvtop version 3.0.0

i am curious if you have windows 10 or 11?

@lransona
Copy link
Author

lransona commented Nov 5, 2024

Windows 11 Pro.

@najibveto
Copy link

Windows 11 Pro.

i tried both, i got the same error when i used win10 but it worked fine when i used windows 11 pro.

@jeffgsf
Copy link

jeffgsf commented Jan 25, 2025

Same issue. Windows 11 Pro, 2 Nvidia GPUs. I am running on WSL2 Ubuntu 22.04.3 LTS. "watch nvidia-smi" shows GPUs correctly. But nvtop responds with "No GPU to monitor."
NVML appears to see and report my GPUs correctly. I've run through all of the reinstalling checks, but can't figure it out. Is it Ubuntu 22.04.3? WSL2?

~$ nvtop -v --> nvtop version 3.1.0

Possibly --> no WSL2 kernel support for /dev/nvidia*

@bramvera
Copy link

bramvera commented Feb 1, 2025

nvtop-git latest still no luck on Arch Linux WSL

[100%] Built target nvtop
❯ sudo make install
[100%] Built target nvtop
Install the project...
-- Install configuration: "Release"
-- Installing: /usr/local/bin/nvtop
-- Set non-toolchain portion of runtime path of "/usr/local/bin/nvtop" to "/usr/local/lib"
-- Installing: /usr/local/share/man/man1/nvtop.1
-- Installing: /usr/local/share/icons/nvtop.svg
-- Installing: /usr/local/share/applications/nvtop.desktop
-- Installing: /usr/local/share/metainfo/nvtop.metainfo.xml
❯ nvtop
No GPU to monitor.

@iddrew
Copy link

iddrew commented Feb 3, 2025

I'm having the same issue. my ollama docker container running from my Ubuntu 24.04.1 LTS WSL distro on a windows 11 PC is seeing the GPU just fine and utilizing it. nvidia-smi on ubuntu also sees the GPU. it's just nvtop (v3.0.2) which is showing "no GPU to monitor"

@rynoian
Copy link

rynoian commented Feb 7, 2025

Same issue with linux mint

@perseity
Copy link

perseity commented Feb 7, 2025

Same.
nvtop 3.0.2
Ubuntu 24.04.1 LTS WSL2 on Windows 11 Pro

+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.77.01              Driver Version: 566.36         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 2080 Ti     On  |   00000000:01:00.0  On |                  N/A |
|  0%   34C    P8             28W /  300W |    1317MiB /  11264MiB |     15%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|  No running processes found                                                             |
+-----------------------------------------------------------------------------------------+

@wallentx
Copy link

It looks like this just started happening on Arch Linux as well.
Running nvidia-open 570.86.16-3.1.

It worked on the 6th..
But I'm noticing that even nvidia-smi is giving me:

Failed to initialize NVML: Driver/library version mismatch
NVML library version: 570.86

And that worked on Jan 28.

My GPU is def working.

/opt/cuda/extras/demo_suite/bandwidthTest

[CUDA Bandwidth Test] - Starting...
Running on...

 Device 0: NVIDIA RTX A4000
 Quick Mode

 Host to Device Bandwidth, 1 Device(s)
 PINNED Memory Transfers
   Transfer Size (Bytes)	Bandwidth(MB/s)
   33554432			2904.0

 Device to Host Bandwidth, 1 Device(s)
 PINNED Memory Transfers
   Transfer Size (Bytes)	Bandwidth(MB/s)
   33554432			2729.6

 Device to Device Bandwidth, 1 Device(s)
 PINNED Memory Transfers
   Transfer Size (Bytes)	Bandwidth(MB/s)
   33554432			397060.8

Result = PASS

NOTE: The CUDA Samples are not meant for performance measurements. Results may vary when GPU Boost is enabled.

Maybe nvidia broke something

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

9 participants