您现在的位置是:首页 >技术杂谈 >nvidia-smi 失效解决网站首页技术杂谈

nvidia-smi 失效解决

only_liji 2024-06-17 11:25:18
简介nvidia-smi 失效解决

服务器重启后,跑模型发现:

RuntimeError: Attempting to deserialize object on a CUDA device but torch.cuda.is_available() is False. If you are running on a CPU-only machine, please use torch.load with map_location=torch.device('cpu') to map your storages to the CPU.

然后使用 nvidia-smi来查看:

>>nvidia-smi 
NVIDIA-SMI has failed because it couldn't communicate with the NVIDIA driver. Make sure that the latest NVIDIA driver is installed and running.
 

很明显,驱动掉了。

然后查看cuda:

>> nvcc -V
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2022 NVIDIA Corporation
Built on Tue_May__3_18:49:52_PDT_2022
Cuda compilation tools, release 11.7, V11.7.64
Build cuda_11.7.r11.7/compiler.31294372_0

再看驱动:

>> ls /usr/src | grep nvidia
nvidia-525.89.02

还好,都在。应该是ubuntu 重启时自动更新了。那就好解决了~

>> sudo apt-get install dkms

>>sudo dkms install -m nvidia -v 525.89.02

再运行 nvidia-smi ,然后熟悉的界面出现了

 

风语者!平时喜欢研究各种技术,目前在从事后端开发工作,热爱生活、热爱工作。