Since the last nvidia drivers update to 570 I’m getting system freezes. The screen blanks and graphic card fans spin on full blast at 100%. System seems to still be running in the background as sound continues to play, but I have no way of accessing it and have to press&hold the power button for a forced shutdown.
Log shows while watching Youtube in Firefox:
kernel: nvidia-modeset: ERROR: GPU:0: Failed to query display engine channel state: 0x0000c57e:0:0:0x0000000f
kernel: NVRM: GPU 0000:26:00.0: GPU has fallen off the bus.
kernel: NVRM: Xid (PCI:0000:26:00): 79, pid=940, name=kwin_x11, GPU has fallen off the bus.
kernel: NVRM: GPU at PCI:0000:26:00: GPU-e2d0765f-5262-15ae-1cb0-50a8a116d2e7
Tried reproducing error by playing a game on Steam and after less than 10 minutes the system froze with error:
kernel: ucsi_ccg 0-0008: i2c_transfer failed -110
kernel: nvidia-gpu 0000:26:00.3: i2c timeout error ffffffff
nvidia-settings[3129]: ERROR: The GPU has fallen off the bus or has otherwise become inaccessible
kernel: xhci_hcd 0000:26:00.2: HC died; cleaning up
kernel: xhci_hcd 0000:26:00.2: PCI post-resume error -19!
kernel: xhci_hcd 0000:26:00.2: Controller not ready at resume -19
kernel: xhci_hcd 0000:26:00.2: Unable to change power state from D3cold to D0, device inaccessible
kernel: xhci_hcd 0000:26:00.2: Unable to change power state from D3hot to D0, device inaccessible
kernel: nvidia-modeset: ERROR: GPU:0: Failed to query display engine channel state: 0x0000c57e:2:0:0x0000000f
kernel: nvidia-modeset: ERROR: GPU:0: Failed to query display engine channel state: 0x0000c57e:0:0:0x0000000f
kernel: NVRM: Xid (PCI:0000:26:00): 154, GPU recovery action changed from 0x0 (None) to 0x1 (GPU Reset Required)
kernel: NVRM: A GPU crash dump has been created. If possible, please run
NVRM: nvidia-bug-report.sh as root to collect this data before
NVRM: the NVIDIA kernel module is unloaded.
kernel: NVRM: GPU 0000:26:00.0: GPU has fallen off the bus.
kernel: NVRM: Xid (PCI:0000:26:00): 79, GPU has fallen off the bus.
kernel: NVRM: GPU at PCI:0000:26:00: GPU-e2d0765f-5262-15ae-1cb0-50a8a116d2e7
kernel: nvidia-gpu 0000:26:00.3: Unable to change power state from D3hot to D0, device inaccessible
When I open Nvidia Settings then journalctl is permanently flooded with errors multiple times a second:
nvidia-settings[3129]: ERROR: An internal driver error occurred
While using Nvidia Settings under Thermal Settings I can’t enable Enable GPU Fan Settings as error Failed to Enable GPU Fan Settings! shows. And Fan Information shows:
ID Speed (RPM) Target Speed (%) Control Type Cooling Target
0 Unsupported 0 Variable GPU, Memory , and Power Supply
1 Unsupported 0 Variable GPU, Memory , and Power Supply
Also I can’t do a reinstall of the nvidia drivers
mhwd -f -i pci video-nvidia:
> Removing video-nvidia...
Sourcing /etc/mhwd-x86_64.conf
Has lib32 support: true
Sourcing /var/lib/mhwd/local/pci/video-nvidia/MHWDCONFIG
Processing classid: 0300
Sourcing /var/lib/mhwd/scripts/include/0300
Processing classid: 0302
checking dependencies...
error: failed to prepare transaction (could not satisfy dependencies)
:: removing nvidia-utils breaks dependency 'nvidia-utils=570.133.07' required by linux612-nvidia
Error: pacman failed!
Error: script failed!
pacman -Q | grep nvidia:
lib32-nvidia-utils 570.133.07-1
linux612-nvidia 570.133.07-3
mhwd-nvidia 570.133.07-2
mhwd-nvidia-390xx 390.157-19
mhwd-nvidia-470xx 470.256.02-10
nvidia-driver-assistant 0.20.124.06-1
nvidia-settings 570.133.07-2.2
nvidia-utils 570.133.07-2
opencl-nvidia 570.133.07-2
nvidia-smi:
Sun Mar 30 15:08:35 2025
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 570.133.07 Driver Version: 570.133.07 CUDA Version: 12.8 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 2080 Off | 00000000:26:00.0 On | N/A |
|ERR! 46C P5 N/A / N/A | 1347MiB / 8192MiB | 2% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+