This section provides solutions to common problems associated with the NVIDIA Linux armv7l Driver.
My X server fails to start, and my X log file contains the error: (EE) NVIDIA(0): The NVIDIA kernel module does not appear to (EE) NVIDIA(0): be receiving interrupts generated by the NVIDIA graphics (EE) NVIDIA(0): device PCI:x:x:x. Please see the COMMON PROBLEMS (EE) NVIDIA(0): section in the README for additional information. |
|||||||||||
This can be caused by a variety of problems, such as PCI IRQ routing errors, I/O APIC problems, conflicts with other devices sharing the IRQ (or their drivers), or MSI compatibility problems. If possible, configure your system such that your graphics card does not share its IRQ with other devices (try moving the graphics card to another slot if applicable, unload/disable the driver(s) for the device(s) sharing the card's IRQ, or remove/disable the device(s)). Depending on the nature of the problem, one of (or a combination of) these kernel parameters might also help:
The problem may also be caused by MSI compatibility problems. See MSI Interrupts for details. |
|||||||||||
My X server fails to start, and my X log file contains the error: (EE) NVIDIA(0): The interrupt for NVIDIA graphics device PCI:x:x:x (EE) NVIDIA(0): appears to be edge-triggered. Please see the COMMON (EE) NVIDIA(0): PROBLEMS section in the README for additional information. |
|||||||||||
An edge-triggered interrupt means that the kernel has programmed the interrupt as edge-triggered rather than level-triggered in the Advanced Programmable Interrupt Controller (APIC). Edge-triggered interrupts are not intended to be used for sharing an interrupt line between multiple devices; level-triggered interrupts are the intended trigger for such usage. When using edge-triggered interrupts, it is common for device drivers using that interrupt line to stop receiving interrupts. This would appear to the end user as those devices no longer working, and potentially as a full system hang. These problems tend to be more common when multiple devices are sharing that interrupt line. This occurs when ACPI is not used to program interrupt routing in the APIC. It may also occur when ACPI is disabled, or fails to initialize. In these cases, the Linux kernel falls back to tables provided by the system BIOS. In some cases the system BIOS assumes ACPI will be used for routing interrupts and configures these tables to incorrectly label all interrupts as edge-triggered. The current interrupt configuration can be found in /proc/interrupts. Available workarounds include: updating to a newer system BIOS, a more recent Linux kernel with ACPI enabled, or passing the 'noapic' option to the kernel to force interrupt routing through the traditional Programmable Interrupt Controller (PIC). The Linux kernel also provides an interrupt polling mechanism you can use to attempt to work around this problem. This mechanism can be enabled by passing the 'irqpoll' option to the kernel. Currently, the NVIDIA driver will attempt to detect edge triggered interrupts and X will purposely fail to start (to avoid stability issues). This behavior can be overridden by setting the "NVreg_RMEdgeIntrCheck" NVIDIA Linux kernel module parameter. This parameter defaults to "1", which enables the edge triggered interrupt detection. Set this parameter to "0" to disable this detection. |
|||||||||||
X starts for me, but OpenGL applications terminate immediately. |
|||||||||||
If X starts but you have trouble with OpenGL, you most likely
have a problem with other libraries in the way, or there are stale
symlinks. See Chapter 5,
Listing of Installed Components for details. Sometimes,
all it takes is to rerun You should also check that the correct extensions are present; % xdpyinfo should show the “GLX” and “NV-GLX” extensions present. If these two extensions are not present, then there is most likely a problem loading the glx module, or it is unable to implicitly load GLcore. Check your X config file and make sure that you are loading glx (see Chapter 6, Configuring X for the NVIDIA Driver). If your X config file is correct, then check the X log file for warnings/errors pertaining to GLX. Also check that all of the necessary symlinks are in place (refer to Chapter 5, Listing of Installed Components). |
|||||||||||
When Xinerama is enabled, my stereo glasses are shuttering only when the stereo application is displayed on one specific X screen. When the application is displayed on the other X screens, the stereo glasses stop shuttering. |
|||||||||||
This problem occurs with DDC and "blue line" stereo glasses, that get the stereo signal from one video port of the graphics card. When a X screen does not display any stereo drawable the stereo signal is disabled on the associated video port. Forcing stereo flipping allows the stereo glasses to shutter continuously. This can be done by enabling the OpenGL control "Force Stereo Flipping" in nvidia-settings, or by setting the X configuration option "ForceStereoFlipping" to "1". |
|||||||||||
Stereo is not in sync across multiple displays. |
|||||||||||
There are two cases where this may occur. If the displays are attached to the same GPU, and one of them is out of sync with the stereo glasses, you will need to reconfigure your monitors to drive identical mode timings; see Chapter 18, Programming Modes for details. If the displays are attached to different GPUs, the only way to synchronize stereo across the displays is with a Quadro Sync device, which is only supported by certain Quadro cards. See Chapter 29, Configuring Frame Lock and Genlock for details. |
|||||||||||
I just upgraded my kernel, and now the NVIDIA kernel module will not load. |
|||||||||||
The kernel interface layer of the NVIDIA kernel module must be compiled specifically for the configuration and version of your kernel. If you upgrade your kernel, then the simplest solution is to reinstall the driver. ADVANCED: You can install the NVIDIA kernel module for a non running kernel (for example: in the situation where you just built and installed a new kernel, but have not rebooted yet) with a command line such as this: # sh NVIDIA-Linux-armv7l-367.44.run --kernel-name='KERNEL_NAME' Where 'KERNEL_NAME' is what uname -r would report if the target kernel were running. |
|||||||||||
Installing the driver fails with: Unable to load the kernel module 'nvidia.ko'. My X server fails to start, and my X log file contains the error: (EE) NVIDIA(0): Failed to load the NVIDIA kernel module! |
|||||||||||
If the NVIDIA kernel module fails to load, you should check the
output of
|
|||||||||||
Installing the NVIDIA kernel module gives an error message like: #error Modules should never use kernel-headers system headers #error but headers from an appropriate kernel-source |
|||||||||||
You need to install the source for the Linux kernel. In most situations you can fix this problem by installing the kernel-source or kernel-devel package for your distribution |
|||||||||||
OpenGL applications crash and print out the following warning: WARNING: Your system is running with a buggy dynamic loader. This may cause crashes in certain applications. If you experience crashes you can try setting the environment variable __GL_SINGLE_THREADED to 1. For more information, consult the FREQUENTLY ASKED QUESTIONS section in the file /usr/share/doc/NVIDIA_GLX-1.0/README.txt. |
|||||||||||
The dynamic loader on your system has a bug which will cause
applications linked with pthreads, and that % export __GL_SINGLE_THREADED=1 and in csh and derivatives use: % setenv __GL_SINGLE_THREADED 1 Previous releases of the NVIDIA Accelerated Linux Graphics Driver attempted to work around this problem. Unfortunately, the workaround caused problems with other applications and was removed after version 1.0-1541. |
|||||||||||
Quake3 crashes when changing video modes. |
|||||||||||
You are probably experiencing a problem described above. Please
check the text output for the “WARNING” message described in the previous
hint. Setting |
|||||||||||
I cannot build the NVIDIA kernel module, or, I can build the NVIDIA kernel module, but modprobe/insmod fails to load the module into my kernel. |
|||||||||||
These problems are generally caused by the build using the wrong
kernel header files (i.e. header files for a different kernel
version than the one you are running). The convention used to be
that kernel header files should be stored in |
|||||||||||
Compiling the NVIDIA kernel module gives this error: You appear to be compiling the NVIDIA kernel module with a compiler different from the one that was used to compile the running kernel. This may be perfectly fine, but there are cases where this can lead to unexpected behavior and system crashes. If you know what you are doing and want to override this check, you can do so by setting IGNORE_CC_MISMATCH. In any other case, set the CC environment variable to the name of the compiler that was used to compile the kernel. |
|||||||||||
You should compile the NVIDIA kernel module with the same
compiler version that was used to compile your kernel. Some Linux
kernel data structures are dependent on the version of gcc used to
compile it; for example, in ... * Most gcc versions have a nasty bug with empty initializers. */ #if (__GNUC__ > 2) typedef struct { } rwlock_t; #define RW_LOCK_UNLOCKED (rwlock_t) { } #else typedef struct { int gcc_is_buggy; } rwlock_t; #define RW_LOCK_UNLOCKED (rwlock_t) { 0 } #endif If the kernel is compiled with gcc 2.x, but gcc 3.x is used when the kernel interface is compiled (or vice versa), the size of rwlock_t will vary, and things like ioremap will fail. To check what version of gcc was used to compile your kernel, you can examine the output of: % cat /proc/version To check what version of gcc is currently in your % gcc -v |
|||||||||||
X fails with error Failed to allocate LUT context DMA |
|||||||||||
This is one of the possible consequences of compiling the NVIDIA kernel interface with a different gcc version than used to compile the Linux kernel (see above). |
|||||||||||
I recently updated various libraries on my system using my Linux distributor's update utility, and the NVIDIA graphics driver no longer works. |
|||||||||||
Conflicting libraries may have been installed by your distribution's update utility; see Chapter 5, Listing of Installed Components for details on how to diagnose this. |
|||||||||||
I have rebuilt the NVIDIA kernel module, but when I try to insert it, I get a message telling me I have unresolved symbols. |
|||||||||||
Unresolved symbols are most often caused by a mismatch between your kernel sources and your running kernel. They must match for the NVIDIA kernel module to build correctly. Make sure your kernel sources are installed and configured to match your running kernel. |
|||||||||||
OpenGL applications leak significant amounts of memory on my system! |
|||||||||||
If your kernel is making use of the -rmap VM, the system may be leaking memory due to a memory management optimization introduced in -rmap14a. The -rmap VM has been adopted by several popular distributions, the memory leak is known to be present in some of the distribution kernels; it has been fixed in -rmap15e. If you suspect that your system is affected, try upgrading your kernel or contact your distribution's vendor for assistance. |
|||||||||||
Some OpenGL applications (like Quake3 Arena) crash when I start them on Red Hat Linux 9.0. |
|||||||||||
Some versions of the glibc package shipped by Red Hat that
support TLS do not properly handle using |
|||||||||||
When changing settings in games like Quake 3 Arena, or Wolfenstein Enemy Territory, the game crashes and I see this error: ...loading libGL.so.1: QGL_Init: dlopen libGL.so.1 failed: /usr/lib/tls/libGL.so.1: shared object cannot be dlopen()ed: static TLS memory too small |
|||||||||||
These games close and reopen the NVIDIA OpenGL driver (via
|
|||||||||||
When I try to install the driver, the installer claims that X is running, even though I have exited X. |
|||||||||||
The installer detects the presence of an X server by checking
for the X server's lock files: |
|||||||||||
Why does the VBIOS fail to load on my Optimus system? |
|||||||||||
On some notebooks with Optimus graphics, the NVIDIA driver may not be able to retrieve the Video BIOS due to interactions between the System BIOS and the Linux kernel's ACPI subsystem. On affected notebooks, applications that require the GPU will fail, and messages like the following may appear in the system log: NVRM: failed to copy vbios to system memory. NVRM: RmInitAdapter failed! (0x30:0xffffffff:858) NVRM: rm_init_adapter(0) failed Such problems are typically beyond the control of the NVIDIA driver, which relies on proper cooperation of ACPI and the System BIOS to retrieve important information about the GPU, including the Video BIOS. |
|||||||||||
OpenGL applications do not work with driver version 364.xx and later, which worked with previous driver versions |
|||||||||||
Release 361 of the NVIDIA Linux driver introduced OpenGL libraries built upon the libglvnd (GL Vendor Neutral Dispatch) architecture, to allow for the coexistence of multiple OpenGL implementations on the same system. The .run installer package includes both GLVND and non-GLVND GLX client libraries, and beginning with release 364, the GLVND libraries are installed by default. By design, GLVND conforms with the Linux OpenGL ABI version 1.0 as defined at https://www.opengl.org/registry/ABI/ and exposes all required entry points; however, applications which depend upon specifics of the NVIDIA OpenGL implementation which fall outside of the OpenGL ABI may be incompatible with a GLVND-based OpenGL implementation. If you encounter an application which is incompatible with
GLVND, you may install a legacy, non-GLVND GLX client library by
adding the |
|||||||||||
OpenGL applications are running slowly |
|||||||||||
The application is probably using a different library that still remains on your system, rather than the NVIDIA supplied OpenGL library. See Chapter 5, Listing of Installed Components for details. |
|||||||||||
X takes a long time to start (possibly several minutes). |
|||||||||||
Most of the X startup delay problems we have found are caused by incorrect data in video BIOSes about what display devices are possibly connected or what i2c port should be used for detection. You can work around these problems with the X config option IgnoreDisplayDevices. |
|||||||||||
Fonts are incorrectly sized after installing the NVIDIA driver. |
|||||||||||
Incorrectly sized fonts are generally caused by incorrect DPI (Dots Per Inch) information. You can check what X thinks the physical size of your monitor is, by running: % xdpyinfo | grep dimensions This will report the size in pixels, and in millimeters. If these numbers are wrong, you can correct them by modifying the X server's DPI setting. See Appendix E, Dots Per Inch for details. |
|||||||||||
OpenGL applications don't work, and my X log file contains the error: (EE) NVIDIA(0): Unable to map device node /dev/zero with read and write (EE) NVIDIA(0): privileges. The GLX extension will be disabled on this (EE) NVIDIA(0): X screen. Please see the COMMON PROBLEMS section in the (EE) NVIDIA(0): README for more information. |
|||||||||||
The NVIDIA OpenGL driver must be able to map anonymous memory with read and write execute privileges in order to function correctly. The driver needs this ability to allocate aligned memory, which is used for certain optimizations. Currently, GLX cannot run without these optimizations. |
|||||||||||
X doesn't start, and my log file contains a message like the following: (EE) NVIDIA(0): Failed to allocate primary buffer: failed to set CPU access (EE) NVIDIA(0): for surface. Please see Chapter 8: Common Problems in (EE) NVIDIA(0): the README for troubleshooting suggestions. |
|||||||||||
The NVIDIA X driver needs to be able to access the buffers it allocates from the CPU, but wasn't able to set up this access. This commonly fails if you're using a large virtual desktop size. Although your GPU may have enough onboard video memory for the buffer, the amount of usable memory may be limited if the IndirectMemoryAccess option is disabled, or if not enough address space was reserved for indirect memory access (this commonly occurs on 32-bit systems). If you're seeing this problem and are using a 32-bit operating system, it may be resolved by switching to a 64-bit operating system. |
|||||||||||
My log file contains a message like the following: (WW) NVIDIA(GPU-0): Unable to enter interactive mode, because non-interactive (WW) NVIDIA(GPU-0): mode has been previously requested. The most common (WW) NVIDIA(GPU-0): cause is that a GPU compute application is currently (WW) NVIDIA(GPU-0): running. Please see the README for details. |
|||||||||||
This indicates that the X driver was not able to put the GPU in interactive mode, because another program has requested non-interactive mode. The GPU watchdog will not run, and long-running GPU compute programs may cause the X server and OpenGL programs to hang. If you intend to run long-running GPU compute programs, set the Interactive option to "off" to disable interactive mode. |
|||||||||||
I see a blank screen or an error message instead of a login screen or desktop session |
|||||||||||
Installation or configuration problems may prevent the X server, a login/session manager, or a desktop environment from starting correctly. If your system is failing to display a login screen, or failing to start a desktop session, try the following troubleshooting steps:
|
|||||||||||
The display settings I configured in nvidia-settings do not persist. |
|||||||||||
Depending on the type of configuration being performed, nvidia-settings will save configuration changes to one of several places:
Settings in Even after nvidia-settings has been run to restore
any settings set in These tools may also override some types of settings that are stored in and loaded from the X configuration file, such as any MetaMode strings that may specify the initial display layouts of NVIDIA X screens. Although the configuration of the initial MetaMode is static, it is possible to dynamically switch to a different MetaMode after X has started. This can have the effect of making the set of active displays, their resolutions, and layout positions as configured in the nvidia-settings control panel appear to be ineffective, when in reality, this configuration was active when starting X and then overridden later by the desktop environment. If you believe that your desktop environment is overriding settings that you configured in nvidia-settings, some possible solutions are:
Some systems may have multiple different display configuration utilities, each with its own way of managing settings. In addition to conflicting with nvidia-settings, such tools may conflict with each other. If your system uses more than one tool for configuring displays, make sure to check the configuration of each tool when attempting to determine the source of any unexpected display settings. |
|||||||||||
My displays are reconfigured in unexpected ways when I plug in or unplug a display, or power a display off and then power it on again. |
|||||||||||
This is a special case of the issues described in “The display settings I configured in nvidia-settings do not persist.”. Some desktop environments which include advanced display configuration tools will automatically configure the display layout in response to detected configuration changes. For example, when a new display is plugged in, such a desktop environment may attempt to restore the previous layout that was used with the set of currently connected displays, or may configure a default layout based upon its own policy. On X servers with support for RandR 1.2 or later, the NVIDIA X driver reports display hotplug events to the X server via RandR when displays are connected and disconnected. These hotplug events may trigger a desktop environment with advanced display management capabilities to change the display configuration. These changes may affect settings such as the set of active displays, their resolutions and positioning relative to each other, per-display color correction settings, and more. In addition to hotplug events generated by connecting or disconnecting displays, DisplayPort displays will generate a hot unplug event when they power off, and a hotplug event when they power on, even if no physical plugging in or unplugging takes place. This can lead to hotplug-induced display configuration changes without any actual hotplug action taking place. If display hotplug events are resulting in undesired configuration changes, try the solutions and workarounds listed in “The display settings I configured in nvidia-settings do not persist.”. Another workaround would be to disable the NVIDIA X driver's reporting of hotplug events with the UseHotplugEvents X configuration option. Note that this option will have no effect on DisplayPort devices, which must report all hotplug events to ensure proper functionality. |