VirtualGL

VirtualGL redirects an application's OpenGL/GLX commands to a separate X server (that has access to a 3D graphics card), captures the rendered images, and then streams them to the X server that actually handles the application.

The main use-case is to enable server-side hardware-accelerated 3D rendering for remote desktop set-ups where the X server that handles the application is either on the other side of the network (in the case of X11 forwarding), or a "virtual" X server that cannot access the graphics hardware (in the case of VNC).

Installation and setup

Install the virtualgl package using pacman, then follow Configuring a Linux or Unix Machine as a VirtualGL Server to configure it. On arch, /opt/VirtualGL/bin/vglserver_config is just vglserver_config and /opt/VirtualGL/bin/glxinfo is vglxinfo.

Using VirtualGL with X11 forwarding

  server:                                              client:
 ······································               ·················
 : ┌───────────┐ X11 commands         :               : ┌───────────┐ :
 : │application│━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━▶│X server 2)│ :
 : │           │        ┌───────────┐ :               : │           │ :
 : │           │        │X server 1)│ :               : ├┈┈┈┈┈┈┈┈┈╮ │ :
 : │ ╭┈┈┈┈┈┈┈┈┈┤ OpenGL │ ╭┈┈┈┈┈┈┈┈┈┤ : image stream  : │VirtualGL┊ │ :
 : │ ┊VirtualGL│━━━━━━━▶│ ┊VirtualGL│━━━━━━━━━━━━━━━━━━▶│client   ┊ │ :
 : └─┴─────────┘        └─┴─────────┘ :               : └─────────┴─┘ :
 ······································               ·················
  1. "3D" rendering happens here
  2. "2D" rendering happens here

Advantages of this set-up, compared to using VirtualGL with VNC:

  • seamless windows
  • uses a little less CPU resources on the server side
  • supports stereo rendering (for viewing with "3D glasses")

Instructions

1. Preparation

In addition to setting up VirtualGL on the remote server as described above, this usage scenario requires you to:

  • install the virtualgl package on the client side as well (but no need to set it up like on the server side, we just need the vglconnect and vglclient binaries on this end).
  • set up SSH with X11 forwarding (confirm that connecting from the client to the server via ssh -X user@server and running GUI applications in the resulting shell works)

2. Connecting

Now you can use vglconnect on the client computer whenever you want to connect to the server:

$ vglconnect user@server     # X11 traffic encrypted, VGL image stream unencrypted
$ vglconnect -s user@server  # both X11 traffic and VGL image stream encrypted

This opens an SSH session with X11 forwarding just like would, and also automatically starts the VirtualGL Client (vglclient) with the right parameters as a background daemon. This daemon will handle incoming VirtualGL image streams from the server, and will keep running in the background even after you close the SSH shell - you can stop it with .

3. Running applications

Once connected, you can run remote applications with VirtualGL rendering enabled for their OpenGL parts, by starting them inside the SSH shell with as described in Running Applications below.

You do not need to restrict yourself to the shell that vglconnect opened for you; any or shell you open from the same X session on the client to the same user@server should work. will detect that you are in an SSH shell, and make sure that the VGL image stream is sent over the network to the IP/hostname belonging to the SSH client (where the running vglclient instance will intercept and process it).

Using VirtualGL with VNC

  server:                                                              client:
 ······················································               ················
 : ┌───────────┐ X11 commands         ┌─────────────┐ : image stream  : ┌──────────┐ :
 : │application│━━━━━━━━━━━━━━━━━━━━━▶│VNC server 2)│━━━━━━━━━━━━━━━━━━▶│VNC viewer│ :
 : │           │        ┌───────────┐ └─────────────┘ :               : └──────────┘ :
 : │           │        │X server 1)│        ▲        :               :              :
 : │ ╭┈┈┈┈┈┈┈┈┈┤ OpenGL │ ╭┈┈┈┈┈┈┈┈┈┤ images ┃        :               :              :
 : │ ┊VirtualGL│━━━━━━━▶│ ┊VirtualGL│━━━━━━━━┛        :               :              :
 : └─┴─────────┘        └─┴─────────┘                 :               :              :
 ······················································               ················
  1. "3D" rendering happens here
  2. "2D" rendering happens here

Advantages of this set-up, compared to using VirtualGL with X11 Forwarding:

  • can maintain better performance in case of low-bandwidth/high-latency networks
  • can send the same image stream to multiple clients ("desktop sharing")
  • the remote application can continue running even when the network connection drops
  • better support for non-Linux clients, as the architecture does not depend on a client-side X server

Instructions

After setting up VirtualGL on the remote server as described above, and establishing a working remote desktop connection using the VNC client/server implementation of your choice, no further configuration should be needed.

Inside the VNC session (e.g. in a terminal emulator within the VNC desktop or even directly in ~/.vnc/xstartup), simply run selected applications with as described in Running Applications below.

You can also run your entire session with , so that all opengl applications work by default. For example, if you use xfce, you can run instead of in your X startup scripts (~/.vnc/xstartup, or equivalent), or copy and edit a .desktop file in /usr/share/xsessions if you are using a display manager.

Choosing an appropriate VNC package

VirtualGL can provide 3D rendering for any general-purpose vncserver implementation (e.g. TightVNC, RealVNC, ...).

However, if you want to really get good performance out of it (e.g. to make it viable to watch videos or play OpenGL games over VNC), you might want to use one of the VNC implementations that are specifically optimized for this use-case:

  • : Developed by the same team as VirtualGL, with the explicit goal of providing the best performance in combination with it. However, its vncserver implementation does not support all features a normal Xorg server provides, thus some applications will run unusually slow or not at all in it.
  • TigerVNC: Also developed with VirtualGL in mind and achieves good performance with it, while providing better Xorg compatibility than TurboVNC.

Using VirtualGL with Xpra

On your host

  • Prepare a minimal xinit resource file that just locks your Xorg server's display:
  • Start an Xorg server using the prepared resource file startx ~/.xinitrc-vgl. Now, should return two Xorg instances.
  • Run an appliction via vglrun command specifying the Xorg display used by your Xpra, e.g. . The apllication will not be visible yet.

On your client

  • Setup Xpra at the client and attach to it. Now you should see the glxspheres64 application started above.
Note: Use a ssh connection when attaching via network. X-forwarding not necessary. Also, set up a ssh-keypair on client and copy it to the remote host.

Running applications

Once you have set up your remote desktop connection with VirtualGL support, you can use to run selected applications with VirtualGL-accelerated rendering of their OpenGL parts:

$ vglrun glxgears

This has to be executed on the remote computer of course (where the application will run), i.e. inside your SSH or VNC session. The X servers that will be used, are determined from the following two environment variables:

The X server that will handle the application, and render its non-OpenGL parts.

If using VNC, this refers to the VNC server. In the case of SSH forwarding, it is a virtual X server number on the remote computer that SSH internally maps to the real X server on the client. There is nothing VirtualGL-specific about this variable, and it will already be set to the correct value within your SSH or VNC session.

The X server to which VirtualGL should redirect OpenGL rendering.

See Installation and setup above. If not set, the value is assumed. Note that the number after the dot can be used to select the graphics card.

Many more environment variables and command-line parameters are available to fine-tune - refer to the user manual and for reference. VirtualGL's behavior furthermore depends on which of its two main modes of operation is active (which will choose automatically, based on the environment in which it is executed):

In this mode, a compressed image stream of the rendered OpenGL scenes is sent through a custom network protocol to a vglclient instance. By default it uses JPEG compression at 90% quality, but this can be fully customized, e.g.:
There is also a GUI dialog that lets you change the most common VirtualGL rendering/compression options for an application on the fly, after you have already started it with - simply press while the application has keyboard focus, to open this dialog.
In this mode, VirtualGL feeds raw (uncompressed) images through the normal X11 protocol directly to the X server that handles the application - e.g. a VNC server running on the same machine. Many of 's command-line options (e.g. those relating to image stream compression or stereo rendering) are not applicable here, because there is no vglclient running on the other end. It is now the VNC server that handles all the image stream optimization/compression, so it is there that you should turn to for fine-tuning.
Tip: vglrun is actually just a shell script that (temporarily) sets some environment variables before running the requested application - most importantly it adds the libraries that provide all the VirtualGL functionality to LD_PRELOAD. If it better suits your workflow, you could just set those variables yourself instead. The following command lists all environment variables that vglrun would set for your particular set-up:
comm -1 -3 <(env | sort) <(vglrun env | grep -v '^\[' | sort)

Confirming that VirtualGL rendering is active

If you set the VGL_LOGO environment variable before starting an application, a small logo reading "VGL" will be shown in the bottom-right corner of any OpenGL scene that is rendered through VirtualGL in that application:

$ VGL_LOGO=1 vglrun glxgears

If the application runs but the logo does not appear, it means VirtualGL has failed to take effect (see #Troubleshooting below) and the application has probably fallen back to software rendering.

Measuring performance

Many OpenGL programs or games can display an embedded FPS ("frames per second") counter - however when using VirtualGL these values will not be very useful, as they merely measure the rate at which frames are rendered on the server side (through the 3D-capable X server), not the rate at which frames actually end up being rendered on the client side.

The "Performance Measurement" chapter of the user manual describes how to get a measurement of the throughput at various stages of the VirtualGL image pipeline, and how to identify bottlenecks (especially when using VirtualGL with X11 forwarding). When using VNC, the VNC client should be able to tell you its rendering frame-rate as well.

Troubleshooting

vglrun aborts with "Could not open display"

If exits with an error messages like...

[VGL] ERROR: Could not open display :0.

...in the shell output, then this means that the 3D-capable X server on the server side (that is supposed to handle the OpenGL rendering) is either not running, or not properly set up for use with VirtualGL (see Installation and setup), or is not set correctly (see Running Applications). If it used to work but not anymore, a package upgrade may have overwritten files modified by vglserver_config, so run that script again and then restart the server-side X server.

vglrun seems to have no effect at all

Symptoms:

  • no VirtualGL-accelerated 3D rendering - the program either aborts, or falls back to software rendering (how to check)
  • at the same time, no VirtualGL related error messages or info is printed to the shell

This may happen when something blocks VirtualGL from getting preloaded into the application's executable(s). The way pre-loading works, is that adds the names of some VirtualGL libraries to the environment variable before running the command that starts the application. Now when an application binary is executed as part of this command, the Linux kernel loads the dynamic linker which in turn detects the variable and links the specified libraries into the in-memory copy of the application binary before anything else. This will obviously not work if the environment variable is not propagated to the dynamic linker, e.g. in the following cases:

  • The application is started through a script that explicitly unsets/overrides LD_PRELOAD
Solution: Edit the script to comment out or fix the offending line. (You can put the modified script in to prevent it from being reverted on the next package upgrade.)
  • The application is started through multiple layers of scripts, and environment variables get lost along the way
Solution: Modify the final script that actually runs the application, to make it run the application with .
  • The application is started through a loader binary (possibly itself!), in a way that fails to propagate LD_PRELOAD
Solution: If possible, bypass the loader binary and start the actual OpenGL application directly with - an example is VirtualBox where you need to start your virtual machine session directly with rather then through the VirtualBox main program GUI. If it is a matter of LD_PRELOAD being explicitly unset within the binary, running with the -ge command-line switch can prevent that in some cases.

See the "Application Recipes" section in the user manual for a list of some applications that are known to require such work-arounds.

vglrun fails with ld.so errors

If VirtualGL-accelerated 3D rendering does not work (like with the previous section), but in addition you see error messages like...

ERROR: ld.so: object 'libdlfaker.so' from LD_PRELOAD cannot be preloaded: ignored.
ERROR: ld.so: object 'librrfaker.so' from LD_PRELOAD cannot be preloaded: ignored.

...in the shell output, then the dynamic linker is correctly receiving instructions to preload the VirtualGL libraries into the application, but something prevents it from successfully performing this task. Three possible causes are:

  • The VirtualGL libraries for the correct architecture are not installed
To run a 32-bit application (like Wine) with VirtualGL, you need to install from the multilib repository.
  • The application executable has the setuid/setgid flag set
You can confirm whether this is the case by inspecting the executable's file permissions using : It will show the letter in place of the user executable bit if setuid is set (for example ), and in place of the group executable bit if setgid is set. For such an application any preloading attempts will fail, unless the libraries to be preloaded have the setuid flag set as well. You can set this flag for the VirtualGL libraries in question by executing the following as root:
$ chmod u+s /usr/lib/lib{rr,dl}faker.so    # for the native-architecture versions provided by <span class="plainlinks archwiki-template-pkg">[https://archlinux.org/packages/?name=virtualgl virtualgl]</span>
$ chmod u+s /usr/lib32/lib{rr,dl}faker.so  # for the multilib versions provided by <span class="plainlinks archwiki-template-pkg">[https://archlinux.org/packages/?name=lib32-virtualgl lib32-virtualgl]</span>
However, make sure you fully understand the security implications of setuid before deciding to do this in a server environment where security is critical.
  • You might need to specify the full path of the VirtualGL libraries
Open /usr/bin/vglrun and specify the libraries' full path in the LD_PRELOAD variable. Example:

vglrun fails with ERROR: Could not connect to VGL client.

If your 'client' program is running on the same server as virtualGL (e.g. if you are using virtualGL for VNC), try using .

Error messages about /etc/opt/VirtualGL/vgl_xauth_key not existing

This means that vglgenkey is either not being run at all for your virtualGL X server, or that it is being run again by another X server. For me, lightdm was running vglgenkey on the wrong (vnc remote) X servers, because vglserver_config adds the following: Changing it to so it only runs on the first X server fixed my problem.

vglrun fails with ERROR: VirtualGL attempted to load the real glXCreatePbuffer function and got the fake one instead.

This means that VirtualGL is trying to load a function from the wrong library. You can specify which OpenGL library to use by setting to the path of the library. appears to work for 64-bit applications. Keep in mind that 32-bit applications (like Steam or Wine) will require 32-bit OpenGL. If you need to use both 32-bit and 64-bit libraries, you can load them both with .

All applications run with 1 frame per second

If you use newer Nvidia drivers (e.g., version 440) you might be affected by a screen locking problem, which will reduce the framerate to approx. 1 frame per second according to the VirtualGL mailing list. Instead of downgrading the Nvidia driver one workaround is to set HardDPMS to in your X server configuration (see NVIDIA/Troubleshooting#Driver 415: HardDPMS for details).

rendering glitches, unusually poor performance, or application errors

OpenGL has a really low-level and flexible API, which means that different OpenGL applications may come up with very different rendering techniques. VirtualGL's default strategy for how to redirect rendering and how/when to capture a new frame works well with most interactive 3D programs, but may prove inefficient or even problematic for some applications. If you suspect that this may be the case, you can tweak VirtualGL's mode of operation by setting certain environment variables before starting your application with . For example you could try setting some of the following values (try them one at a time, and be aware that each of them could also make things worse!):

VGL_ALLOWINDIRECT=1
VGL_FORCEALPHA=1
VGL_GLFLUSHTRIGGER=0
VGL_READBACK=pbo
VGL_SPOILLAST=0
VGL_SYNC=1  # use VNC with this one, it is very slow with X11 forwarding

A few OpenGL applications also make strong assumptions about their X server environment or loaded libraries, that may not be fulfilled by a VirtualGL set-up - thus causing those applications to fail. The environment variables , VGL_GLLIB, , , can be used to fix this in some cases.

See the "Advanced Configuration" section in the user manual for a proper explanation of all supported environment variables, and the "Application Recipes" section for info on some specific applications that are known to require tweaking to work well with VirtualGL.

Xpra: vglrun uses rendering device llvmpipe only

You need two Xorg servers running: One that Xpra attaches to, e.g. at display :10. And a second one to do the actual rendering using you graphics card, e.g. your default Xorg server that is run using startx at display :0.

gollark: Spreading to disks, anyway.
gollark: It used to have that but the admins insisted I turn it off.
gollark: I'm always looking for ideas for the potato operational system.
gollark: Also, potatOS feature ideas anyone?
gollark: If someone somehow makes a virus which spreads through potatOS they should call it the "potato blight".

See also

This article is issued from Archlinux. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.