vGPU on XenServer is awsome technology. But as with any other software, it can happen that you run into issues. For example, when starting a VM which has a vGPU profile assigned, the VM can fail with the dreaded “vgpu exited unexpectedly” error. To be more precise, such an error could read:
Last week, NVIDIA released an update for the vGPU driver and manager (also described on The Citrix Blog). Besides bug fixes, this release introduces 3 new vGPU profiles, 1 new profile for the GRID K2 card (K280Q) and 2 new profiles for the GRID K1 card (K160Q and K180Q).
Virtual display heads
Max resolution per head
Max vGPUs per GPU
Max vGPUs per board
As you can see, the K180Q and K280Q profiles will assign an entire GPU to a VM. This can be compared to GPU pass-through, but leveriging the functionality of vGPU.
A few days ago, I wrote a post about Citrix’s fix for fullscreen applications in HDX 3D Pro connections. In that post I mentioned the fact that by default, an application launched in a resolution larger than the session’s resolution, will be trimmed off. Rachel Berry (Product Manager HDX Graphics for XenDesktop/XenApp, twitter @rhbBSE) provided me with some feedback on my article. Especially this part:
I understand why the application behaves in this way. The best way to fix such an challenge is to “scale” the application (I don’t even know if this would be possible programmatically) so that it scales while it maintains the aspect ratio. This way you would always see the entire application, no matter what the resolution is.
One feature of Citrix Receiver, which I knew exists, but never used, is the ability to scale the session completely. I investigated this feature as solution for the trimming of high-resolution applications.
In my previous post about full screen applications in HDX 3D Pro, I explained possible solutions to the “Full Screen Applications are not supported in Citrix HDX 3D Pro” message. In short, the issue is due to the fact that the application does not use the same resolution as your HDX session.
Implementing GPUs in your virtual desktop environment (for example vGPU or GPU pass-though) allows better use of graphical intensive applications (like Catia, AutoCAD, Photoshop, etc.). These applications have an impact on utilization (CPU, memory, disk I/O) in your infrastructure. Not only the applications will impact your infrastructure, sending the graphical intensive data over the LAN or WAN will at least impact your network and the CPU on the end-point and inside your virtual desktop. In this article I will explain the impact on network level.
Implementing a graphics card in your server allows better performance for GPU-intensive applications in VDI and SBC environments. With Citrix and NVIDIA’s vGPU innovations, this is becoming more popular. However, if you don’t plan your GPU deployment correctly, user acceptance of these new innovations will be difficult. This article will contain some things to keep in mind while planning your GPU deployment in virtualized environment. While I created this article with XenDesktop in mind, it does apply to any other VDI or SBC technology like Vmware View, RemoteFX and XenApp.
If you’re used to working with VMware, you probably know about the nifty little utility esxtop. This tool allows you to view real-time performance data on a ESX host. It also allows you to gather performance metrics in batch mode, which allows you to analyze the data in eg. Excel or ESXplot.
For those who are working with XenServer, a similar utility is available called xentop. The xentop utility allows you to monitor real-time performance metrics, like esxtop does. However, gathering data like in esxtop is a challenge.