My application is quite interactive application with lots of colors and drag-and-drop functionality, but no fancy 3D-stuff or animations or video, so I only used plain GDI (no GDI Plus, No DirectX).
In the past my applications ran in desktops or laptops, and I suggested my customers to invest in a decent video card, with:
- a minimum resolution of 1280x1024
- a minimum color depth of 24 pixels
- X Megabytes of memory on the video card
Now my users are switching more and more to terminal servers, therefore my question:
What is the importance of a video card on a terminal server?
- Is a video card needed anyway on the terminal server?
- If it is, is the resolution of the remote desktop client limited to the resolutions supported by the video card on the server?
- Can the choice of a video card in the server influence the performance of the applications running on the terminal server (but shown on a desktop PC)?
- If I start to make use of graphical libraries (like Qt) or things like DirectX, will this then have an influence on the choice of video card on the terminal server?
- Are calculations in that case 'offloaded' to the video card? Even on the terminal server?
Thanks.