1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Nvidia GTX 980 Owners Thread

Discussion in 'Graphics Cards' started by smogsy, Sep 19, 2014.

  1. smogsy


    Joined: Dec 9, 2006

    Posts: 7,661

    Location: @ManCave



    Tech Slides


    Benchmark Threads

    Other Threads

    Useful Information
    Direct X 11 & 12 http://www.anandtech.com/show/8544/microsoft-details-direct3d-113-12-new-features


    Overclockers NVIDIA Range can be found here

    What is G-SYNC
    NVIDIA G-SYNC is ground breaking new display technology that delivers the smoothest and fastest gaming experience ever.
    G-SYNC’s revolutionary performance is achieved by synchronizing display refresh rates to the GPU in your GeForce GTX-powered PC, eliminating screen tearing and minimizing display stutter and input lag.
    The result: scenes appear instantly, objects look sharper, and game play is super smooth, giving you a stunning visual experience and a serious competitive edge.

    The Problem: Old Tech
    When TVs were first developed they relied on CRTs which work by scanning a beam of electrons across the surface of a phosphorus tube. This beam causes a pixel on the tube to glow, and when enough pixels are activated quickly enough the CRT can give the impression of full motion video. Believe it or not, these early TVs had 60Hz refresh rates primarily because the United States power grid is based on 60Hz AC power. Matching TV refresh rates to that of the power grid made early electronics easier to build, and reduced power interference on the screen.
    By the time PCs came to market in the early 1980s, CRT TV technology was well established and was the easiest and most cost effective technology for utilize for the creation of dedicated computer monitors. 60Hz and fixed refresh rates became standard, and system builders learned how to make the most of a less than perfect situation. Over the past three decades, even as display technology has evolved from CRTs to LCD and LEDs, no major company has challenged this thinking, and so syncing GPUs to monitor refresh rates remains the standard practice across the industry to this day.
    Problematically, graphics cards don’t render at fixed speeds. In fact, their frame rates will vary dramatically even within a single scene of a single game, based on the instantaneous load that the GPU sees. So with a fixed refresh rate, how do you get the GPU images to the screen? The first way is to simply ignore the refresh rate of the monitor altogether, and update the image being scanned to the display in mid cycle. This we call ‘VSync Off Mode’ and it is the default way most gamers play. The downside is that when a single refresh cycle show 2 images, a very obvious “tear line” is evident at the break, commonly referred to as screen tearing. The established solution to screen tearing is to turn VSync on, to force the GPU to delay screen updates until the monitor cycles to the start of a new refresh cycle. This causes stutter whenever the GPU frame rate is below the display refresh rate. And it also increases latency, which introduces input lag, the visible delay between a button being pressed and the result occurring on-screen.
    Worse still, many players suffer eyestrain when exposed to persistent VSync stuttering, and others develop headaches and migraines, which drove us to develop Adaptive VSync, an effective, critically-acclaimed solution. Despite this development, VSync’s input lag issues persist to this day, something that’s unacceptable for many enthusiasts, and an absolute no-go for eSports pro-gamers who custom-pick their GPUs, monitors, keyboards, and mice to minimize the life-and-death delay between action and reaction.

    Enter NVIDIA G-SYNC, which eliminates screen tearing, VSync input lag, and stutter. To achieve this revolutionary feat, we build a G-SYNC module into monitors, allowing G-SYNC to synchronize the monitor to the output of the GPU, instead of the GPU to the monitor, resulting in a tear-free, faster, smoother experience that redefines gaming.
    Industry luminaries John Carmack, Tim Sweeney, Johan Andersson and Mark Rein have been bowled over by NVIDIA G-SYNC’s game-enhancing technology. Pro eSports players and pro-gaming leagues are lining up to use NVIDIA G-SYNC, which will expose a player’s true skill, demanding even greater reflexes thanks to the unnoticeable delay between on-screen actions and keyboard commands. And in-house, our diehard gamers have been dominating lunchtime LAN matches, surreptitiously using G-SYNC monitors to gain the upper hand.
    Online, if you have a NVIDIA G-SYNC monitor you’ll have a clear advantage over others, assuming you also have a low ping.
    How To Upgrade To G-SYNC
    If you’re as excited by NVIDIA G-SYNC as we are, and want to get your own G-SYNC monitor, you can buy a modded monitor now. Select NVIDIA System Builders will be offering ASUS VG248QE monitor

    How To Upgrade To G-SYNC
    If you’re as excited by NVIDIA G-SYNC as we are, and want to get your own G-SYNC monitor, you can buy a modded monitor now.
    Select NVIDIA System Builders will be offering ASUS VG248QE monitors that have been specially upgraded with an NVIDIA G-Sync module. These are now available to buy here.


    What is NVIDIA PhysX Technology?
    NVIDIA® PhysX® is a powerful physics engine enabling real-time physics in leading edge PC games.
    PhysX software is widely adopted by over 150 games and is used by more than 10,000 developers. PhysX is optimized for hardware acceleration by massively parallel processors.
    GeForce GPUs with PhysX provide an exponential increase in physics processing power taking gaming physics to the next level.

    What is physics for gaming and why is it important?
    Physics is the next big thing in gaming. It's all about how objects in your game move, interact, and react to the environment around them. Without physics in many of today's games, objects just don't seem to act the way you'd want or expect them to in real life. Currently, most of the action is limited to pre-scripted or ‘canned' animations triggered by in-game events like a gunshot striking a wall. Even the most powerful weapons can leave little more than a smudge on the thinnest of walls; and every opponent you take out, falls in the same pre-determined fashion.
    Players are left with a game that looks fine, but is missing the sense of realism necessary to make the experience truly immersive. With NVIDIA PhysX technology, game worlds literally come to life: walls can be torn down, glass can be shattered, trees bend in the wind, and water flows with body and force.
    NVIDIA GeForce GPUs with PhysX deliver the computing horsepower necessary to enable true, advanced physics in the next generation of game titles making canned animation effects a thing of the past.

    Which NVIDIA GeForce GPUs support PhysX?
    The minimum requirement to support GPU-accelerated PhysX is a GeForce 8-series or later GPU with a minimum of 32 cores and a minimum of 256MB dedicated graphics memory. However, each PhysX application has its own GPU and memory recommendations. In general, 512MB of graphics memory is recommended unless you have a GPU that is dedicated to PhysX.

    How does PhysX work with SLI and multi-GPU configurations?
    When two, three, or four matched GPUs are working in SLI, PhysX runs on one GPU, while graphics rendering runs on all GPUs. The NVIDIA drivers optimize the available resources across all GPUs to balance PhysX computation and graphics rendering. Therefore users can expect much higher frame rates and a better overall experience with SLI.
    A new configuration that’s now possible with PhysX is 2 non-matched (heterogeneous) GPUs. In this configuration, one GPU renders graphics (typically the more powerful GPU) while the second GPU is completely dedicated to PhysX. By offloading PhysX to a dedicated GPU, users will experience smoother gaming.

    Why is a GPU good for physics processing?
    The multithreaded PhysX engine was designed specifically for hardware acceleration in massively parallel environments.
    GPUs are the natural place to compute physics calculations because, like graphics, physics processing is driven by thousands of parallel computations.
    Today, NVIDIA's GPUs, have as many as 480 cores, so they are well-suited to take advantage of PhysX software.
    NVIDIA is committed to making the gaming experience exciting, dynamic, and vivid. The combination of graphics and physics impacts the way a virtual world looks and behaves.

    Can I use an NVIDIA GPU as a PhysX processor and a non-NVIDIA GPU for regular display graphics?
    No. There are multiple technical connections between PhysX processing and graphics that require tight collaboration between the two technologies.
    To deliver a good experience for users, NVIDIA PhysX technology has been fully verified and enabled using only NVIDIA GPUs for graphics.

    can i use an AMD card?
    PhysX supports both CPU and GPU simulation. GPU simulation is only available on Nvidia graphics card irrespective of CPU.
    Well written application usually check for GPU simulation support and allow for CPU fallback if its not present.
    However, CPU PhysX can seriously effect Framerate

    Games that support physX
    See link below:


    Quick Understanding
    SLI is the technology that combats AMD's Crossfire Technology

    NVIDIA SLI intelligently scales graphics performance by combining multiple GeForce GPUs on an SLI certified motherboard. With over 1,000 supported applications and used by over 94% of multi-GPU PCs on Steam, SLI is the technology of choice for gamers who demand the very best.
    SLI features an intelligent communication protocol embedded in the GPU, a high-speed digital interface to facilitate data flow between the two graphics cards, and a complete software suite providing dynamic load balancing, advanced rendering, and compositing to ensure maximum compatibility and performance in today’s latest games.

    Beyond just better performance, SLI offers a host of advanced features. For PhysX games, SLI can assign the second GPU for physics computation, enabling stunning effects such as life-like fluids, particles, and destruction. For CUDA applications, a second GPU can be used for compute purposes such as Folding@home or video transcoding. Finally, for the ultimate in image quality, SLI antialiasing offers up to 64xAA with two GPUs, 96xAA with three GPUs, or 128xAA with four GPUs.

    Thanks to Fermi’s architectural innovations, SLI scaling is higher than ever. Across many popular titles, over 80%, and at times 100% performance improvement, can be obtained by adding a second GPU.

    Not just Multiple Video Cards
    When SLI was first introduced the technology was used only to connect multiple video cards. In 2005, however, Gigabyte introduced a video card that used SLI technology to connect two different Nvidia GPUs located on the same video card.
    This arrangement has become more common over time. Both Nvidia and AMD have released reference design cards featuring two GPUs in the same video card connected via SLI or CrossFire.
    This has confused things a bit because two video cards with two GPUs each would technically be a quad-SLI arrangement even though only two video cards are involved. With that said, these cards are expensive and thus rare, so you can generally assume that if someone is talking about SLI they are talking about the use of two or more video cards.
    SLI usually describes a desktop solution but it is available in gaming laptops. AMD sometimes pairs its APUs with a discrete Radeon GPU, which means you’ll sometimes run across CrossFire laptops that only cost $600 to $800 bucks.
    Nvidia has also paired a discrete GPU with an integrated GPU in the past. This was branded with the term Hybrid SLI. Nvidia was forced out of the chipset business soon after, however, which meant the company no longer offered integrated graphics. Hybrid SLI is effectively dead as a result.


    Everything you need to know about SLI

    Benchmarks & Guide

    Titan Quad SLI Demo

    NVIDIA GameWorks™
    pushes the limits of gaming by providing a more interactive and cinematic game experience and thus enabling next gen gaming for current games.
    We provide technologies e.g. PhysX and VisualFX, which are easy to integrate into games as well as tutorials and tools to quickly generate game content.
    In addition we also provide tools to debug, profile and optimize your code.

    Upcoming Technology
    NVIDIA FlameWorks enables cinematic smoke, fire and explosions. It combines a state-of-the-art grid based fluid simulator with an efficient volume rendering engine.
    The system is highly customizable, and supports user-defined emitters, force fields, and collision objects

    PhysX FleX is a particle based simulation technique for real-time visual effects. It will be introduced as a new feature in the upcoming PhysX SDK v3.4.
    The FleX pipeline encapsulates a highly parallel constraint solver that exploits the GPU’s compute capabilities effectively.

    NVIDIA GameWorks technology in released games
    Call of Duty: Ghosts is using NVIDIA HairWorks to provide a more realistic Riley and wolves. Each hair asset has about 400-500K hair strands.
    Most of these hair assets are created on the fly inside the GPU from roughly 10K guide hairs.
    Additional technologies used in the game are NVIDIA Turbulence for the smoke bombs, as well as TXAA.

    Batman Arkham Origins is loaded with NVIDIA GameWorks technologies; NVIDIA Turbulence for the snow, steam/fog and shock gloves as well as PhysX Cloth for ambient cloth.
    In addition NVIDIA ShadowWorks for HBAO+ and advanced soft shadows and NVIDIA CameraWorks for TXAA and DoF.

    ShadowPlay From Nvidia
    Don't just brag about your gaming wins. Show the world with GeForce ShadowPlay.
    Available only with NVIDIA® GeForce Experience™, ShadowPlay records game action as you play—automatically—with minimal impact on performance. It's fast, simple, and free!
    Shadow every game.
    Share every victory.
    ShadowPlay records the up to the last 20 minutes of your gameplay.
    Just pulled off an amazing stunt? Hit a hotkey and the game video will be saved to disk. Or, use the manual mode to capture video for as long as you like.
    You now have a video record of your gaming awesomeness to share or post anywhere.
    ShadowPlay even lets you instantly broadcast your gameplay in an HD-quality stream through Twitch.tv.

    How It Works

    ShadowPlay has two user-configurable modes. The first, shadow mode, continuously records your gameplay, saving up to 20 minutes of high-quality 1920×1080 footage to a temporary file.
    So, if you pull off a particularly impressive move in-game, just hit the user-defined hotkey and the footage will be saved to your chosen directory.
    The file can then be edited with the free Windows Movie Maker application, or any other .mp4-compatible video editor, and uploaded to YouTube to share with friends or gamers galore.
    Alternatively, in manual mode, which acts like traditional gameplay recorders, you can save your entire session to disk.
    The beauty of ShadowPlay is that, because it takes advantage of the hardware built into every GTX GPU, you don’t have to worry about any major impact on frame rates compared to other, existing applications.


    GPU accelerated H.264 video encoder
    Records up to the last 20 minutes of gameplay in Shadow Mode
    Records unlimited length video in Manual Mode
    Broadcasts to Twitch
    Outputs 1080p at up to 50 Mbps
    Minimal performance impact
    Full desktop capture (desktop GPUs only)

    General System Requirements

    Operating System:
    Windows 8, 8.1
    Windows 7
    Windows Vista (DirectX 11 Runtime required)
    Windows XP SP3 (driver updates only)
    2GB system memory
    Supported Hardware:

    Intel Pentium G Series, Core 2 Duo, Quad Core i3, i5, i7, or higher
    AMD Phenom II, Athlon II, Phenom X4, FX or higher

    Disk Space Required:
    20MB minimum
    Internet Connectivity:
    Display Resolution:
    Any display with 1024x768 to 3840×2160 resolution

    GPU Series Driver Updates Game Optimization ShadowPlay and SHIELD PC Streaming
    GeForce TITAN, GTX 700,
    GTX 600 Yes Yes Yes
    GeForce GTX 800M, GTX 700M, select GTX 600M Yes Yes Yes
    GeForce 800, 700, 600, 500, 400 Yes Yes No
    GeForce 600M, 500M, 400M Yes Yes No
    GeForce 300, 200, 100, 9, 8 Yes No No
    GeForce 300M, 200M, 100M, 9M, 8M Yes No No

    - See more at: http://blogs.nvidia.com/blog/2013/10/18/shadowplay/#sthash.l0GzGizW.dpuf

    3D Vision

    3D Vision Vision
    3D Vision (previously GeForce 3D Vision) is a stereoscopic gaming kit from Nvidia which consists of LC shutter glasses and driver software which enables stereoscopic vision for any Direct3D game,
    with various degrees of compatibility. There have been many examples of shutter glasses over the past decade, but the NVIDIA 3D Vision gaming kit introduced in 2008 made this technology available for mainstream consumers and PC gamers.
    The kit is specially designed for 120 Hz LCD monitors but is compatible with CRT monitors (some of which may work at 1024×768×120 Hz and even higher refresh rates), DLP-projectors, and others.
    It requires a compatible graphics card from Nvidia (GeForce 200 series or later).

    Shutter Glasses
    The glasses use wireless IR protocol and can be charged from a USB cable, allowing around 60 hours of continuous use.
    The wireless emitter connects to the USB port and interfaces with the underlying driver software. It also contains a VESA Stereo port for connecting supported DLP TV sets, although standalone operation without a PC with installed Nvidia 3D Vision driver is not allowed.
    NVIDIA includes one pair of shutter glasses in their 3D Vision kit, SKU 942-10701-0003. Each lens operates at 60 Hz, and alternate to create a 120 Hz 3-dimensional experience.
    This version of 3D Vision supports select 120 Hz monitors, 720p DLP projectors, and passive-polarized displays from Zalman.

    Stereo Driver
    The stereo driver software can perform automatic stereoscopic conversion by using the 3D models submitted by the application and rendering two stereoscopic views instead of the standard mono view. The automatic driver works in two modes: fully "automatic" mode, where 3D Vision driver controls screen depth (convergence) and stereo separation, and "explicit" mode, where control over screen depth, separation, and textures is performed by the game developer with the use of proprietary NVAPI.
    The quad-buffered mode allows developers to control the rendering, avoiding the automatic mode of the driver and just presenting the rendered stereo picture to left and right frame buffers with associated back buffers.

    3D Vision Requirements
    see here:

    What is supported? (Games/Web/Photos/Blu-Rays)
    see here for the full list:


    Performance Review

    NVIDIA Surround Technology
    In NVIDIA Surround three displays are supported up to 2560x1600 resolution on each display,
    the same as AMD’s Eyefinity, although Eyefinity can currently support up to six displays at 2560x1600 each.

    How many displays do I need to run 3D Vision Surround?
    3D Vision Surround requires three displays or projectors. You can use a variety of displays, including LCD monitors, TVs, or projectors. Please view the 3D Vision Surround system requirements for a full list of supported products
    Please note: NVIDIA 3D Vision Surround and NVIDIA Surround does not support a two display surround configuration. Both NVIDIA 3D Vision Surround and NVIDIA Surround require three supported displays as defined in the system requirements above.

    Does 3D Vision Surround support 2D displays?
    Yes, you can run Surround in 2D mode. We call this mode Surround (2D). Make sure that 3D Vision is disabled from the Stereoscopic 3D control panel to use Surround (2D) mode.

    What GPUs does 3D Vision Surround support?
    Please view the 3D Vision Surround system requirements for a full list of supported products.
    Please note that in the first Beta v258.69 driver for 3D Vision Surround, we do not recommend running GeForce GTX 295 Quad SLI because there are bugs in the driver which may result in system instability. We will be providing a future driver update to support this configuration.

    Can I use HDMI connectors to run 3D Vision Surround?
    3D Vision-Ready LCDs require dual-link DVI connectors and will not work with HDMI connectors. However, you can use HDMI connectors with 3D Vision Projectors that support HDMI input.
    Note: 3D DLP HDTVs do use HDMI connectors, but 3D Vision Surround does not support using 3D DLP TVs.

    What resolutions does Surround support?
    Surround requires three displays to operate. Surround resolutions are calculated by arranging your monitors in portrait or landscape mode and then multiplying the horizontal resolution by three times a single display in that mode (portrait or landscape).
    The following table illustrates some common aspect ratios and resolutions:

    Note: 3D Vision-Ready LCDs do not support portrait mode. However, you can run 3D Vision Surround with 3D projectors in portrait mode.

    Can I use 3D Vision Surround with three displays and setup an Accessory display that is not in 3D?
    At this time, accessory displays are only supported in Surround (2D) configuration.

    CUDA™ is a parallel computing platform and programming model invented by NVIDIA. It enables dramatic increases in computing performance by harnessing the power of the graphics processing unit (GPU).
    With millions of CUDA-enabled GPUs sold to date, software developers, scientists and researchers are finding broad-ranging uses for GPU computing with CUDA. Here are a few examples:
    Identify hidden plaque in arteries: Heart attacks are the leading cause of death worldwide. Harvard Engineering, Harvard Medical School and Brigham & Women's Hospital have teamed up to use GPUs to simulate blood flow and identify hidden arterial plaque without invasive imaging techniques or exploratory surgery.
    Analyze air traffic flow: The National Airspace System manages the nationwide coordination of air traffic flow. Computer models help identify new ways to alleviate congestion and keep airplane traffic moving efficiently.
    Using the computational power of GPUs, a team at NASA obtained a large performance gain, reducing analysis time from ten minutes to three seconds.
    Visualize molecules: A molecular simulation called NAMD (nanoscale molecular dynamics) gets a large performance boost with GPUs. The speed-up is a result of the parallel architecture of GPUs, which enables NAMD developers to port compute-intensive portions of the application to the GPU using the CUDA Toolkit.

    Widely Used By Researchers
    Since its introduction in 2006, CUDA has been widely deployed through thousands of applications and published research papers, and supported by an installed base of over 500 million CUDA-enabled GPUs in notebooks, workstations, compute clusters and supercomputers.

    How to get started
    Software developers, scientists and researchers can add support for GPU acceleration in their own applications using one of three simple approaches:
    Drop in a GPU-accelerated library to replace or augment CPU-only libraries such as MKL BLAS, IPP, FFTW and other widely-used libraries
    Automatically parallelize loops in Fortran or C code using OpenACC directives for accelerators
    Develop custom parallel algorithms and libraries using a familiar programming language such as C, C++, C#, Fortran, Java, Python, etc.

    What is NVIDIA Tesla™?

    With the world’s first teraflop many-core processor, NVIDIA® Tesla™ computing solutions enable the necessary transition to energy efficient parallel computing power. With thousands of CUDA cores per processor , Tesla scales to solve the world’s most important computing challenges—quickly and accurately.

    What is OpenACC?
    OpenACC is an open industry standard for compiler directives or hints which can be inserted in code written in C or Fortran enabling the compiler to generate code which would run in parallel on multi-CPU and GPU accelerated system. OpenACC directives are easy and powerful way to leverage the power of GPU Computing while keeping your code compatible for non-accelerated CPU only systems. Learn more at https://developer.nvidia.com/openacc.

    What kind of performance increase can I expect using GPU Computing over CPU-only code?
    This depends on how well the problem maps onto the architecture. For data parallel applications, accelerations of more than two orders of mangitude have been seen. You can browse research, developer, applications and partners on our CUDA In Action Page

    What operating systems does CUDA support?

    CUDA supports Windows, Linux and Mac OS. For full list see the latest CUDA Toolkit Release Notes.The latest version is available at http://docs.nvidia.com

    Which GPUs support running CUDA-accelerated applications?
    CUDA is a standard feature in all NVIDIA GeForce, Quadro, and Tesla GPUs as well as NVIDIA GRID solutions. A full list can be found on the CUDA GPUs Page.

    What is the "compute capability"?
    The compute capability of a GPU determines its general specifications and available features. For a details, see the Compute Capabilities section in the CUDA C Programming Guide.

    Where can I find a good introduction to parallel programming?
    There are several university courses online, technical webinars, article series and also several excellent books on parallel computing. These can be found on our CUDA Education Page.

    Hardware and Architecture
    Will I have to re-write my CUDA Kernels when the next new GPU architecture is released?
    No. CUDA C/C++ provides an abstraction; it’s a means for you to express how you want your program to execute. The compiler generates PTX code which is also not hardware specific. At runtime the PTX is compiled for a specific target GPU - this is the responsibility of the driver which is updated every time a new GPU is released. It is possible that changes in the number of registers or size of shared memory may open up the opportunity for further optimization but thats optional. So write your code now, and enjoy it running on future GPU's

    Does CUDA support multiple graphics cards in one system?
    Yes. Applications can distribute work across multiple GPUs. This is not done automatically, however, so the application has complete control. See the "multiGPU" example in the GPU Computing SDK for an example of programming multiple GPUs.

    Nvidia Game Stream
    NVIDIA GameStream harnesses the power of the most advanced GeForce® GTX™ graphics cards or GRID cloud Beta and exclusive game-speed streaming technologies to bring low-latency PC gaming experience to your SHIELD portable.

    With streaming support for over 100 of the hottest PC titles at up to 1080p and 60 frames per second, SHIELD offers a truly unique handheld gaming experience.

    NEW: Remotely access your PC to play your games away from your home.*
    Play your favorite PC games on your HDTV at up to 1080p at 60 FPS using a wireless Bluetooth controller with SHIELD Console Mode.
    NEW: Stream games like World of Warcraft, League of Legends, and DOTA 2 to your HDTV using Wi-Fi and play with full Bluetooth keyboard and mouse support.
    Download GeForce Experience to see if your PC is ready for.

    How GAMESTREAM Works
    NVIDIA uses the H.264 encoder built into GeForce GTX 650 or higher desktop GPUs and GeForce GTX Kepler and Maxwell notebooks, along with efficient wireless streaming software protocol integrated into GeForce Experience, to stream games from the PC to SHIELD over the user's home Wi-Fi network with ultra-low latency. Gamers then use SHIELD as the controller and display for their favorite PC games, as well as for Steam Big Picture.
    In addition to streaming the game, NVIDIA also configures PC games for streaming using GeForce Experience to deliver a seamless out of the box experience. There are three key activities involved here:

    Generating optimal game settings for streaming games from the PC to SHIELD
    We use our GeForce Experience servers to determine the best quality settings based on the user's CPU and GPU, and target higher frame rates than 'normal' optimal settings to ensure the lowest latency gaming experience. These settings are automatically applied when the game is launched so gamers don't have to worry about configuring these settings themselves.

    Enabling controller support in games
    This avoids gamers having to manually configure controller support in the game – it 'just works' out of the box.

    Optimizing the game launch process
    This is important so that gamers can get into the game as quickly and seamlessly as possible, without hitting launchers or other 2 foot UI interactions that can be difficult to interface with a controller.

    Steam Big Picture
    Big Picture shows up as a selection in the list of SHIELD-optimized games. Launching it gives you access to your Steam games – both supported and unsupported – plus the Steam Store and all the Steam community features. Supported games are recommended for the best, optimized streaming experience – the latest list of supported games can be found on SHIELD.nvidia.com starting at launch.
    Unsupported games may work if they have native controller support, but will not be optimized for streaming and may have other streaming compatibility issues.

    System Requirements for PC Game Streaming:
    > GPU:
    - Desktop: GeForce GTX 650 or higher GPU, or
    - Notebook (Beta): GeForce GTX 800M, GTX 700M and select Kepler-based GTX 600M GPUs
    > CPU: Intel Core i3-2100 3.1GHz or AMD Athlon II X4 630 2.8 GHz or higher
    > System Memory: 4 GB or higher
    > Software: GeForce Experience™ application and latest GeForce drivers
    > OS: Windows 8 or Windows 7
    > Routers: 802.11a/g router (minimum). 802.11n dual band router (recommended). See list of GameStream ready routers.

    GameStream Ready Games

    Setup Guide

    Nvidia Workstation cards
    The NVIDIA® Quadro® family of products is designed and built specifically for professional workstations, powering more than 200 professional applications across a broad range of industries. From Manufacturing, Sciences and Medical Imaging, and Energy, to Media and Entertainment, Quadro solutions deliver unbeatable performance and reliability that make them the graphics processors of choice for professionals around the globe.

    Get the power to realise your vision with the new family of NVIDIA Quadro professional graphics. They’re fueled by Kepler, NVIDIA’s most powerful GPU architecture ever, bringing a whole new level of performance and innovative capabilities to modern workstations.
    Whether you’re creating revolutionary products, designing groundbreaking architecture, navigating massive geological datasets, or telling spectacularly vivid visual stories, Quadro graphics solutions give you the power to do it better and faster.

    Guaranteed compatibility through support for the latest OpenGL, DirectX, and NVIDIA CUDA® standards, deep professional software developer engagements, and certification with over 200 applications by software companies.
    Maximised performance through the unique capabilities of the latest Kepler GPU—including the SMX next-generation multiprocessor engine, advanced temporal anti-aliasing (TXAA) and fast approximate anti-aliasing (FXAA) modes, and innovative bindless textures technology—as well as larger on-board GPU memory and optimised software drivers.
    Transformed and accelerated workflows that drive faster time-to-results in product design or digital content creation. This is made possible through simultaneous design and rendering or simulation using Quadro and NVIDIA Tesla® cards in the same system—called NVIDIA Maximus™.


    Quadro solutions combine the most advanced display technologies and ecosystem interfaces to provide the ultimate visual workspace for maximum productivity.
    Stunning image quality with movie-quality antialiasing techniques and enhanced color depth, higher refresh rates, and ultra-high screen resolution offered by the DisplayPort standard.
    Simplified display scaling through increased display outputs per board, choice of display connections, and multi-display blending and synchronization made possible by NVIDIA Mosaic technology.
    Enhanced desktop workspace across multiple displays using intuitive placement of windows, multiple virtual desktops, and user profiles offered by the NVIDIA nView® visual workspace manager.

    Reliability is core to all Quadro solutions, and one of the keys to the decade-long industry leadership of Quadro-powered professional desktops. Every product is designed to deliver the peace of mind you need to focus on what you do best—changing the way we all look at the world.
    Highest-quality products through power-efficient hardware designs and component selection for optimum operational performance, durability, and longevity.
    Simplified software driver deployment for the IT team through a regular cadence of long-life, stable driver releases and quality-assurance processes.
    Maximum uptime through exhaustive testing in partnership with leading OEMs and system integrators that simulates the most demanding real-world conditions.

    The NVIDIA Tesla-accelerated computing co-processor pairs with Quadro products to fundamentally transform traditional workflows. Take advantage of the NVIDIA Maximus solution to simulate and visualize more design options, explore more innovative entertainment ideas, and accelerate time to market for a competitive advantage.

    Quadro graphics card for desktop workstations
    Designed and built specifically for professional workstations, NVIDIA® Quadro® GPUs power more than 200 professional applications across a broad range of industries, including Manufacturing, Media and Entertainment, Sciences, and Energy. Professionals like you trust Quadro solutions to deliver the best possible experience with applications such as Adobe CS6, Avid Media Composer, Autodesk Inventor, Dassault Systemes CATIA and SolidWorks, Siemens NX, PTC Creo, and many more. For maximum application performance, add an NVIDIA Tesla® GPU to your workstation and experience the power of NVIDIA Maximus™ technology.

    The cards
    QUADRO K6000
    The most powerful pro graphics on the planet built for tackling the largest visual projects, 12GB of on-board memory, advanced display capabilities for large-scale visualization, and support for high-performance video I/O.

    Quadro K5000
    Extreme performance to handle demanding workloads, 4 GB of on-board memory, advanced quad-display capabilities for large-scale visualization, and support for high-performance video I/O.

    Quadro K5000 for Mac
    Extreme performance on the Apple Mac Pro platform for accelerating professional design, animation, and video applications, 4 GB of on-board memory, and advanced quad-display capabilities for large-scale visualization.

    Quadro K4000
    Supercharged performance for graphics-intensive applications, large 3 GB on-board memory, multi-monitor support, and stereo capability in a single-slot configuration.

    Quadro K2000
    Outstanding performance with a range of professional applications, substantial 2 GB on-board memory to hold large models, and multi-monitor support for enhanced desktop productivity.

    Quadro K2000D
    Outstanding performance with a range of professional applications, dual-link DVI capability, 2 GB of GDDR5 memory, and multi-monitor support for enhanced desktop productivity.

    Quadro K600
    Great performance for leading professional applications, 1 GB of on-board memory, and a low- profile form factor for maximum usage flexibility.

    Quadro 410
    Entry-level professional graphics with ISV certifications.


    Last edited: Oct 1, 2014
  2. Silent_Scone


    Joined: Sep 5, 2011

    Posts: 12,393

    Location: Surrey

    Nice one. Was wondering why there wasn't one yet ;)

    Bought one EVGA reference.
  3. ratters


    Joined: Jan 18, 2009

    Posts: 620

    2 ref 980's inc... cant wait for tomorrow to follow the dpd man around the town :)
  4. DLD


    Joined: Jun 28, 2014

    Posts: 199

    I just got my 780 Classy. So I'm going to wait a while until EVGA offers something in the step up program.

    980 Classified boosting to 1405mhz out of the box. :eek:

    To bad they only offer reference cards in step up. :D
  5. thefogo


    Joined: Jul 20, 2012

    Posts: 3,286

    Location: Manchester

    Went to buy one but missed them :(
  6. smogsy


    Joined: Dec 9, 2006

    Posts: 7,661

    Location: @ManCave

    updated Will require pictures for proof though :p
  7. whyscotty


    Joined: May 6, 2012

    Posts: 9,396

    Location: Wigan

  8. JBod


    Joined: Sep 1, 2007

    Posts: 5,212

    Location: Santa Monica, California

    I got a EVGA GTX 980 arriving tomorrow. Still deciding whether to go SLI or not.
  9. Protoype


    Joined: Jan 25, 2014

    Posts: 120

    Location: - Home

    ^Just hit the button already =] good luck
  10. Besty


    Joined: Oct 18, 2002

    Posts: 3,348

    I bought a EVGA SC and a MSI card for 'evaluation'.
  11. smogsy


    Joined: Dec 9, 2006

    Posts: 7,661

    Location: @ManCave

  12. ratters


    Joined: Jan 18, 2009

    Posts: 620

    just to add my 2 980's are the GALAX ones :)

    Wondering the differing ways i can spend the time before they arrive.
  13. a.tunnard


    Joined: Jun 27, 2012

    Posts: 138

    I ordered an EVGA SC, might come Monday
  14. Clov!s


    Joined: Dec 1, 2011

    Posts: 17,303


    Nice. :) Did you go pick them up from the shop?
  15. Gregster


    Joined: Sep 24, 2008

    Posts: 37,728

    Location: Essex innit!

    Nice guys and I told you to wait Smogsy ;) Good call bud and I don't mind admitting that I am very jealous of you guys and those getting the 970s. Plenty of feedback please and plenty of benchmarks for the bench threads :)
  16. smogsy


    Joined: Dec 9, 2006

    Posts: 7,661

    Location: @ManCave

    980's :p

    just kidding :)
  17. Gregster


    Joined: Sep 24, 2008

    Posts: 37,728

    Location: Essex innit!

    They will keep your man cave cool :p You would need all of that air con unit working OT if you went AMD :p
  18. bradUK


    Joined: Jul 14, 2013

    Posts: 884

    Looks like I'll be loosing my hard won top fifteen "on air" fire strike score in a mater of hours!

    You bunch of *******s!!

    Anyone want to buy a double binned 780ti lol

    Seriously though guys, we need real world performance experiences ASAP. I'm hoping they're below par as I can't really justify injecting anymore cash into my rig over the next couple weeks!

    Very, very jealous!
    Last edited: Sep 19, 2014
  19. smogsy


    Joined: Dec 9, 2006

    Posts: 7,661

    Location: @ManCave

    trust me, this Air con powers whole Office floors normally it can handle a lot.
    might get some decent overclocks on these though!. :)
  20. ~Mike~


    Joined: Jan 17, 2011

    Posts: 8,737

    Location: the King's city

    Finally decided on your new cards then Smogsy :) why gigabyte cards thought you hated them :p

    Very nice as well Neil.