Austin, Texas— “This is the largest chip project endeavor in history of humanity, with a budget of several billion dollars,” Nvidia CEO Jen-Hsun Huang said proudly onstage, a smile on his face. ““I’m pretty sure you can go to Mars [for that budget].” And the result? Nvidia’s new GeForce GTX 1080 is far faster than the Titan X, the fastest graphics card in the world today—and even faster than two GTX 980s in SLI.
And with that, the next-generation graphics card wars are officially on. Hot damn.
That sort of performance and efficiency drives home the sort of graphical glories that the long-awaited leap to 16 nanometer process technology in GPU transistors enable. After four long years of being stuck on the 28nm transistor process node, both of the big graphical powerhouses in PC gaming are finally transitioning to GPU architectures with smaller transistors—Nvidia with the Pascal GPU in the GTX 1080 and GTX 1070, and AMD with Polaris, which is built with on 14nm process. Without getting too technical, the long-awaited jump should let AMD and Nvidia’s graphics cards provide big leaps in power efficiency and performance.
While AMD’s early Polaris demos showed off the power-efficiency advantages of the technological advancement—with an early Polaris GPU running Star Wars Battlefront at 1080p at just 86 watts of full system power, compared to 140 watts on an otherwise identical system running Nvidia’s GTX 950 —Team Green opted to come out with the big guns blazing. Again: The GeForce GTX 1080 will be faster than two GTX 980s in SLI, according to Huang. And the GTX 1070 will outpunch a Titan X.
The GeForce GTX 1080 is Nvidia’s first consumer graphics cards built around the 16 nanometer manufacturing process—a technological leap bolstered by the addition of FinFET technology, as well, as first revealed in the monstrous Tesla P100 in April.
Unlike the Tesla P100, the GeForce GTX 1080 skips the revolutionary high-bandwidth memory first revealed in AMD’s Fury lineup. Instead, it features 8GB of Micron’s new GDDR5X technology—essentially a more advanced version of the GDDR5 memory used in graphics cards for years now—clocked at a whopping 10GHz. Before you shake your head in disappointment, the first-gen HBM available today is limited to 4GB in capacity, and second-gen HBM2 isn’t expected to land until the beginning of 2017. Even AMD’s next HBM-equipped graphics cards aren’t expected until 2017.
Huang didn’t reveal many technical details, but Epic’s Tim Sweeney took the stage to run a demo of the studio’s upcoming MOBA Paragon. At the end of the demo, Huang called for a overlay of the system’s specifications to be pulled up. The GeForce GTX 1080 was clocked at a whopping 2,114MHz—one of the highest GPU clock speeds ever recorded, Huang said—and staying still at an icy cool 67 degrees Celcius. “It’s overclockability is just incredible,” Huang boasted.
Even better: Nvidia’s new cards are coming soon and won’t break the bank, especially when you compare their performance to the GTX 900-series. The GTX 1080 will cost $600, or $700 for an Nvidia designed edition, on May 27. The GTX 1070, meanwhile, will cost $380—or $450 for an Nvidia edition—on June 10.
But what are you supposed to do with all that firepower?
Nvidia showcased a fancy new technology called “simultaneous multi-projection” that puts Pascal’s oomph to good work. Simultaneous multi-projection improves how game looks on multiple displays, while also boosting your frame rate. Traditionally, when you’re using a multi-display setup, the GPU still projects a flat image across all three screens, which creates distortion in the image. Think of folding a paper with a picture on it—rather than looking at a straight line across it, it looks like two slightly bent lines. Simultaneous multi-projection dynamically adjusts the image to look natural across your screens, so those straight lines appear straight yet again.
It sounds silly, but it could potentially fix a problem point in PC gaming—and it works across a whopping 16 viewpoints, be they traditional displays or VR eye pieces. And to create the experience, Nvidia’s GPU’s pre-distort the image in a manner similar to Nvidia VRWorks’ multi-resolution shading, and only renders the area you’re looking at in full resolution. That lets the system create higher framerates by focusing on what’s important—in one demo, the frame rate jumped from roughly 60fps to 96fps with simultaneous multi-projection enabled.
But that wasn’t all.
Huang also revealed some snazzy new software tricks, starting with Ansel, which Huang touted as the “world’s first in-game 3D camera system” while mocking the traditional method of pressing Print screen or F12 in Steam to capture your gaming glories. Ansel—which Huang says was inspired by digital game artists let Duncan “Dead End Thrills” Harris—is an SDK that plugs into games and is built directly into Nvidia’s drivers.
Ansel lets you do far more than simply capture a still image of a scene. The SDK supports free-floating cameras, various vignettes and filters, support for custom resolutions up to 32 times higher than standard monitors—in a demo, resolutions up to 61,000 pixels wide were shown—and perhaps most interestingly, 360 degree image support, which lets you explore wrap-around scenes you captured in a VR headset like HTC Vive or even a Google Cardboard headset.
Look for support for Ansel to land in games like Witcher 3, Lawbreakers, Paragon, and The Division at some point in the future. It’s kind of a bummer that (it appears) developers have to explicitly build support into their games for the technology, but what a cool, exciting technology it is.
Nvidia also revealed “VRWorks Audio,” a new addition to the Gameworks/VRWorks toolset that uses fancy software tricks to deliver true positional audio in virtual reality games. We won’t get into the details, but it should make VR games even more immersive—and you can try it yourself in a new VR app coming out dubbed Nvidia Funhouse, which is essentially a series of carnival games that showcase Nvidia’s VRWorks technology.
All impressive stuff indeed. It’s your move, AMD—I suspect we’ll hear more about Polaris-based graphics cards sooner than later.