AMD's take and the future of Mantle
From the day Microsoft announced DirectX 12, AMD has made it clear that it's fully behind the new API. Its message is simple: Direct3D 12 "supports and celebrates" the push toward lower-level abstraction that AMD began with Mantle last year—but D3D12 won't be ready right away, and in the meantime, developers can use Mantle in order to get some of the same gains out of AMD hardware.
At GDC, AMD's Corpus elaborated a little bit on that message. He told me Direct3D 12's arrival won't spell the end of Mantle. D3D12 doesn't get quite as close to the metal of AMD's Graphics Core Next GPUs as Mantle does, he claimed, and Mantle "will do some things faster." Mantle may also be quicker to take advantage of new hardware, since AMD will be able to update the API independently without waiting on Microsoft to release a new version of Direct3D. Finally, AMD is talking to developers about bringing Mantle to Linux, where it would have no competition from Microsoft.
Corpus was adamant that developers will see value in adopting Mantle even today, with D3D12 on the horizon and no explicit support for Linux or future AMD GPUs. Because the API is similar to D3D12, it will give developers a "big head start," he said, and we may see D3D12 launch titles "very early" as a result.
Naturally, AMD can motivate developers in other ways, too. While Corpus didn't address that side of the equation, VG247 reported last year that Battlefield 4's inclusion in the Gaming Evolved program—and its support for Mantle—involved a $5-8 million payment from AMD. That figure was never confirmed officially, but it's no secret AMD's and Nvidia's developer relations and co-marketing programs often involve financial incentives. Supporting Mantle may be a financially lucrative proposition for some game studios.
Nvidia seems to see lower-level graphics APIs as less of a panacea than AMD does. Tamasi told us that, while such APIs are "great," they're "not the only answer" because they're "not necessarily great for everyone." This statement goes back to what we said earlier about developers having manual control over things currently handled by the API and driver, such as GPU memory management. Engine programming gurus like DICE's Johan Andersson and Epic's Tim Sweeney might be perfectly happy to manage resources manually, but according to Tamasi, "a lot of folks wouldn't."
Nvidia also believes there's still some untapped potential for efficiency improvements and overhead reduction in D3D11. Since Mantle's debut six months ago, Nvidia has "redoubled" its efforts to curb CPU overhead, improve multi-core scaling, and use shader caching to address stuttering problems. (Tamasi freely admitted that Mantle's release spurred the initiative. "AMD and Mantle should get credit for revitalizing . . . and getting people fired up," he said.)
We saw first-hand the results of Nvidia's work two months ago. In a CPU-limited Battlefield 4 test, Nvidia's Direct3D driver clearly performed better than AMD's. That optimization work is still ongoing:
The performance data above, supplied to us by Nvidia, shows performance improvements over successive GeForce driver releases in Oxide Game's Star Swarm stress test. That test also supports Mantle, which helps put Nvidia's D3D11 optimizations in context. Tamasi conceded AMD's Mantle version "still has less slow frames" and that D3D11 "still [has] some limiting factors," but he reiterated his overarching point, which is that it's possible to "do a much better job" with D3D11. Even going by our own, perhaps less flattering numbers, we'd say that's a fair assessment.
What about OpenGL?
Direct3D 12 holds a lot of promise, but it won't help folks running Linux-based operating systems like SteamOS. Game developers seeking to write native ports for those OSes will need to use OpenGL, and they will have to extract whatever optimizations they can out of that API.
Tamasi told us Nvidia, AMD, and Intel have all been "working hard" to help developers achieve "super high efficiency" with OpenGL. In a GDC session entitled "Approaching Zero Driver Overhead in OpenGL," folks from all three companies demonstrated best practices for OpenGL optimizations. The techniques they outlined can be exploited with the current version of the API on today's hardware with existing drivers, and they can result in large performance gains.
During the session, we saw performance numbers obtained with APItest, an open-source benchmark developed by Blizzard's Patrick Doane. In Nvidia's words, APItest is "designed to showcase and compare between different approaches to common problems encountered in real-time rendering applications." The results showed order-of-magnitude performance differences between a "naive" approach, which Tamasi described as "writing OpenGL like Direct3D," and the best practices advocated by GPU manufacturers.
In the graph above, the baseline "naive" approach is the top bar, while the last bar is what Tamasi describes as "writing good code." The difference amounts to an 18X speedup. Obviously, this is an isolated test case rather than a comprehensive, game-like scenario. But I'd say the difference is large enough to make at least some OpenGL developers rethink the way they optimize their code.
The important takeaway here, I think, is that despite their involvement with D3D12, the big three makers of PC graphics hardware—AMD, Intel, and Nvidia—all have a stake in keeping OpenGL competitive. That's good news for Linux users, and it's especially good news for those of us hoping to see SteamOS become a real competitor to Windows in the realm of PC gaming.
Of course, SteamOS isn't due out until the summer, and the first D3D12 titles aren't expected until the 2015 holiday season. We'll have to revisit these matters in the future, when we can see for ourselves how next-gen games really perform on the two platforms.
114 comments — Last by sschaem at 12:21 AM on 04/29/14
|HyperX's Alloy FPS mechanical gaming keyboard reviewedBlasting away at fluff||31|
|Nvidia's GeForce GTX 1080 Ti graphics card reviewedI like big chips and I cannot lie||191|
|Where minimum-FPS figures mislead, frame-time analysis shinesA new way to go Inside the Second||248|
|Aorus' GeForce GTX 1080 Xtreme Edition 8G graphics card reviewedFlying high||28|
|The curtain comes up on AMD's Vega architectureRadeons get ready for the workloads of the future||156|
|Nvidia unveils its GTX 1050 and GTX 1050 Ti for laptopsThe pint-size Pascal empowers portable players||16|
|AMD opens up machine learning with Radeon InstinctVega lights the way||65|
|Radeon Software Crimson ReLive Edition: an overviewStream, capture, Chill||103|
|Gigabyte has two A320 boards for bread-and-butter Ryzen builds||1|
|MSI GTX 1080 Ti Armor 11G is the first custom card on e-tail shelves||4|
|Google points deep-learning machines at audio effect subtitles||4|
|Throw a Quadro card on Gigabyte's Z270X-Designare||10|
|Deals of the week: an RX 480 4GB for $150 and more||18|
|Dell UltraSharp 32 8K embarrasses 4K monitors||47|
|EVGA readies a Hybrid Waterblock for Nvidia GP102 cards||5|
|Elgato Stream Deck lets streamers play news desk||7|
|Puppy Day Shortbread||27|
|I need this because of reasons.||+41|