When we first reviewed the GeForce GTX 680, we were in our usual rush to push out an article at the time of the product's unveiling, after a solid week of poking and prodding Nvidia's new GPU. We were also attempting some of the more complex analysis of real-time graphics performance we'd ever done, and these things take time to produce.
Something had to give.
Our choice was to take a mulligan on writing up several of the new hardware and software features built into the GTX 680 and its GK104 graphics processor, with the intention of fixing the oversight later. We figured omitting some stuff and patching it later was pretty fitting for a GPU introduction, anyhow.
Well, we've been working on writing up those features, and given additional time to describe them, we indulged our penchant for ridiculous detail. The word count shot up like Apple's stock price, and the result is this article, because there's no way we're not totally milking this thing for more readership. Read on for more detail than you probably deserve to have inflicted on you concerning the GeForce GTX 680's new dynamic clocking scheme, antialiasing capabilities, adaptive vsync, and more.
Nvidia silicon gets forced induction
As evidenced by the various "turbo" schemes in desktop CPUs, dynamic voltage and frequency tech is all the rage these days. The theory is straightforward enough. Not all games and other graphics workloads make use of the GPU in the same way, and even relatively "intensive" games may not cause all of the transistors to flip and thus heat up the GPU quite like the most extreme cases. As a result, there's often some headroom left in a graphics card's designated thermal envelope, or TDP (thermal design power), which is generally engineered to withstand a worst-case peak workload. Dynamic clocking schemes attempt to track this headroom and to take advantage of it by raising clock speeds opportunistically.
Although the theory is fairly simple, the various implementations of dynamic clocking vary widely in their specifics, which can make them hard to sort. Intel's Turbo Boost is probably the gold standard at present; it uses a network of thermal sensors spread across the die in conjunction with a programmable, on-chip microcontroller that governs Turbo policy. Since it's a hardware solution with direct inputs from the die, Turbo Boost reacts very quickly to changes in thermal conditions, and its behavior may differ somewhat from chip to chip, since the thermal properties of the chips themselves can vary.
Although distinct from one another in certain ways, both AMD's Turbo Core (in its CPUs) and PowerTune (in its GPUs) combine on-chip activity counters with pre-production chip testing to establish a profile for each model of CPU or GPU. In operation, then, power draw for the chip is estimated based on the activity counters, and clocks are adjusted in response to the expected thermal situation. AMD argues the predictable, deterministic behavior of its dynamic voltage and frequency (DVFS) schemes is an admirable trait. The price of that consistency is that it can't squeeze every last drop of performance out of each individual slab of silicon.
Nvidia's take on this sort of capability is called GPU Boost, and it combines some traits of each of the competing schemes. Fundamentally, its logic is more like the two Turbos than it is like AMD's PowerTune. With PowerTune, AMD runs its GPUs at a relatively high base frequency, but clock speeds are sometimes throttled back during atypically high GPU utilization. By contrast, GPU Boost starts with a more conservative base clock speed and ranges into higher frequencies when possible.
The inputs for Boost's decision-making algorithm include power draw, GPU and memory utilization, and GPU temperatures. Most of this information is collected from the GPU itself, but I believe the power use information comes from external circuitry on the GTX 680 board. In fact, Nvidia's Tom Petersen told us board makers will be required to include this circuitry in order to get the GPU maker's stamp of approval. The various inputs for Boost are then processed in software, in a portion of the GPU driver, not in an on-chip microcontroller. This combination of software control and external power circuitry is likely responsible for Boost's relatively high clock-change latency. Stepping up or down in frequency takes about 100 milliseconds, according to Petersen. A tenth of a second is a very long time in the life of a gigahertz-class chip, and Petersen was frank in admitting that this first generation of GPU Boost isn't everything Nvidia hopes it will become in the future.
Graphics cards with Boost will be sold with a couple of clock speed numbers on the side. The base clock is the lower of the two—1006MHz on the GeForce GTX 680—and represents the lowest operating speed in thermally intensive workloads. Curiously enough, the "boost clock"—which is 1058MHz on the GTX 680—isn't the maximum speed possible. Instead, it's "sort of a promise," says Petersen; it's the clock speed at which the GPU should run during typical operation. GPU Boost performance will vary slightly from card to card, based on factors like chip quality, ambient temperatures, and the effectiveness of the cooling solution. GeForce GTX 680 owners should expect to see their cards running at the Boost clock frequency fairly regularly, regardless of these factors. Beyond that, GPU Boost will make its best effort to reach even higher clock speeds when feasible, stepping up and down in increments of 13MHz.
Petersen demoed several interesting scenarios to illustrate Boost behavior. In a very power-intensive scene, 3DMark11's first graphics test, the GTX 680 was forced to remain at its base clock throughout. When playing Battlefield 3, meanwhile, the chip spent most of its time at about 1.1GHz—above both the base and boost levels. In a third application, the classic DX9 graphics demo "rthdribl," the GTX throttled back to under 1GHz, simply because additional GPU performance wasn't needed. One spot where Nvidia intends to make use of this throttling capability is in-game menu screens—and we're happy to see it. Some menu screens can cause power use and fan speeds to shoot skyward as frame rates reach quadruple digits.
Nvidia has taken pains to ensure GPU Boost is compatible with user-driven tweaking and overclocking. A new version of its NVAPI allows third-party software, like EVGA's slick Precision software, control over key Boost parameters. With Precision, the user may raise the GPU's maximum power limit by as much as 32% above the default, in order to enable operation at higher clock speeds. Interestingly enough, Petersen said Nvidia doesn't consider cranking up this slider overclocking, since its GPUs are qualified to work properly at every voltage-and-frequency point along the curve. (Of course, you could exceed the bounds of the PCIe power connector specification by cranking this slider, so it's not exactly 100% kosher.) True overclocking happens by grabbing hold of a separate slider, the GPU clock offset, which raises the chip's frequency at a given voltage level. An offset of +200MHz, for instance, raised our GTX 680's clock speed while running Skyrim from 1110MHz (its usual Boost speed) to 1306MHz. EVGA's tool allows GPU clock offsets as high as +549MHz and memory clock offsets up to +1000MHz, so users are given quite a bit of room for experimentation.
Although GPU Boost is only in its first incarnation, Nvidia has some big ideas about how to take advantage of these dynamic clocking capabilities. For instance, Petersen openly telegraphed the firm's plans for future versions of Boost to include control over memory speeds, as well as GPU clocks.
More immediately, one feature exposed by EVGA's Precision utility is frame-rate targeting. Very simply, the user is able to specify his desired frame rate with a slider, and if the game's performance exceeds that limit, the GPU steps back down the voltage-and-frequency curve in order to conserve power.
We were initially skeptical about the usefulness of this feature for one big reason: the very long latency of 100 ms for clock speed adjustments. If the GPU has dialed back its speed because the workload is light and then something changes in the game—say, an explosion that adds a bunch of smoke and particle effects to the mix—ramping the clock back up could take quite a while, causing a perceptible hitch in the action. We think that potential is there, and as a result, we doubt this feature will appeal to twitch gamers and the like.
However, in our initial playtesting of this feature, we've not noticed any problems. We need to spend more time with it, but Kepler's frame rate targeting may prove to be useful, even in this generation, so long as its clock speed leeway isn't too wide. At some point in the future, when the GPU's DVFS logic is moved into hardware and frequency change delays are measured in much smaller numbers, we expect features like this one to become standard procedure, especially for mobile systems.
|MSI's Z87-GD65 Gaming motherboard reviewed||23|
|Microsoft backtracks on Xbox One disc sharing, connection limitations||60|
|Here's about 10 minutes of Thief gameplay||6|
|LG says mass production of flexible displays will begin this year||32|
|Steam beta file hints at game sharing feature||28|
|Asus mulling wearable devices||14|
|Nvidia to license Kepler GPU core to Android device makers||66|
|Refuted: BF4, other Frostbite 3 games to be 'optimized exclusively for AMD'||185|