Games have included basic object physics for years, but real-time fluid simulations have remained elusive. Sure, we've seen some tech demos of tentative approaches, but simulated liquids have always looked more like jello or tapioca pudding than actual water—close, but no cigar.
Well, hold on to your hats, because Nvidia staffers have come up with a new technique to simulate fluids using PhysX, and the results are pretty true-to-life. See for yourself:
The demo is the result of research by Miles Macklin and Matthias Müller, who have published a paper (PDF) detailing their work. The paper will be presented at SIGGRAPH 2013 in a couple of months. Here's the abstract:
In ﬂuid simulation, enforcing incompressibility is crucial for realism; it is also computationally expensive. Recent work has improved efﬁciency, but still requires time-steps that are impractical real-time applications. In this work we present an iterative density solver integrated into the Position Based Dynamics framework (PBD). By formulating and solving a set of positional constraints that enforce constant density, our method allows similar incompressibility and convergence to modern smoothed particle hydrodynamic (SPH) solvers, but inherits the stability of the geometric, position based dynamics method, allowing large time steps suitable for real-time applications. We incorporate an artificial pressure term that improves particle distribution, creates surface tension, and lowers the neighborhood requirements of traditional SPH. Finally, we address the issue of energy loss by applying vorticity confinement as a velocity post process.
Yep. Uh-huh. Totally understood every word of that.
The lighthouse simulation is shown running at 15 FPS on a GeForce GTX 680, which isn't outlandishly fast by today's standards. (It's over a year old.) Of course, aside from the water, the 3D scene being rendered isn't that complex. It may take substantially more horsepower to implement the same simulation in a modern game.
Speaking of which, I hope game developers will be able to use the algorithms described in the paper to do similar work using a cross-vendor API—say, DirectCompute. There's nothing wrong with PhysX, but limiting this kind of technology to a single GPU vendor may reduce the likelihood of a major game using it in a meaningful way.
|1. GKey13 - $650||2. JohnC - $600||3. davidbowser - $501|
|4. cmpxchg - $500||5. DeadOfKnight - $400||6. danny e. - $375|
|7. the - $360||8. Ryszard - $351||9. rbattle - $350|
|10. Ryu Connor - $350|
|Here's the not-so-live video version of The TR Podcast 164||16|
|Here's what's cooking in Damage Labs||34|
|Deal of the week: An IPS ultra-wide for $420, plus cheap SSDs and more||29|
|Microsoft's quarterly revenue up 25% on strong Surface, Xbox sales||23|
|Assassin's Creed Unity PC requires 6GB of RAM, GTX 680||235|
|Join us as we attempt to live stream The TR Podcast tonight||13|
|Civ: Beyond Earth with Mantle aims to end multi-GPU microstuttering||75|
|CPU startup claims to achieve 3x IPC gains with VISC architecture||62|
|I just found this AMAZING trick! Call of Duty takes up 0GB if you just don't buy it!||+122|