Thanks to a paper Intel is presenting at this year's Siggraph conference, we now know more about Intel's in-development graphics processor and do-everything data-parallel computing engine known as Larrabee. Dr. Larry Seiler, Senior Principal Engineer on the Larrabee project, and a number of his co-authors provided an overview of the paper to the press late last week, and in the process, they revealed quite a bit about how Larrabee will work and what Intel's design philosophy looks like. We still don't have enough information to piece together any performance projections, but Intel has detailed quite a bit about the basic building blocks of Larrabee's architecture and about its approach to rendering.
Sadly to say, I was working on write-up about this new Larrabee info, but I just wasn't able to get it done last night. Sometimes in life, you just have to punt, and for various reasons, today is one of those times. But you should not miss out on the Larrabee goods if you're into this stuff. Let me suggest reading the Larrabee write-up by the guys over at AnandTech for more detail on the chip's inner workings. The highlights include:
This Larrabee info is fascinating, because it makes for head-spinning potential in many areas and yet... one could easily see the first products falling well short of existing GPUs in terms of performance and area efficiency in most games and other graphics apps. Custom hardware is awfully difficult to beat.
One potential positive of Larrabee's fully coherent memory subsystem is the possibility of much more efficient multi-chip implementations. Nvidia and AMD are essentially managing the coherency problem manually via custom game profiles for multi-GPU setups right now. When I asked about this issue, Intel said it didn't expect to have the same pain as its competitors in this area.
The possible downsides of all-software implementations of things like the render back-end are also rather apparent. We saw this illustrated nicely when the Radeon HD 4800 series brought a vast improvement over the shader-based MSAA resolve used in the Radeon HD 2900 and 3800 series products.
Somewhat comically, the initial reactions to the Larrabee architecture disclosures from Intel are mixed along clear dividing lines. When I pinged David Kanter, the CPU guru at Real World Tech, about it, he was very much impressed with the choices Intel had made and generally positive about the prospects. Meanwhile, Rys over at Beyond3D expressed quite a bit of skepticism about the chip's likely performance and area efficiency for graphics versus more traditional architectures.
Handicapping Larrabee's performance is very difficult right now for many reasons, including the fact that we still don't know some key specifications: how many cores Larrabee chips will have, how the memory interfaces will look, the number and capacity of the custom texture units, and what sort of clock speeds to expect. We also don't yet know exactly how Intel will be extending the x86 instruction set for graphics. We should have more answers as Larrabee's release approaches. What we do know for sure is that things are about to get a lot more interesting.
|A first look at USB 3.1 performance||15|
|People want thicker smartphones with bigger batteries; Apple doesn't||79|
|FREAK vulnerability can affect Windows, as well||6|
|Deal of the week: A G-Sync display, Shield Tablet, and more||28|
|Wolfenstein: The Old Blood trailer is delightfully pulpy||50|
|Thursday Night Shortbread||37|
|We learned more about Vulkan at Valve's glNext presentation||78|
|Steam Controller gets November release, $50 price tag||25|
|And Samsung makes new phone with no sd slot lol whaw whaw whaw||+59|