Question about the:
[ul][li] ATI Radeon HD5970 Eyefinity 12[*] ATI Radeon HD5870 Eyefinity 6[/ul]Anybody worked with the latter in OpenGL? [/li]
Nearly all the material I’ve found is marketing fluff, but best I can tell they allocate one ginormous framebuffer and then just partition the output to separate display outputs (basically, you just crank up your FOV and resolution, and pretend that it’s all-the-same to the GPU). AFAICT, there’s just one clump of “GPU memory” shared by all displays, one clump of GPU cores processing the vertex data going to all displays, one clump of GPU cores doing fragment processing and downsample for all displays, and one GPU command stream, not 6 (or 12) possibly driven by different CPU cores in parallel.
Is this correct?
If so, then seens this setup is interesting for simple WoW-style games. But if your app really pushes the GPU (lots of verts, lots of fill, lots of view-dependent texture, high AA, different projections per display, etc.) then it seems it would bring this setup to its knees.
That said, this is what I’m inferring. I’d much appreciate some hard facts from some developer that knows details.