Wednesday, September 23rd 2009

LucidLogix Fires up Multi-GPU Computing With Faster, More Flexible HYDRA 200 Chip

LucidLogix (Lucid) today introduced the HYDRA 200 real time distributed processing engine designed to bring multi-GPU computing to the masses.

For the first time ever, motherboard, graphics card manufacturers and users can have the flexibility to use different combinations of GPUs from AMD (ATI) and NVIDIA in notebooks and PCs. The solution delivers faster 3D graphics at consumer price points.

The new HYDRA 200 SoC is Lucid's second generation parallel graphics processor that works with any GPU, CPU or chipset to provide scalable 3D graphics performance in a multi-GPU computing environment. On display for the first time at IDF 2009 (booth 213) HYDRA 200 is faster, more flexible, smaller and more power-efficient than its predecessor silicon, the HYDRA 100.
"We've further refined our HYDRA engine and made it faster and more flexible, allowing for a near limitless combination of GPU's," said Lucid vice president of research and development, David Belz. "HYDRA 200 allows the consumer to get more 'bang for their GPU buck' by extending the life of their current GPU investment, providing even faster graphics performance and later upgrading their system with whatever card they choose." Until now, multi-GPU systems have been graphics vendor specific and generally require the consumer to be fairly technically savvy. With Lucid HYDRA 200, OEMs can offer custom configurations at different price/performance targets, and consumers will be able to easily add graphics hardware to achieve an overall performance boost without the worry of compatibility.

Gamers with a need for speed now have a solution that's optimized for their performance requirements by allowing them to choose the right chip for the job, or simply upgrade without throwing away the old one. HYDRA 200 currently supports Windows Vista and Windows 7 operating systems as well as DirectX 9.0c and 10.1 standard APIs and is DirectX 11 ready.

HYDRA 200 Tech Specs
The adaptive and dynamic parallel graphics load-balancing scheme of HYDRA 200 resolves bottlenecks in the system and optimizes the usage of GPU resources with minimal power consumption. HYDRA 200 is a 65nm PCIe compatible SoC that also features:
  • Low power use of under 6W, making it ideal for graphic cards, notebooks or desktops as there is no need for a special heatsink
  • A small footprint (18-22mm) that allows for a compact design
  • GPU connector free, making it easy to integrate into systems
  • Supports multiple display configurations
  • Universal GPU, CPU and chipset support
  • For systems using dual, tri or quad GPU combinations
"I cannot say the graphics wars are over, but technology demonstrated by Lucid clearly levels the playing field when it comes to GPU scaling," said Dr. Jon G Peddie. "And with multi-core graphics growing at a rate of ten percent year over year, it will be interesting to see what kind of momentum a mass market M GPU can have in the marketplace."

HYDRA 200 Availability
With universal GPU support and a variety of configurations, the HYDRA 200 makes system integration swift and worry-free. HYDRA 200 is available now for reference designs in three models:
Source: LucidLogix
Add your own comment

38 Comments on LucidLogix Fires up Multi-GPU Computing With Faster, More Flexible HYDRA 200 Chip

#26
pr0n Inspector
KenshaiWhy would it be adding any latency. It potentially does the same thing as nvidia's chip.
Hydra receives orders from the CPU, splits them into workloads suitable for different GPU's, then send them to GPU's. When the GPU's finished their jobs they sent the results back to Hydra. Hydra combines them then sent it to the primary GPU for display. Why wouldn't this add latency?
Also, NF200 does not do the same thing.
Posted on Reply
#27
human_error
pr0n InspectorHydra receives orders from the CPU, splits them into workloads suitable for different GPU's, then send them to GPU's. When the GPU's finished their jobs they sent the results back to Hydra. Hydra combines them then sent it to the primary GPU for display. Why wouldn't this add latency?
Also, NF200 does not do the same thing.
Current xfire/SLI:
1. Get draw commands
2. Split commands to both cards (if not doing AFR then dissect the image for the gpus to render their part).
3. Send data to both cards
4. Both cards render their part
5. Card 2 sends its data to card one for combining before sending to VDU.

Hydra:
1. Get draw commands
2. Detect different load capabilities on cards
3. Split commands to both cards (this will use a similar part to AFR i believe, so each card does a whole frame instead of tiled/split frames like super AA can do).
4. Send data to both cards
5. Both cards render their part
6. Card 2 sends its data to hydra which rediects it to card 1 (simple connection, not any latency) and the frame is interjected between frames generated by card 1 to the VDU (again very simple, no real latency).

With both of these setups the overall latency is going to be so close i'd say it will be indistinguishable. The hydra chip is meant to split the directX commands between the cards which are allowed to render the image using their own methods (if different). The image is then simply sent to be interjected between frames by card 1 - due to the hydra splitting the workload properly then you don't have to worry about rejoining parts of the same frame from different cards, or having to assume the cards run at a similar/same speed and having to sync the frames between the two properly (which is the cause of a lot of the overhead in current setups).

To split the data the hydra chip doesn't have to do too much work - once it knows the relative capabilities of each card it can simply direct the directx commands between them with no extra work needed - i.e. if card 1 is twice as fast as card 2 you just do:

Draw 1 -> card 1
Draw 2 -> card 1
Draw 3 -> card 2
Draw 4 -> card 1
Draw 5 -> card 1
Draw 6 -> card 2

Which isn't too expensive as you just direct the command down the apropriate pci-e connection.
Posted on Reply
#28
Deleted member 3
DaMultaI wonder when Nvidia will buy them out......
I doubt they will, Intel is a big investor, I see them buying it before nv gets a chance.
Posted on Reply
#29
Mussels
Freshwater Moderator
what i'm interested in is how its actually done - if they dont use crossfire/SLI but use their own method, then we could start seeing ram being additive in multi GPU setups (EG, if they take half the screen each, but actually get to use all of the ram per card - 512MB card + 1024MB card stops dropping to 512MB, and gives us 1.5GB of usable ram)
Posted on Reply
#31
human_error
Musselswhat i'm interested in is how its actually done - if they dont use crossfire/SLI but use their own method, then we could start seeing ram being additive in multi GPU setups (EG, if they take half the screen each, but actually get to use all of the ram per card - 512MB card + 1024MB card stops dropping to 512MB, and gives us 1.5GB of usable ram)
From what i've read about the system it is basically an AFR rendering style that they adopt - this is how they overcome compatability issues between different card vendors and is why you can add any combination of cards without worrying about overhead. Looking at how they described the system i doubt the gram will be additive in this implementation, but i do think that the cards would be able to use all of their own ram - even if you had a 512mb and a 2gb card the 2gb can still use of of its ram (the system basically makes the cards act as if they were in a single card config).

I may be understanding their implementation wrong of course but from what i've read this is how it will be working.
Posted on Reply
#32
Deleted member 3
Musselswhat i'm interested in is how its actually done - if they dont use crossfire/SLI but use their own method, then we could start seeing ram being additive in multi GPU setups (EG, if they take half the screen each, but actually get to use all of the ram per card - 512MB card + 1024MB card stops dropping to 512MB, and gives us 1.5GB of usable ram)
This is a second generation chip. There are plenty of demos out there for the first, google it. They use their own method, that's the whole point.
Posted on Reply
#33
theubersmurf
1Kurgan1Forgot about that, but P55 multiple GPU's just doesn't get me all excited :(
What I'd like to see, not so much because of the Lucid Logix chips, but more over pcie 3.0, usb 3.0, and sata 3.0, is a refresh of the x58 motherboards. If manufacturers could get rid of the stocks they have and make new ones with those updated interfaces, and hopefully lucid logix chips (if they work as advertised) That would be great.
Posted on Reply
#34
RallyRoach
So, if I'm understanding this correctly, say we have 30 frames to render in a second, and a GTX 295 and a 4850 installed on a system with this chip. The GTX is roughly twice as fast as the 4850 on a given game/setting. So for every 1 frame sent to the 4850, the GTX gets 2? The 4850 renders 10 frames, while the GTX 295 renders 20 frames, all in the same second and the Hydra chip just outputs this as 30 frames per second?
Posted on Reply
#36
Disparia


Showing output of one GPU on the right, completed frame on the left.


Instead of a brute method of splitting work, the scene is intelligently broken down. Different amounts of RAM per card will work because textures don't need to be in memory on all cards.
Posted on Reply
#37
PP Mguire
THe problem though is still within drivers of multi vendors. Putting ATI and Nvidia together on the same platform and making them work together at a driver level is almost impossible.
Posted on Reply
#38
Mussels
Freshwater Moderator
PP MguireTHe problem though is still within drivers of multi vendors. Putting ATI and Nvidia together on the same platform and making them work together at a driver level is almost impossible.
true, as well as directX levels.

you cant mix a DX11 and a DX10 card and expect to run DX11 games, for example.


as for the rest of it, it gives performance boosts (each card can use all its ram, hence its 'additive' - poor word, but you get the meaning) and it lets you say, crossfire a 3870 with a 4870
Posted on Reply
Add your own comment
Apr 25th, 2024 18:06 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts