• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

LucidLogix Fires up Multi-GPU Computing With Faster, More Flexible HYDRA 200 Chip

Joined
Aug 15, 2008
Messages
5,941 (1.02/day)
Location
Watauga, Texas
System Name Univac SLI Edition
Processor Intel Xeon 1650 V3 @ 4.2GHz
Motherboard eVGA X99 FTW K
Cooling EK Supremacy EVO, Swiftech MCP50x, Alphacool NeXXos UT60 360, Black Ice GTX 360
Memory 2x16GB Corsair Vengeance LPX 3000MHz
Video Card(s) Nvidia Titan X Tri-SLI w/ EK Blocks
Storage HyperX Predator 240GB PCI-E, Samsung 850 Pro 512GB
Display(s) Dell UltraSharp 34" Ultra-Wide (U3415W) / (Samsung 48" Curved 4k)
Case Phanteks Enthoo Pro M Acrylic Edition
Audio Device(s) Sound Blaster Z
Power Supply Thermaltake 1350watt Toughpower Modular
Mouse Logitech G502
Keyboard CODE 10 keyless MX Clears
Software Windows 10 Pro
It shouldnt be any less exciting than x58. And besides, if this chip does what its supposed to do, the p55 boards will rape.
 
Joined
Dec 8, 2008
Messages
1,334 (0.23/day)
Why would it be adding any latency. It potentially does the same thing as nvidia's chip.

Hydra receives orders from the CPU, splits them into workloads suitable for different GPU's, then send them to GPU's. When the GPU's finished their jobs they sent the results back to Hydra. Hydra combines them then sent it to the primary GPU for display. Why wouldn't this add latency?
Also, NF200 does not do the same thing.
 
Joined
Nov 10, 2008
Messages
1,989 (0.35/day)
Processor Intel Core i9 9900k @ 5.1GHZ all core load (8c 16t)
Motherboard MSI MEG Z390 ACE
Cooling Corsair H100i v2 240mm
Memory 32GB Corsair 3200mhz C16 (2x16GB)
Video Card(s) Powercolor RX 6900 XT Red Devil Ultimate (XTXH) @ 2.6ghz core, 2.1ghz mem
Storage 256GB WD Black NVME drive, 4TB across various SSDs/NVMEs, 4TB HDD
Display(s) Asus 32" PG32QUX (4k 144hz mini-LED backlit IPS with freesync & gsync & 1400 nit HDR)
Case Corsair 760T
Power Supply Corsair HX850i
Mouse Logitech G502 Lightspeed on powerplay mousemat
Keyboard Logitech G910
VR HMD Wireless Vive Pro & Valve knuckles
Software Windows 10 Pro
Hydra receives orders from the CPU, splits them into workloads suitable for different GPU's, then send them to GPU's. When the GPU's finished their jobs they sent the results back to Hydra. Hydra combines them then sent it to the primary GPU for display. Why wouldn't this add latency?
Also, NF200 does not do the same thing.

Current xfire/SLI:
1. Get draw commands
2. Split commands to both cards (if not doing AFR then dissect the image for the gpus to render their part).
3. Send data to both cards
4. Both cards render their part
5. Card 2 sends its data to card one for combining before sending to VDU.

Hydra:
1. Get draw commands
2. Detect different load capabilities on cards
3. Split commands to both cards (this will use a similar part to AFR i believe, so each card does a whole frame instead of tiled/split frames like super AA can do).
4. Send data to both cards
5. Both cards render their part
6. Card 2 sends its data to hydra which rediects it to card 1 (simple connection, not any latency) and the frame is interjected between frames generated by card 1 to the VDU (again very simple, no real latency).

With both of these setups the overall latency is going to be so close i'd say it will be indistinguishable. The hydra chip is meant to split the directX commands between the cards which are allowed to render the image using their own methods (if different). The image is then simply sent to be interjected between frames by card 1 - due to the hydra splitting the workload properly then you don't have to worry about rejoining parts of the same frame from different cards, or having to assume the cards run at a similar/same speed and having to sync the frames between the two properly (which is the cause of a lot of the overhead in current setups).

To split the data the hydra chip doesn't have to do too much work - once it knows the relative capabilities of each card it can simply direct the directx commands between them with no extra work needed - i.e. if card 1 is twice as fast as card 2 you just do:

Draw 1 -> card 1
Draw 2 -> card 1
Draw 3 -> card 2
Draw 4 -> card 1
Draw 5 -> card 1
Draw 6 -> card 2

Which isn't too expensive as you just direct the command down the apropriate pci-e connection.
 

Mussels

Freshwater Moderator
Staff member
Joined
Oct 6, 2004
Messages
58,413 (8.08/day)
Location
Oystralia
System Name Rainbow Sparkles (Power efficient, <350W gaming load)
Processor Ryzen R7 5800x3D (Undervolted, 4.45GHz all core)
Motherboard Asus x570-F (BIOS Modded)
Cooling Alphacool Apex UV - Alphacool Eisblock XPX Aurora + EK Quantum ARGB 3090 w/ active backplate
Memory 2x32GB DDR4 3600 Corsair Vengeance RGB @3866 C18-22-22-22-42 TRFC704 (1.4V Hynix MJR - SoC 1.15V)
Video Card(s) Galax RTX 3090 SG 24GB: Underclocked to 1700Mhz 0.750v (375W down to 250W))
Storage 2TB WD SN850 NVME + 1TB Sasmsung 970 Pro NVME + 1TB Intel 6000P NVME USB 3.2
Display(s) Phillips 32 32M1N5800A (4k144), LG 32" (4K60) | Gigabyte G32QC (2k165) | Phillips 328m6fjrmb (2K144)
Case Fractal Design R6
Audio Device(s) Logitech G560 | Corsair Void pro RGB |Blue Yeti mic
Power Supply Fractal Ion+ 2 860W (Platinum) (This thing is God-tier. Silent and TINY)
Mouse Logitech G Pro wireless + Steelseries Prisma XL
Keyboard Razer Huntsman TE ( Sexy white keycaps)
VR HMD Oculus Rift S + Quest 2
Software Windows 11 pro x64 (Yes, it's genuinely a good OS) OpenRGB - ditch the branded bloatware!
Benchmark Scores Nyooom.
what i'm interested in is how its actually done - if they dont use crossfire/SLI but use their own method, then we could start seeing ram being additive in multi GPU setups (EG, if they take half the screen each, but actually get to use all of the ram per card - 512MB card + 1024MB card stops dropping to 512MB, and gives us 1.5GB of usable ram)
 
Joined
Jan 2, 2008
Messages
3,296 (0.55/day)
System Name Thakk
Processor i7 6700k @ 4.5Ghz
Motherboard Gigabyte G1 Z170N ITX
Cooling H55 AIO
Memory 32GB DDR4 3100 c16
Video Card(s) Zotac RTX3080 Trinity
Storage Corsair Force GT 120GB SSD / Intel 250GB SSD / Samsung Pro 512 SSD / 3TB Seagate SV32
Display(s) Acer Predator X34 100hz IPS Gsync / HTC Vive
Case QBX
Audio Device(s) Realtek ALC1150 > Creative Gigaworks T40 > AKG Q701
Power Supply Corsair SF600
Mouse Logitech G900
Keyboard Ducky Shine TKL MX Blue + Vortex PBT Doubleshots
Software Windows 10 64bit
Benchmark Scores http://www.3dmark.com/fs/12108888
So will this still use "profiles" for efficient performance?
 
Joined
Nov 10, 2008
Messages
1,989 (0.35/day)
Processor Intel Core i9 9900k @ 5.1GHZ all core load (8c 16t)
Motherboard MSI MEG Z390 ACE
Cooling Corsair H100i v2 240mm
Memory 32GB Corsair 3200mhz C16 (2x16GB)
Video Card(s) Powercolor RX 6900 XT Red Devil Ultimate (XTXH) @ 2.6ghz core, 2.1ghz mem
Storage 256GB WD Black NVME drive, 4TB across various SSDs/NVMEs, 4TB HDD
Display(s) Asus 32" PG32QUX (4k 144hz mini-LED backlit IPS with freesync & gsync & 1400 nit HDR)
Case Corsair 760T
Power Supply Corsair HX850i
Mouse Logitech G502 Lightspeed on powerplay mousemat
Keyboard Logitech G910
VR HMD Wireless Vive Pro & Valve knuckles
Software Windows 10 Pro
what i'm interested in is how its actually done - if they dont use crossfire/SLI but use their own method, then we could start seeing ram being additive in multi GPU setups (EG, if they take half the screen each, but actually get to use all of the ram per card - 512MB card + 1024MB card stops dropping to 512MB, and gives us 1.5GB of usable ram)

From what i've read about the system it is basically an AFR rendering style that they adopt - this is how they overcome compatability issues between different card vendors and is why you can add any combination of cards without worrying about overhead. Looking at how they described the system i doubt the gram will be additive in this implementation, but i do think that the cards would be able to use all of their own ram - even if you had a 512mb and a 2gb card the 2gb can still use of of its ram (the system basically makes the cards act as if they were in a single card config).

I may be understanding their implementation wrong of course but from what i've read this is how it will be working.
 
Joined
May 20, 2004
Messages
10,487 (1.42/day)
what i'm interested in is how its actually done - if they dont use crossfire/SLI but use their own method, then we could start seeing ram being additive in multi GPU setups (EG, if they take half the screen each, but actually get to use all of the ram per card - 512MB card + 1024MB card stops dropping to 512MB, and gives us 1.5GB of usable ram)

This is a second generation chip. There are plenty of demos out there for the first, google it. They use their own method, that's the whole point.
 
Joined
Mar 11, 2009
Messages
865 (0.15/day)
Location
Dawn
Forgot about that, but P55 multiple GPU's just doesn't get me all excited :(
What I'd like to see, not so much because of the Lucid Logix chips, but more over pcie 3.0, usb 3.0, and sata 3.0, is a refresh of the x58 motherboards. If manufacturers could get rid of the stocks they have and make new ones with those updated interfaces, and hopefully lucid logix chips (if they work as advertised) That would be great.
 

RallyRoach

New Member
Joined
Dec 31, 2005
Messages
11 (0.00/day)
So, if I'm understanding this correctly, say we have 30 frames to render in a second, and a GTX 295 and a 4850 installed on a system with this chip. The GTX is roughly twice as fast as the 4850 on a given game/setting. So for every 1 frame sent to the 4850, the GTX gets 2? The 4850 renders 10 frames, while the GTX 295 renders 20 frames, all in the same second and the Hydra chip just outputs this as 30 frames per second?
 

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
46,792 (7.63/day)
Location
Hyderabad, India
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard ASUS ROG Strix B450-E Gaming
Cooling DeepCool Gammax L240 V2
Memory 2x 8GB G.Skill Sniper X
Video Card(s) Palit GeForce RTX 2080 SUPER GameRock
Storage Western Digital Black NVMe 512GB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
Joined
Aug 10, 2007
Messages
4,267 (0.69/day)
Location
Sanford, FL, USA
Processor Intel i5-6600
Motherboard ASRock H170M-ITX
Cooling Cooler Master Geminii S524
Memory G.Skill DDR4-2133 16GB (8GB x 2)
Video Card(s) Gigabyte R9-380X 4GB
Storage Samsung 950 EVO 250GB (mSATA)
Display(s) LG 29UM69G-B 2560x1080 IPS
Case Lian Li PC-Q25
Audio Device(s) Realtek ALC892
Power Supply Seasonic SS-460FL2
Mouse Logitech G700s
Keyboard Logitech G110
Software Windows 10 Pro


Showing output of one GPU on the right, completed frame on the left.


Instead of a brute method of splitting work, the scene is intelligently broken down. Different amounts of RAM per card will work because textures don't need to be in memory on all cards.
 
Joined
Aug 15, 2008
Messages
5,941 (1.02/day)
Location
Watauga, Texas
System Name Univac SLI Edition
Processor Intel Xeon 1650 V3 @ 4.2GHz
Motherboard eVGA X99 FTW K
Cooling EK Supremacy EVO, Swiftech MCP50x, Alphacool NeXXos UT60 360, Black Ice GTX 360
Memory 2x16GB Corsair Vengeance LPX 3000MHz
Video Card(s) Nvidia Titan X Tri-SLI w/ EK Blocks
Storage HyperX Predator 240GB PCI-E, Samsung 850 Pro 512GB
Display(s) Dell UltraSharp 34" Ultra-Wide (U3415W) / (Samsung 48" Curved 4k)
Case Phanteks Enthoo Pro M Acrylic Edition
Audio Device(s) Sound Blaster Z
Power Supply Thermaltake 1350watt Toughpower Modular
Mouse Logitech G502
Keyboard CODE 10 keyless MX Clears
Software Windows 10 Pro
THe problem though is still within drivers of multi vendors. Putting ATI and Nvidia together on the same platform and making them work together at a driver level is almost impossible.
 

Mussels

Freshwater Moderator
Staff member
Joined
Oct 6, 2004
Messages
58,413 (8.08/day)
Location
Oystralia
System Name Rainbow Sparkles (Power efficient, <350W gaming load)
Processor Ryzen R7 5800x3D (Undervolted, 4.45GHz all core)
Motherboard Asus x570-F (BIOS Modded)
Cooling Alphacool Apex UV - Alphacool Eisblock XPX Aurora + EK Quantum ARGB 3090 w/ active backplate
Memory 2x32GB DDR4 3600 Corsair Vengeance RGB @3866 C18-22-22-22-42 TRFC704 (1.4V Hynix MJR - SoC 1.15V)
Video Card(s) Galax RTX 3090 SG 24GB: Underclocked to 1700Mhz 0.750v (375W down to 250W))
Storage 2TB WD SN850 NVME + 1TB Sasmsung 970 Pro NVME + 1TB Intel 6000P NVME USB 3.2
Display(s) Phillips 32 32M1N5800A (4k144), LG 32" (4K60) | Gigabyte G32QC (2k165) | Phillips 328m6fjrmb (2K144)
Case Fractal Design R6
Audio Device(s) Logitech G560 | Corsair Void pro RGB |Blue Yeti mic
Power Supply Fractal Ion+ 2 860W (Platinum) (This thing is God-tier. Silent and TINY)
Mouse Logitech G Pro wireless + Steelseries Prisma XL
Keyboard Razer Huntsman TE ( Sexy white keycaps)
VR HMD Oculus Rift S + Quest 2
Software Windows 11 pro x64 (Yes, it's genuinely a good OS) OpenRGB - ditch the branded bloatware!
Benchmark Scores Nyooom.
THe problem though is still within drivers of multi vendors. Putting ATI and Nvidia together on the same platform and making them work together at a driver level is almost impossible.

true, as well as directX levels.

you cant mix a DX11 and a DX10 card and expect to run DX11 games, for example.


as for the rest of it, it gives performance boosts (each card can use all its ram, hence its 'additive' - poor word, but you get the meaning) and it lets you say, crossfire a 3870 with a 4870
 
Top