Re: 580Gbits/s Bus
First of all, thank you very much for your reply. It is very informative and I understand the limitations better.
>>
Let us know what exactly that it is you are looking for so that we may have better solutions for you.
My intent is to use a graphics card to perform fast, parallelized data processing. I chose nVidia because they have a library maintained internally for GPGPU.
However, this data needs to be output to an external device. However the implementation, the external device needs to receive 580Gbit/s constantly.
My issue is how to output the data at such high data rates. I must use some sort of distributed computing. But, at this point, even if option 1 (output from graphics card to DVI) does work, it will take 580Gbit / 14.8Gbit = 40 cards (and many PCs)
At this point, I am unsure of what device can
receive data at these speeds. But if I can at least send the data that fast, I can most likely hire a company to design a custom ASIC to receive the data and process it.
-----------------------------------------------------------------
My mistake, I meant the
Geforce 8800 Ultra.
-The bandwidth of the Geforce will be 86.4GB/sec * 8 bits = 691.2Gbit/sec.
Option 1: 2xDVI Dual Mode --> 7.4 * 2 = 14.8Gbits/sec
Option 2: Transferring to host memory, it would still be limited by the PCI-E 2.0 bus (5GBits * 16 lanes * 8bits/10bits_encoded) = 64Gbit/sec
However, the actual transfer speed is dependent on the implementation.
Also, the CPU would be overloaded from all the copying...which...will probably slow things down a lot.
-----------------------------------------------------------------
I know that option 2 is not the best approach. But it seems like that will provide a higher throughput. I have been looking into 10Gb fiber optic NICs. The reason for that is because:
-10Gb fiber NICs are readily available
-10Gb fiber transceivers are available for the Xilinx Virtex family
-This at least addresses both sending and receiving data at such high speeds. Plus, I could avoid needing an ASIC device and store all processing in the FPGA.
-I also just discovered a 10Gb card that supports RDMA (Remote Direct Memory Access
https://en.wikipedia.org/wiki/Remote_Direct_Memory_Access), which I don't know if applicable, but I will look into it. But, if it is, then the CPU being overworked shouldn't be an issue.
###OR###
Dual 20Gbit Infiniband NICs (40Gbit). But I haven't discovered a transceiver that is compatible with FPGAs. (580 / 40 = 15 NICs)
-----------------------------------------------------------------
Eventually when 100Gbit NICs come out, it will be much easier to implement the above. But, this is the best solution that I could come up with.
Any ideas on other implementations?