Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×

Comment Questions answered (Score 2) 63

Just like SIMD operations on a CPU, GPUs are typically designed to execute the same instructions in parallel over a set of data. Even if the hardware algorithm is slower, parallelizing across a data stream can end up being much faster than executing it faster one at a time especially since hardware buses perform better with bursts of data.

The advantages of FPGAs and custom/special purpose ASICs is that you can choose to optimize less for generality and more for handling specialized tasks. If you have enough transistors/LUTs you can utilize more and more of them to optimize and reduce how many clock cycles your algorithm executes all the way until it takes only a single clock cycle (as long as you're not bumping in to path planning length limitations). FPGA speeds are typically less than 800Mhz so even single cycle operations can't get any faster than your FPGAs maximum speed whereas an ASIC can be designed to run at much higher clock speeds.

ASICs are typically simulated on fairly beefy FPGAs, often several working in concert, before being produced so calculating speed is obviously doable.

Slashdot Top Deals

"When it comes to humility, I'm the greatest." -- Bullwinkle Moose

Working...