by theover » Fri May 10, 2013 12:18 pm
I recall having looked into some of the literature as a student (the universities DSP teacher at the time was rather a mess, and didn't do anything relevant himself), and there are various considerations to think about when making an FFT implementation, like accuracy of the repeated butterfly, and the reordering being done right as in not to throw away required sub-computations. Of course reusing the results of the previous computations makes the computations complexity go from something like square of the size to size times log(size), which is possibly lost when going parallel without reusing results...
@Yaniv: It's just that I found it marvelous I could use my graphics card for in priciple "plug-ing" in the fftw3 library, which the Cuda makers claim to be available in the same form from their libs. I don't know how much I'd need it and didn't try out making "Jamin" with graphics card acceleration, even though some of the specs of my humble GT640 Kepler card are probably reason enough to use the power of it. It's a nice thing to built a powerful, working Linux system from parts, getting Fedora on it, and using the motherboard great 192/24 audio converters to get studio quality processing, but it's a big machine, and many in Open Source appear to want to play as much "divide and conquer" as say Microsoft.
The idea of using the low power Parallella for the same purpose of the whole 100W Linux machine (actually less because of the platinum plus fanless supply), AND having the FPGA connect to for instance custom converters, meters, mixer control etc, is appealing.
T.