by piotr5 » Fri May 29, 2015 9:47 pm
when I first learned about raspberry pi I experienced the same thing as you with parallella: not enough memory, problems with usb-connections to hardware, lack of ready-made hardware-components to plug in. now the pi has a camera and loads of comercially quite successful addons. so now I await the same to happen for epiphany on the software side.
as for fft, it's nonsense to use epiphany for that alone. even if I had a graphics-card instead, with simd to program, why would I want to restrict myself to using it for fft? for fft I'd feed it with data at a certain rate, and I'd get the result at the same speed. then it really doesn't matter beyond a certain amount of cores if the card has more power to offer, the transfer-speed is always the bottleneck, with every co-processor chip. and the fft-algorithm is so blazingly fast that with a handful of cores this bottleneck becomes relevant. fft is really better implemented in the main processor or fpga. a multi-core co-processor is supposed to get small compressed input, unpack it on demand, and then return again some small compressed data. any other way your application wont scale up well with rising number of cores. the approach parallella is using now is: create a lib, called pal, which runs on the epiphany chip, on each core, and which offers fft there! so, just like at the old times where fpu or mmx/sse were introduced, programmers are needed to make use of it instead of relying on the compiler to apply some limited set of algorithms when needed. in maths I've seen only one scaleable application for multi-core: partial sums/products in logarithmic time! (well, of course input still could be large, would need some compression, but output can be a single number!) but this doesn't need to be the only application, I'm curious to find other ideas...