
on 11.08.2009 at 21:52 joel wrote :
Best approach is always do, show, convince. I was burned with my own Boost.SIMD proposal. It's better to work first THEN show off.
For my own work, see :
http://www.lri.fr/~falcou/pub/falcou-ACIVS-2004.pdf http://www.springerlink.com/content/l4r4462r25740127/
and the old page at sourceforge : http://nt2.sourceforge.net
I had a bit of discussion in the ML before, with comaprison w/r to Eigen
i'll run through that links for now listen as i can see you are familiar to simd programming so when i researched the expression template based impementation of vector operations i tried to use sse2 though compiler intrinsics (since they are de facto portable) but i did not get any benefit for doubles i think that's because load/store operations consumed boost from actual add_pd's etc. so evaluating something like (pairwise for sse2) vec + vec*scalar - vec + element_wise_mul(vec, vec) yields the same time for sse2 and plain implementations however for floats i got significant boost (~2 however not 4) so since i prefer doubles i droped such simd optimizations (but they are still easily possible) i'm interested about what you can say on this -- Pavel