Uhm? The x86 output is perfectly decent, as long as it's not using openMP - slightly slower than the best gcc results, but very seldom anything you'd actually notice in a real-life situation.
Originally Posted by allquixotic
I think these results are actually quite impressive. Sure, it isn't matching the performance of dedicated hardware, but did anyone seriously expect that at this point? The gap between the CPU and GPU is remarkably small considering that the CPU is a fully generic processor.
I am particularly interested in what this might hold for the future. The AVX2 instruction set extension has four times the floating-point vector throughput, two times the integer vector throughput, and gather instructions which replace 18 legacy instructions!
So what are the LLVMpipe developers' thoughts on the future of real-time CPU rendering?
There isn't really an llvmpipe roadmap, nobody is really pushing features on it at the moment, vmware seem to be using it but no new major speedups have shown up.
Originally Posted by smitty3268
I was doing GL3 support as a spare time project, but my spare time decided it would rather do something else, so I might get back to it eventually,
The kernel stuff was only for making texture-from-pixmap faster so gnome-shell can go faster, it doesn't make llvmpipe itself go faster at all.
This would interest me too. I always read here how the LLVMpipe is better than the old softpipe in using all cores and new CPU instructions. But I never see that in real usage. If I kill direct rendering (e.g. by chmod -w /dev/dri/card*) there should be SW rendering. I only compile LLVMpipe and R600 (both gallium3D) so I think the SW rendering should be using LLVMpipe. But no 3D program is shown in top as using more threads to use all cores (and the programs are rendering slowly).
Originally Posted by sandy8925
What am I doing wrong? Some bad arguments to Mesa build? Or is this only starting at some specific LLVM version?
glxinfo will say which driver you have, and you can force softpipe with the GALLIUM_DRIVER=softpipe var IIRC.
I seem to recall there was a var to limit llvmpipe's threads too, but not sure on that.
OK, so I have this:
Originally Posted by curaga
OpenGL vendor string: VMware, Inc.
OpenGL renderer string: Gallium 0.4 on llvmpipe (LLVM 0x300)
OpenGL version string: 2.1 Mesa 8.1-devel
OpenGL shading language version string: 1.20
But I have tested it just now on celestia and it now has 5 threads (on 4 core cpu) which I haven't noticed before. They not using the cores at 100%, but about 25% each and the program is slow. But I haven't noticed this threading before so there may be progress. Maybe it is due to me upgrading to X.org server 1.12 and llvm 3.0 recently.