Hyperthreading? Of the threshold is 16 but you're really only getting 8 cores, you might only get scaling up to 8.
> On Dec 4, 2014, at 3:24 PM, Viral Shah <vi...@mayin.org> wrote: > > >> On 05-Dec-2014, at 1:32 am, Douglas Bates <dmba...@gmail.com> wrote: >> >> On Thursday, December 4, 2014 1:50:06 PM UTC-6, Viral Shah wrote: >>> On 05-Dec-2014, at 1:16 am, Douglas Bates <dmb...@gmail.com <javascript:>> >>> wrote: >>> >>> Thanks, I'll try that. I'm still curious as to why there is so little >>> difference between 8 and 16 threads. >> >> peakflops() just performs a matrix multiplication to estimate the flops. It >> uses a 2000x2000 matrix by default, which is good for most laptops, but for >> bigger machines with more cores, one often needs to use a larger matrix to >> see the speedup. >> >> peakflops(8000) should give a good indication. I am not sure what the >> running time will be, so you may want to gradually increase the size. >> >> >> 8000 is reasonable on this machine and it does stabilize the results from >> repeated timings. But I still have essentially no difference between 8 and >> 16 threads. I wonder if somehow the NUM_THREADS is being set to 8, although >> looking in the deps/Makefile it does seem that it should be 16 > > > I tried on julia.mit.edu, and I do see a scale up from 1->16 processors with > peakflops(4000). That seems to suggest that the build is ok, and openblas can > scale. I think it would be best to check with Xianyi about this - perhaps > file an issue against OpenBLAS? > > Perhaps someone here may have some other ideas too. > > -viral > > >> >> julia> blas_set_num_threads(4) >> >> julia> [peakflops(8000)::Float64 for i in 1:6] >> 6-element Array{Float64,1}: >> 8.66823e10 >> 8.65584e10 >> 8.65692e10 >> 8.64753e10 >> 8.64083e10 >> 8.63359e10 >> >> julia> blas_set_num_threads(8) >> >> julia> [peakflops(8000)::Float64 for i in 1:6] >> 6-element Array{Float64,1}: >> 1.68008e11 >> 1.67772e11 >> 1.67378e11 >> 1.67397e11 >> 1.6746e11 >> 1.67623e11 >> >> julia> blas_set_num_threads(16) >> >> julia> [peakflops(8000)::Float64 for i in 1:6] >> 6-element Array{Float64,1}: >> 1.66779e11 >> 1.70068e11 >> 1.698e11 >> 1.70419e11 >> 1.70601e11 >> 1.67226e11 >> >> >> >> -viral >> >> >> >>> >>> -viral >>> >>> On Friday, December 5, 2014 1:00:39 AM UTC+5:30, Douglas Bates wrote: >>> I have been working on a package https://github.com/dmbates/ParalllelGLM.jl >>> <https://github.com/dmbates/ParalllelGLM.jl> and noticed some peculiarities >>> in the timings on a couple of shared-memory servers, each with 32 cores. >>> In particular changing from 16 workers to 32 workers actually slowed down >>> the fitting process. So I decided to check how changing the number of >>> OpenBLAS threads affected the peakflops() result. I end up with >>> essentially the same results for 8, 16 and 32 threads on this machine with >>> 32 cores. Is that to be expected? >>> >>> _ _ _(_)_ | A fresh approach to technical computing >>> (_) | (_) (_) | Documentation: http://docs.julialang.org >>> <http://docs.julialang.org/> >>> _ _ _| |_ __ _ | Type "help()" for help. >>> | | | | | | |/ _` | | >>> | | |_| | | | (_| | | Version 0.4.0-dev+1944 (2014-12-04 15:06 UTC) >>> _/ |\__'_|_|_|\__'_| | Commit 87e9ee1* (0 days old master) >>> |__/ | x86_64-unknown-linux-gnu >>> >>> julia> [peakflops()::Float64 for i in 1:6] >>> 6-element Array{Float64,1}: >>> 1.41151e11 >>> 1.1676e11 >>> 1.27597e11 >>> 1.27607e11 >>> 1.27518e11 >>> 1.27478e11 >>> >>> julia> CPU_CORES >>> 32 >>> >>> julia> blas_set_num_threads(16) >>> >>> julia> [peakflops()::Float64 for i in 1:6] >>> 6-element Array{Float64,1}: >>> 1.23523e11 >>> 1.27119e11 >>> 1.11381e11 >>> 1.17847e11 >>> 1.28415e11 >>> 1.17998e11 >>> >>> julia> blas_set_num_threads(8) >>> >>> julia> [peakflops()::Float64 for i in 1:6] >>> 6-element Array{Float64,1}: >>> 1.25194e11 >>> 1.20969e11 >>> 1.25777e11 >>> 1.20757e11 >>> 1.26086e11 >>> 1.20958e11 >>> >>> julia> versioninfo(true) >>> Julia Version 0.4.0-dev+1944 >>> Commit 87e9ee1* (2014-12-04 15:06 UTC) >>> Platform Info: >>> System: Linux (x86_64-unknown-linux-gnu) >>> CPU: AMD Opteron(tm) Processor 6328 >>> WORD_SIZE: 64 >>> "Red Hat Enterprise Linux Server release 6.5 (Santiago)" >>> uname: Linux 2.6.32-431.3.1.el6.x86_64 #1 SMP Fri Dec 13 06:58:20 EST 2013 >>> x86_64 x86_64 >>> Memory: 504.78467178344727 GB (508598.8125 MB free) >>> Uptime: 261586.0 sec >>> Load Avg: 0.08740234375 0.19384765625 0.8330078125 >>> AMD Opteron(tm) Processor 6328 : >>> speed user nice sys idle >>> irq >>> #1-32 3199 MHz 1855973 s 23392 s 670932 s 834073187 s >>> 21 s >>> >>> BLAS: libopenblas (USE64BITINT NO_AFFINITY PILEDRIVER) >>> LAPACK: libopenblas >>> LIBM: libopenlibm >>> LLVM: libLLVM-3.5.0 >>> Environment: >>> TERM = screen >>> PATH = >>> /s/cmake-3.0.2/bin:/s/gcc-4.9.2/bin:./u/b/a/bates/bin:/usr/lib64/qt-3.3/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/s/std/bin:/usr/afsws/bin: >>> >>> WWW_HOME = http://www.stat.wisc.edu/ <http://www.stat.wisc.edu/> >>> JULIA_PKGDIR = /scratch/bates/.julia >>> HOME = /u/b/a/bates >>> >>> Package Directory: /scratch/bates/.julia/v0.4 >>> 2 required packages: >>> - Distributions 0.6.1 >>> - Docile 0.3.2 >>> 5 additional packages: >>> - ArrayViews 0.4.8 >>> - Compat 0.2.5 >>> - PDMats 0.3.1 >>> - ParallelGLM 0.0.0- master (unregistered) >>> - StatsBase 0.6.10 >