Realy ? Is imposibly smoething like :
take 1. column and copmute on 1. core, wihout waiting for end of 1. oparation take 2. column and copmpute on 2. cores .etc.... ?
Paul

W dniu 2015-01-31 o 16:32, Tim Holy pisze:
Paul, until the "threads" branch gets merged, I recommend that you just accept
the fact that you'll only have 1 core active for most operations.

--Tim

On Saturday, January 31, 2015 07:15:25 AM paul analyst wrote:
Thx, but, no.
For sparse matrix 10^5,10^4,0.002 is the same . Time for both whiles is
about 48 sek, only 11% o cores is used. I vave 8 cores, 7 sleeps:/
Paul

W dniu sobota, 31 stycznia 2015 15:50:02 UTC+1 użytkownik Sam Kaplan

napisał:
Hi Paul,

If D is allocated on the master, then Julia will need to pass D from the
master to the workers.  I'm guessing that this communication might be more
expensive than the compute in your loops.  It may be useful to take a look
at distributed arrays in the parallel section of the Julia docs.

Hope it helps.

Sam

On Saturday, January 31, 2015 at 7:38:22 AM UTC-6, paul analyst wrote:
Parallel loop, what wroong ? Parallel is slower then normal

julia> @time for i=1:l

        w[i]=var(D[:,i])
        end

elapsed time: 4.443197509 seconds (14074576 bytes allocated)


julia> @time ww=@parallel (hcat) for i=1:l

        var(D[:,i])
        end

elapsed time: 5.287007403 seconds (435449580 bytes allocated, 5.00% gc
time)
1x10000 Array{Float64,2}:

Paul

julia> @time for i=1:l

        w[i]=var(D[:,i])
        end

elapsed time: 4.331569152 seconds (8637464 bytes allocated)

julia> @time ww=@parallel (hcat) for i=1:l

        var(D[:,i])
        end

elapsed time: 4.908234336 seconds (422121448 bytes allocated, 4.85% gc
time)

1x10000 Array{Float64,2}:
  0.000703737  0.000731674  0.000582672  0.00080388    0.000759479

0.000402509  0.0007118  0.000989408

julia> size(D)
(10000,10000)

Reply via email to