1. You should change:
C = complex(zeros(Float64, Ly, Lx)
to:
C = zeros(Complex{Float64}, Ly, Lx)
[the way you are doing it there creates a float version, then a complex
version, then trashes it]
2. The algorithm after the above change allocates 3 * limit * (limit/2) *
samples * 16 bytes in the @parallel loop
[3 being one allocation for A, B, and C ; 16 being sizeof(Complex{Float64})]
[without change #1 it's 4 * limit * (limit/2) * samples * 16]
[so yeah, makes sense that would grow fast as limit increases]
You could reduce this by 50% by running the paralel computation over D and
E directly, then scaling it down afterwards. Which would also be faster,
because you don't have to do that element-wise multiply or the element-wise
scaling. You're walking D 3 times more often than you need to.
3. You could consider going Float32 (factor of 2 savings of course), or
Float16 if those would meet your needs.
4. Style-wise, you should do a deconstructing bind instead of using tuphat:
(A, B, C) = @parallel (add_three_tuples) for ...
or really:
(D, E) = @parallel (add_two_tuples) for ...
5. You'll probably get better numerical stability if you scale down prior
to the addition step (so doing D[i,j] += Aij * Bij / samples / samples
instead of D[i,j] += Aij * Bij and then dividing before returning)
If you make the above changes (specifically 1 and 2) you'll have the
allocation down to the minimum size possible for this approach to
threading... ie (threads) * (size of returned data). If the thread factor
remains a problem, it's not *so* bad compiling julia from source to get
SharedArrays.
The only other thing to consider would be how you are using the return
value. Perhaps you could get away with computing a smaller sized return
value?
On Tue, Feb 4, 2014 at 1:49 PM, Alex C <[email protected]> wrote:
> Sorry for the confusion. I was trying to get a simple example to work so I
> wouldn't get distracted by details. The @everywhere did the trick.
> This is the fastest parallel version of the code that I was able to get
> working. However, I easily run into memory limitations (8GB RAM) as I
> increase the variable 'limit'.
>
> function expensive_hat(S::Array{Complex{Float64},2},limit::Int64)
>
> samples = size(S,2);
> Ly = int64(limit/4);
> Lx = int64(limit/2);
>
> @everywhere add_three_tuple(x,y) = (x[1]+y[1], x[2]+y[2], x[3]+y[3])
>
> tuphat = @parallel (add_three_tuple) for k = 1:samples
> A = B = zeros(Float64,Ly,Lx);
> C = complex(zeros(Float64,Ly,Lx));
> for j = 1:Lx, i = 1:Ly
> Sx = S[i,k];
> Sy = S[j,k];
> Sxy = S[i+j,k];
> Sxyc = conj(Sxy);
> A[i,j] += abs2(Sy*Sx);
> B[i,j] += abs2(sqrt(Sxyc*Sxy));
> C[i,j] += Sxyc*Sy*Sx;
> end
> (A, B, C)
> end
>
> D = tuphat[1].*tuphat[2]/samples/samples;
> E = tuphat[3]/samples;
> return (D, E);
>
> end
>
> srand(77);
> data = rand(24000,64);
> limit = 4000;
> S = rfft(data,1)./24000;
>
> @time (D, E) = expensive_hat(S,limit);
>
>
>
>
> On Tuesday, February 4, 2014 3:01:17 PM UTC-5, David Salamon wrote:
>
>> huh.
>>
>> maybe @everywhere in front of the function definition? I'm not sure
>>
>>
>> On Tue, Feb 4, 2014 at 10:53 AM, Alex C <[email protected]> wrote:
>>
>>> Thanks for the hint. Getting rid of 'mx' and 'my' definitely helps.
>>>
>>> I couldn't figure out how to implement the parallel version of tuple
>>> adding. This is what I've got. It crashes my Julia Studio console when I
>>> try to run it. What am I missing?
>>>
>>> add_two_tuple(x,y) = (x[1]+y[1], x[2]+y[2], x[3]+y[3])
>>>
>>>
>>> @parallel (add_two_tuple) for i = 1:100
>>>
>>>
>>> x = (1,2,3)
>>>
>>> end
>>>
>>>
>>> I also don't have the SharedArray function (running Julia 0.2.0) so I
>>> couldn't implement the other alternative.
>>>
>>> Alex
>>>
>>>
>>> On Monday, February 3, 2014 11:45:57 AM UTC-5, David Salamon wrote:
>>>
>>>> You're not out of the no-slicing woods yet. Looks like you can get rid
>>>> of `mx` and `my`
>>>>
>>>> for i=1:limit, j=1:int64(limit/2)
>>>> end
>>>>
>>>>
>>>>
>>>> As far as parallelizing, you could define:
>>>> three_tup_add(a, b, c) = (a[1] + b[1] + c[1], a[2] + b[2] + c[2], a[3]
>>>> + b[3] + c[3])
>>>>
>>>> and then do a @parallel (three_tup_add) over your sample index?
>>>>
>>>> for that matter, why not compute the two parts of the answer directly
>>>> rather than going via A, B, and C?
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Mon, Feb 3, 2014 at 8:11 AM, Alex C <[email protected]> wrote:
>>>>
>>>>> Thanks. I've re-written the function to minimize the amount of copying
>>>>> (i.e. slicing) that is required. But now, I'm befuddled as to how to
>>>>> parallelize this function using Julia. Any suggestions?
>>>>>
>>>>> Alex
>>>>>
>>>>> function expensive_hat(S::Array{Complex{Float64},2},
>>>>> mx::Array{Int64,2}, my::Array{Int64,2})
>>>>>
>>>>> samples = 64
>>>>> A = zeros(size(mx));
>>>>> B = zeros(size(mx));
>>>>> C = zeros(size(mx));
>>>>>
>>>>> for i = 1:samples
>>>>> Si = S[:,i];
>>>>> Sx = Si[mx];
>>>>> Sy = Si[my];
>>>>> Sxy = Si[mx+my];
>>>>> Sxyc = conj(Sxy);
>>>>>
>>>>> A += abs2(Sy .* Sx);
>>>>> B += abs2(sqrt(Sxyc .* Sxy));
>>>>> C += Sxyc .* Sy .* Sx;
>>>>> end
>>>>>
>>>>> ans = (A .* B ./ samples ./ samples, C./samples)
>>>>> return ans
>>>>>
>>>>> end
>>>>>
>>>>> data = rand(24000,64);
>>>>> limit = 2000;
>>>>>
>>>>> ix = int64([1:limit/2]);
>>>>> iy = ix[1:end/2];
>>>>> mg = zeros(Int64,length(iy),length(ix));
>>>>> mx = broadcast(+,ix',mg);
>>>>> my = broadcast(+,iy,mg);
>>>>> S = rfft(data,1)./24000;
>>>>>
>>>>> @time (AB, C) = expensive_hat(S,mx,my);
>>>>>
>>>>>
>>>>
>>>>
>>