The body of the test_unsafe function is:
- function test_unsafe(n, time=true)
0 a = zeros(Int, (320, 320, 320))
-
- # warmup
0 setk_UnSafe(a);
-
0 Profile.clear_malloc_data()
-
0 for i=1:n
-
0 setk_UnSafe(a)
-
- end
-
0 a
- end
So I clear the malloc data after running my expensive function once and
then run it a second time. There is no difference in the .mem file if I
call Profile.clear_malloc_data and then call test_unsafe a second time.
You are right about the bug but that line of code is never called in these
tests. Correcting that line did not change the allocation patterns.
On Monday, April 20, 2015 at 10:54:49 AM UTC-6, Tim Holy wrote:
>
> First, you need to run it twice, see
>
> http://docs.julialang.org/en/latest/manual/profile/#memory-allocation-analysis
>
> and the part about clear_malloc_data.
>
> Second, I think you have a bug:
> size(s::UnsafeSlice) = size(s.size)
> should presumably be
> size(s::UnsafeSlice) = s.size
>
> --Tim
>
>
> On Monday, April 20, 2015 09:35:33 AM Peter Brady wrote:
> > Here's the results of running --track-allocation=user --inline=no on
> 0.4.
> > Note that I also deleted all the macros which were affecting the
> reported
> > line numbers.
> >
> > I have three questions base on the data below:
> > 1. Why is the call to size which indexes into a tuple so expensive
> > 2. Why is setindex! so expensive?
> > 3. Why is it so expensive to update the 'start' attribute of my
> > unsafeslice?
> > Does anyone have any answers or any suggestions on what tools to use to
> > find the answers?
> >
> > Here's my session:
> >
> >
> > $ ~/gitrepos/julia0.4/julia --track-allocation=user --inline=no
> > _
> > _ _ _(_)_ | A fresh approach to technical computing
> > (_) | (_) (_) | Documentation: http://docs.julialang.org
> > _ _ _| |_ __ _ | Type "help()" for help.
> >
> > | | | | | | |/ _` | |
> > | | |
> > | | |_| | | | (_| | | Version 0.4.0-dev+4385 (2015-04-20 14:52 UTC)
> >
> > _/ |\__'_|_|_|\__'_| | Commit 5499882 (0 days old master)
> >
> > |__/ | x86_64-redhat-linux
> >
> > julia> include("test_alloc.jl")
> > test_unsafe (generic function with 2 methods)
> >
> >
> > julia> test_unsafe(1);
> >
> >
> > julia>
> > [ptb@cyrus julia]$
> >
> >
> > And the output
> >
> >
> > - using ArrayViews
> > - import Base: size, getindex, setindex!, ndims, start, stride,
> > pointer
> > -
> > - type UnsafeSlice{T,N, P<:Ptr} <: AbstractArray
> > - start::Int
> > - stride::Int
> > - size::NTuple{N,Int}
> > - p::P
> > - end
> > -
> > - size(s::UnsafeSlice) = size(s.size)
> > -
> > 7356448 size(s::UnsafeSlice, i::Int) = s.size[i]
> > -
> > - ndims{T,N}(s::UnsafeSlice{T,N}) = N
> > -
> > - getindex(s::UnsafeSlice, i::Int) = unsafe_load(s.p,
> s.start+(i-1)*
> > s.stride)
> > -
> > 1048559648 setindex!(s::UnsafeSlice, x, i::Int) = unsafe_store!(s.p, x,
> s.
> > start+(i-1)*s.stride)
> > -
> > - function UnsafeSlice(a, slicedim::Int, start=1)
> > 0 p = pointer(a)
> > -
> > 0 str = stride(a, slicedim)
> > -
> > 368 UnsafeSlice{eltype(a), ndims(a), typeof(p)}(start, str,
> size(a
> > ),p)
> > - end
> > -
> > - function update(a::UnsafeSlice, idx, off)
> > -
> > 0 for i=1:size(a, idx)
> > -
> > 0 a[i] = -10*off+i
> > - end
> > -
> > 0 a
> > - end
> > -
> > - function setk_UnSafe{T}(a::Array{T,3})
> > 0 us = UnsafeSlice(a, 3)
> > -
> > 0 for j=1:size(a,2),i=1:size(a,1)
> > -
> > 14712896 us.start = sub2ind(size(a), i, j, 1)
> > -
> > 0 update(us, 3, us.start)
> > - end
> > -
> > 0 a
> > - end
> > -
> > - function test_unsafe(n, time=true)
> > 0 a = zeros(Int, (320, 320, 320))
> > -
> > - # warmup
> > 0 setk_UnSafe(a);
> > -
> > 0 Profile.clear_malloc_data()
> > -
> > 0 for i=1:n
> > -
> > 0 setk_UnSafe(a)
> > -
> > - end
> > -
> > 0 a
> > - end
> > -
> >
> > On Monday, April 20, 2015 at 9:04:41 AM UTC-6, Peter Brady wrote:
> > > Accidentally hit reply instead of reply-all. Sorry for the double
> post.
> > >
> > > Ran my script in 0.4 and got these results...
> > >
> > > julia> test_all(5)
> > > test_stride
> > > elapsed time: 2.008043041 seconds (0 bytes allocated)
> > > test_view
> > > elapsed time: 8.871387399 seconds (42 MB allocated, 0.23% gc time in 2
> > > pauses with 1 full sweep)
> > > test_unsafe
> > > elapsed time: 2.308598574 seconds (46 MB allocated, 0.68% gc time in 2
> > > pauses with 1 full sweep)
> > > test_unsafeview
> > > elapsed time: 9.106651158 seconds (0 bytes allocated)
> > >
> > > julia> test_all(10)
> > > test_stride
> > > elapsed time: 4.012240175 seconds (0 bytes allocated)
> > > test_view
> > > elapsed time: 18.085514211 seconds (85 MB allocated, 0.16% gc time in
> 4
> > > pauses with 1 full sweep)
> > > test_unsafe
> > > elapsed time: 4.477773618 seconds (93 MB allocated, 1.12% gc time in 4
> > > pauses with 1 full sweep)
> > > test_unsafeview
> > > elapsed time: 18.146163969 seconds (0 bytes allocated)
> > >
> > > So the allocation for the new unsafeview has been reduced to zero but
> it
> > > has become slower than the regular view.
> > >
> > > Perhaps the compiler optimizations that have been discussed here are
> > > occuring since the only occurence of 'unsafeview' is the argument to a
> > > function.
> > >
> > > On Mon, Apr 20, 2015 at 12:57 AM, René Donner <[email protected]
> <javascript:>> wrote:
> > >> What about something like unsafe_updateview!(view, indices...) ?
> > >>
> > >> It could be used like this (pseudocode):
> > >> view = unsafe_view(data, 1, 1, :) # to construct / allocate
> > >> for i in ..., j in ...
> > >>
> > >> unsafe_updateview!(view, i, j, :)
> > >> # use view
> > >>
> > >> end
> > >>
> > >> In the trivial case of unsafe_view(data, :, :, i) this would boil
> down to
> > >> a single pointer update. Of course passing around these views outside
> of
> > >> their scope is rather discouraged. I use this pattern a lot and it
> proved
> > >> to be very handy / fast.
> > >>
> > >> Am 20.04.2015 um 02:08 schrieb Dahua Lin <[email protected]
> <javascript:>>:
> > >> > My benchmark shows that element indexing has been as fast as it can
> be
> > >>
> > >> for array views (or subarrays in Julia 0.4).
> > >>
> > >> > Now the problem is actually the construction of views/subarrays. To
> > >>
> > >> optimize the overhead of this part, the compiler may need to
> introduce
> > >> additional optimization.
> > >>
> > >> > Dahua
> > >> >
> > >> >
> > >> > On Monday, April 20, 2015 at 6:39:35 AM UTC+8, Sebastian Good
> wrote:
> > >> > —track-allocation still requires guesswork, as optimizations can
> move
> > >>
> > >> the allocation to a different place than you would expect.
> > >>
> > >> > On April 19, 2015 at 4:36:19 PM, Peter Brady ([email protected])
> > >>
> > >> wrote:
> > >> >> So I discovered the --track-allocation option and now I am really
> > >>
> > >> confused:
> > >> >> Here's my session:
> > >> >>
> > >> >> $ julia --track-allocation=all
> > >> >>
> > >> >> _
> > >> >>
> > >> >> _ _ _(_)_ | A fresh approach to technical computing
> > >> >>
> > >> >> (_) | (_) (_) | Documentation: http://docs.julialang.org
> > >> >>
> > >> >> _ _ _| |_ __ _ | Type "help()" for help.
> > >> >>
> > >> >> | | | | | | |/ _` | |
> > >> >> | | |
> > >> >> | | |_| | | | (_| | | Version 0.3.8-pre+13 (2015-04-17 18:08
> UTC)
> > >> >>
> > >> >> _/ |\__'_|_|_|\__'_| | Commit 0df962d* (2 days old release-0.3)
> > >> >>
> > >> >> |__/ | x86_64-redhat-linux
> > >> >>
> > >> >> julia> include("test.jl")
> > >> >> test_all (generic function with 1 method)
> > >> >>
> > >> >> julia> test_unsafe(5)
> > >> >>
> > >> >> And here's the relevant part of the resulting test.jl.mem file.
> Note
> > >>
> > >> that I commented out some calls to 'size' and replaced with the
> > >> appropriate
> > >> hard-coded values but the resulting allocation is the same... Can
> anyone
> > >> shed some light on this while I wait for 0.4 to compile?
> > >>
> > >> >> - function update(a::AbstractArray, idx, off)
> > >> >>
> > >> >> 8151120 for i=1:320 #size(a, idx)
> > >> >>
> > >> >> 0 a[i] = -10*off+i
> > >> >> - end
> > >> >> 0 a
> > >> >> - end
> > >> >> -
> > >> >>
> > >> >> - function setk_UnSafe{T}(a::Array{T,3})
> > >> >>
> > >> >> 760 us = UnsafeSlice(a, 3)
> > >> >>
> > >> >> 0 for j=1:size(a,2),i=1:size(a,1)
> > >> >>
> > >> >> 8151120 us.start = (j-1)*320+i #size(a,1)+i
> > >> >>
> > >> >> - #off = sub2ind(size(a), i, j, 1)
> > >> >> 0 update(us, 3, us.start)
> > >> >> - end
> > >> >> 0 a
> > >> >> - end
> > >> >> - function test_unsafe(n)
> > >> >> 0 a = zeros(Int, (320, 320, 320))
> > >> >> - # warmup
> > >> >> 0 setk_UnSafe(a);
> > >> >> 0 clear_malloc_data()
> > >> >> - #@time (
> > >> >> 0 for i=1:n; setk_UnSafe(a); end
> > >> >> - end
> > >> >>
> > >> >> On Sunday, April 19, 2015 at 2:21:56 PM UTC-6, Peter Brady wrote:
> > >> >> @Dahua, thanks for adding an unsafeview! I appreciate how quickly
> > >>
> > >> this community responds.
> > >>
> > >> >> I've added the following function to my test.jl script
> > >> >> function setk_unsafeview{T}(a::Array{T,3})
> > >> >>
> > >> >> for j=1:size(a,2),i=1:size(a,1)
> > >> >>
> > >> >> off = sub2ind(size(a), i, j, 1)
> > >> >> update(unsafe_view(a, i, j, :), 3, off)
> > >> >>
> > >> >> end
> > >> >> a
> > >> >>
> > >> >> end
> > >> >>
> > >> >> But I'm not seeing the large increase in performance I was
> > >>
> > >> expecting. My timings are now
> > >>
> > >> >> julia> test_all(5);
> > >> >> test_stride
> > >> >> elapsed time: 2.156173128 seconds (0 bytes allocated)
> > >> >> test_view
> > >> >> elapsed time: 9.30964534 seconds (94208000 bytes allocated, 0.47%
> gc
> > >>
> > >> time)
> > >>
> > >> >> test_unsafe
> > >> >> elapsed time: 2.169307471 seconds (16303000 bytes allocated)
> > >> >> test_unsafeview
> > >> >> elapsed time: 8.955876793 seconds (90112000 bytes allocated, 0.41%
> gc
> > >>
> > >> time)
> > >>
> > >> >> To be fair, I am cheating a bit with my custom 'UnsafeSlice' since
> I
> > >>
> > >> make only one instance and simply update the offset on each
> iteration.
> > >> If
> > >> I make it immutable and create a new instance on every iteration (as
> I do
> > >> for the view and unsafeview), things slow down a little and the
> > >> allocation
> > >>
> > >> goes south:
> > >> >> julia> test_all(5);
> > >> >> test_stride
> > >> >> elapsed time: 2.159909265 seconds (0 bytes allocated)
> > >> >> test_view
> > >> >> elapsed time: 9.029025282 seconds (94208000 bytes allocated, 0.43%
> gc
> > >>
> > >> time)
> > >>
> > >> >> test_unsafe
> > >> >> elapsed time: 2.621667854 seconds (114606240 bytes allocated,
> 2.41% gc
> > >>
> > >> time)
> > >>
> > >> >> test_unsafeview
> > >> >> elapsed time: 8.888434466 seconds (90112000 bytes allocated, 0.44%
> gc
> > >>
> > >> time)
> > >>
> > >> >> These are all with 0.3.8-pre. I'll try compiling master and see
> what
> > >>
> > >> happens. I'm still confused about why allocating a single type with
> a
> > >> pointer, 2 ints and a tuple costs so much memory though.
> > >>
> > >> >> On Sunday, April 19, 2015 at 11:38:17 AM UTC-6, Tim Holy wrote:
> > >> >> It's not just escape analysis, as this (new) issue demonstrates:
> > >> >> https://github.com/JuliaLang/julia/issues/10899
> > >> >>
> > >> >> --Tim
> > >> >>
> > >> >> On Sunday, April 19, 2015 12:33:51 PM Sebastian Good wrote:
> > >> >> > Their size seems much decreased. I’d imagine to totally avoid
> > >>
> > >> allocation in
> > >>
> > >> >> > this benchmark requires an optimization that really has nothing
> to
> > >>
> > >> do with
> > >>
> > >> >> > subarrays per se. You’d have to do an escape analysis and see
> that
> > >>
> > >> Aj never
> > >>
> > >> >> > left sumcols. Not easy in practice, since it’s passed to slice
> and
> > >>
> > >> length,
> > >>
> > >> >> > and you’d have to make sure they didn’t squirrel it away or pass
> it
> > >>
> > >> on to
> > >>
> > >> >> > someone else. Then you could stack allocate it, or even
> destructure
> > >>
> > >> it into
> > >>
> > >> >> > a bunch of scalar mutations on the stack. After eliminating dead
> > >>
> > >> code,
> > >>
> > >> >> > you’d end up with a no-allocation loop much like you’d write by
> > >>
> > >> hand. This
> > >>
> > >> >> > sort of optimization seems to be quite tricky for compilers to
> pull
> > >>
> > >> off,
> > >>
> > >> >> > but it’s a common pattern in numerical code.
> > >> >> >
> > >> >> > In Julia is such cleverness left entirely to LLVM, or are there
> > >>
> > >> optimization
> > >>
> > >> >> > passes in Julia itself? On April 19, 2015 at 6:49:21 AM, Tim
> Holy
> > >> >> > ([email protected]) wrote:
> > >> >> >
> > >> >> > Sorry to be slow to chime in here, but the tuple overhaul has
> landed
> > >>
> > >> and
> > >>
> > >> >> > they are still not zero-cost:
> > >> >> >
> > >> >> > function sumcols(A)
> > >> >> > s = 0.0
> > >> >> > for j = 1:size(A,2)
> > >> >> > Aj = slice(A, :, j)
> > >> >> > for i = 1:length(Aj)
> > >> >> > s += Aj[i]
> > >> >> > end
> > >> >> > end
> > >> >> > s
> > >> >> > end
> > >> >> >
> > >> >> > Even in the latest 0.4, this still allocates memory. On the
> other
> > >>
> > >> hand,
> > >>
> > >> >> > while SubArrays allocate nearly 2x more memory than ArrayViews,
> the
> > >>
> > >> speed
> > >>
> > >> >> > of the two (replacing `slice` with `view` above) is, for me,
> nearly
> > >> >> > identical.
> > >> >> >
> > >> >> > --Tim
> > >> >> >
> > >> >> > On Friday, April 17, 2015 08:30:27 PM Sebastian Good wrote:
> > >> >> > > This was discussed a few weeks ago
> > >>
> > >> https://groups.google.com/d/msg/julia-users/IxrvV8ABZoQ/uWZu5-IB3McJ
> > >>
> > >> >> > > I think the bottom line is that the current implementation
> > >>
> > >> *should* be
> > >>
> > >> >> > > 'zero-cost' once a set of planned improvements and
> optimizations
> > >>
> > >> take
> > >>
> > >> >> > > place. One of the key ones is a tuple overhaul.
> > >> >> > >
> > >> >> > > Fair to say it can never be 'zero' cost since there is
> different
> > >>
> > >> inherent
> > >>
> > >> >> > > overhead depending on the type of subarray, e.g. offset,
> slice,
> > >> >> > > re-dimension, etc. however the implementation is quite clever
> > >> >> > > about
> > >> >> > > allowing specialization of those.
> > >> >> > >
> > >> >> > > In a common case (e.g. a constant offset or simple stride) my
> > >> >> > > understanding
> > >> >> > > is that the structure will be type-specialized and likely
> stack
> > >>
> > >> allocated
> > >>
> > >> >> > > in many cases, reducing to what you'd write by hand. At least
> this
> > >>
> > >> is what
> > >>
> > >> >> > > they're after.
> > >> >> > >
> > >> >> > > On Friday, April 17, 2015 at 4:24:14 PM UTC-4, Peter Brady
> wrote:
> > >> >> > > > Thanks for the links. I'll check out ArrayViews as it looks
> like
> > >>
> > >> what I
> > >>
> > >> >> > > > was going to do manually without wrapping it in a type.
> > >> >> > > >
> > >> >> > > > By semi-dim agnostic I meant that the differencing algorithm
> > >>
> > >> itself only
> > >>
> > >> >> > > > cares about one dimension but that dimension is different
> for
> > >>
> > >> different
> > >>
> > >> >> > > > directions. Only a few toplevel routines actually need to
> know
> > >>
> > >> about the
> > >>
> > >> >> > > > dimensionality of the problem.
> > >> >> > > >
> > >> >> > > > On Friday, April 17, 2015 at 2:04:39 PM UTC-6, René Donner
> wrote:
> > >> >> > > >> As far as I have measured it sub in 0.4 is still not cheap,
> as
> > >>
> > >> it
> > >>
> > >> >> > > >> provides the flexibility to deal with all kinds of strides
> and
> > >>
> > >> offsets,
> > >>
> > >> >> > > >> and
> > >> >> > > >> the view object itself thus has a certain size. See
> > >> >> > > >> https://github.com/rened/FunctionalData.jl#efficiency for
> a
> > >>
> > >> simple
> > >>
> > >> >> > > >> analysis, where the speed is mostly dominated by the speed
> of
> > >>
> > >> the
> > >>
> > >> >> > > >> "sub-view" mechanism.
> > >> >> > > >>
> > >> >> > > >> To get faster views which require strides etc you can try
> > >> >> > > >> https://github.com/JuliaLang/ArrayViews.jl
> > >> >> > > >>
> > >> >> > > >> What do you mean by semi-dim agnostic? In case you only
> need
> > >>
> > >> indexing
> > >>
> > >> >> > > >> along the last dimension (like a[:,:,i] and a[:,:,:,i]) you
> can
> > >>
> > >> use
> > >>
> > >>
> > >> https://github.com/rened/FunctionalData.jl#efficient-views-details
> > >>
> > >> >> > > >> which uses normal DenseArrays and simple pointer updates
> > >>
> > >> internally. It
> > >>
> > >> >> > > >> can also update a view in-place, by just incrementing the
> > >>
> > >> pointer.
> > >>
> > >> >> > > >> Am 17.04.2015 um 21:48 schrieb Peter Brady <
> [email protected]
> > >> >> > > >>
> > >> >> > > >> > Inorder to write some differencing algorithms in a
> > >>
> > >> semi-dimensional
> > >>
> > >> >> > > >> agnostic manner the code I've written makes heavy use of
> > >>
> > >> subarrays
> > >>
> > >> >> > > >> which
> > >> >> > > >> turn out to be rather costly. I've noticed some posts on
> the
> > >>
> > >> cost of
> > >>
> > >> >> > > >> subarrays here and that things will be better in 0.4. Can
> > >>
> > >> someone
> > >>
> > >> >> > > >> comment
> > >> >> > > >> on how much better? Would subarray (or anything like it) be
> on
> > >>
> > >> par with
> > >>
> > >> >> > > >> simply passing an offset and stride (constant) and
> computing
> > >>
> > >> the index
> > >>
> > >> >> > > >> myself? I'm currently using the 0.3 release branch.
>
>