Hi Steven,
thank you very much for the clarification. It was indeed possible to pass
the length if the vector and with pointer_to_array it works smoothly.
Tomas
On Tuesday, 2 August 2016 13:46:47 UTC+2, Steven G. Johnson wrote:
>
>
>
> On Tuesday, August 2, 2016 at 7:39:16 AM UTC-4,
Hi All,
I am trying to bind a fortran library for optimization
(http://napsu.karmitsa.fi/lmbm/) to Julia.
To do so, I would like to get a pointer to function written in Julia, which
I try to do as
function fOpt(x::Array{Float64,1},g::Array{Float64,1})
fill!(g,1.0);
Hi Todd,
I have been looking at latte and it does not seem to be useful for me,
since I need some special constructs and they are just not available.
Nevertheless, I would like to ask you, if Latte uses parallelization? In my
own implementation, I am struggling to exploit multi-core hw.
Thank
>
> Hi Todd,
>
I have tried several times to use ParallelAccelerator to speed up my toy
Neural Network library, but I never had any significant performance boost.
I like the idea of the project a lot, sadly I was never able to fully
utilise it.
Best wishes,
Tomas
Hi James,
thank for the reply. Though in your implementation the reading is not in a
separate process / thread, as I expect that you are bound to IO operations.
In my problem there is computationally intensive post-processing. Should I
modify the iotask as
iotask = @task begin
Hi All,
I would like to implement an asynchronous reading from file.
I am doing stochastic gradient descend and while I am doing the
optimisation, I would like to load the data on the background. Since
reading of the data is followed by a quite complicated parsing, it is not
just simple IO
Dear Sam,
the output of the benchmark is following
105.290122 seconds (31.43 k allocations: 1.442 MB, 0.00% gc time)
107.445101 seconds (1.37 M allocations: 251.368 MB, 0.12% gc time)
Tomas
Thank you very much Tim.
I am using the profiler and your package ProfileView quite extensively and
I know where is my Achille heel in the code, and it is cpu bound. That's
why I am so puzzled with threads.
I will try to use @code_warntype, never use it before.
Best wishes,
Tomas
Hello all,
I was polishing my call and I have found the following definition of daxpy!
I was not aware of
function axpy!{Ti<:Integer,Tj<:Integer}(α, x::AbstractArray,
rx::AbstractArray{Ti}, y::AbstractArray, ry::AbstractArray{Tj})
if length(x) != length(y)
Thanks a lot for the suggestions.
As I have mentioned, it was really a toy problem, but I am not getting a
significant speedup on a bigger problem, where threads are nicely separated
either and the problem is very CPU bound either. I would be very interested
to know about tool that would point
Hi All,
I would like to ask if someone has an experience with Threads as they are
implemented at the moment in the master branch.
After the successful compilation (put JULIA_THREADS=1 to Make.user)
I have played with different levels of granularity, but usually the code
was slower or more or
Hi All,
I wanted to try julia with threads, therefore I have cloned GIT repository,
checkout version 0.4.3
git checkout release-0.4
I have put JULIA_THREADS=1, here is my Make.user
CC=/opt/rh/devtoolset-2/root/usr/bin/gcc
CXX=/opt/rh/devtoolset-2/root/usr/bin/g++
JULIA_THREADS=1
and compiled
Hi All,
I know that julia does not have at the moment native support for threads, but
it exist in an experimental branch. Is there any tutorial on this, such that
average user can try it? I would like to try it speed my code. I do separate
processes at the moment and I think the overhead is
Hello,
I have a problem to which I have found a dirty solution and I am keen to
know, if there is a principal one.
I have a composite type defined as
type Outer{T}
A::T
B::T
end
where A and and B are composite types
Then I want to create constructor
function Outer(k::Int)
Hi,
I have a question related to use of Mocha, particularly if I can tweak it
to my problem. I want to use the library for multi-instance learning, which
means that each sample is composed of multiple instances, but the number of
instances in each sample differs from sample to sample. You can
Hi all,
I would like to ask if it is possible to enforce memory alignment for large
arrays, such that that arrays are aligned to 64 bytes and can be
efficiently with SIMD instructions. I assume to call library functions
wirtten in c/c++.
Thanks for the answer.
Tomas
16 matches
Mail list logo