OK, that works fine. Thanks. I think it would be a good idea to drop the 
matlab-ism in future versions.

Patrik

On Thursday, November 17, 2016 at 11:51:09 AM UTC+1, Simon Byrne wrote:

> I'm not familiar with the package in question, but this line:
>
> w = Any[ 0.1*randn(1,13), 0 ]
>
> may be what is causing the problem. It is creating a 2-element Vector, the 
> first element of which is a 1x13 Matrix, and the second element is a scalar 
> 0. The analogous object in R would be:
>
> W = list(matrix(0.1*rnorm(13),nrow=1), 0)
>
> In Julia, extraneous dimensions have an implicit index of 1 (this is a 
> matlab-ism, and may disappear in future), so w[1], w[1,1], w[1,1,1] are 
> all identical (and equivalent to W[[1]] in R). w[1,:] is a bit of an odd 
> case in that it returns a 1-element Vector containing a Matrix, but would 
> be equivalent to W[1] in R.
>
> I think what you may want is actually
>
> w[1][(w[1].<z) & (w[1].>-(z))]
>
> which can be written more clearly as
>
> w[1][-z .< w[1] .< z]
>
>
> -Simon
>
>
>
>
>
> On Thursday, 17 November 2016 09:39:14 UTC, Patrik Waldmann wrote:
>>
>> I guess I should have explained my problem clearer. If I run the code 
>> without w[1,(w[1].<z)&(w[1].>-(z))] = 0, and do:
>> dump(w)
>> Array{Any}((2,))
>>   1: 
>> Array{Float64}((1,13)) 
>> [-0.681392 0.595298 … 0.893845 -3.5044]
>>   2: Float64 22.447679788630705
>>
>> and 
>> println(w[1])
>> [-0.681392 0.595298 -0.394906 0.776983 -1.11178 3.11679 -0.0984956 
>> -2.18501 0.928204 -0.484802 -1.86844 0.893845 -3.5044]
>>
>> println(w[1,1])
>> [-0.681392 0.595298 -0.394906 0.776983 -1.11178 3.11679 -0.0984956 
>> -2.18501 0.928204 -0.484802 -1.86844 0.893845 -3.5044]
>>
>> println(w[1,:])
>> Any[
>> [-0.681392 0.595298 -0.394906 0.776983 -1.11178 3.11679 -0.0984956 
>> -2.18501 0.928204 -0.484802 -1.86844 0.893845 -3.5044]]
>>
>>
>> This is very confusing for an R user like me. How do I access the column 
>> indexes of w[1] and apply the logical expression 
>> w[1,(w[1].<z)&(w[1].>-(z))] = 0 ?
>>
>> Patrik
>>
>>
>>
>>
>> On Thursday, November 17, 2016 at 6:58:14 AM UTC+1, Jeffrey Sarnoff wrote:
>>
>>> good things to know about how indexing works
>>>
>>>
>>> The indices for a Vector, or a column or row of a Matrix start at *1*
>>>
>>> ```
>>> length(avector)   # gets the number of elements in avector
>>>
>>> avector[1]        # gets the first item in avector
>>> avector[end]      # gets the final item in avector 
>>> avector[1:end]    # gets all elements of avector
>>>
>>> int_column_vector = [10, 20, 30]
>>>  10
>>>  20
>>>  30
>>>
>>> int_column_vector[1]
>>>  10
>>> # do not use zero as an index
>>> int_column_vector[ 0 ]
>>> ERROR: BoundsError:
>>> # do not use false, true as indices because avec[ false ] means avec[ 0 ]
>>>
>>> ```
>>>
>>> in ` w[1,(w[1].<z)&(w[1].>-(z))] = 0 `, the second index can simplify to 
>>> `false`   (consider this)
>>> ```
>>> avec = [ 10, 20, 30 ]
>>> avec1 = avec[ 1 ] 
>>> avec1 == avec[ 1 + false ]
>>> avec2 = avec[  2 ]
>>> avec2 == avec[ 1 + true ] 
>>> ```
>>>
>>> As a start, recheck indexing expressions, be more sure they do what you 
>>> want them to do.
>>>
>>>
>>> On Wednesday, November 16, 2016 at 1:36:57 PM UTC-5, Patrik Waldmann 
>>> wrote:
>>>>
>>>> Hi,
>>>>
>>>> I'm an R user trying to learn Julia. I got hold of some code from the 
>>>> Knet package that I was playing around with. My goal is to set values to 
>>>> zero in a loop based on a logical expression, but I cannot figure out how 
>>>> the indexing works. Any help would be appreciated (the problem lies in 
>>>> w[1,(w[1].<z)&(w[1].>-(z))] = 0):
>>>>
>>>> using Knet
>>>> predict(w,x) = w[1]*x .+ w[2]
>>>> lambda = 2
>>>> z = Array{Float64}(1,13)
>>>> loss(w,x,y) = sumabs2(y - predict(w,x)) / size(y,2)
>>>> lossgradient = grad(loss)
>>>> function train(w, data; lr=.1)
>>>>     for (x,y) in data
>>>>         dw = lossgradient(w, x, y)
>>>>         z[:] = lr * lambda
>>>>         w[1] -= lr * dw[1]
>>>>         w[2] -= lr * dw[2]
>>>>         w[1,(w[1].<z)&(w[1].>-(z))] = 0
>>>>     end
>>>>     return w
>>>> end
>>>> url = "
>>>> https://archive.ics.uci.edu/ml/machine-learning-databases/housing/housing.data
>>>> "
>>>> rawdata = readdlm(download(url))
>>>> x = rawdata[:,1:13]'
>>>> x = (x .- mean(x,2)) ./ std(x,2)
>>>> y = rawdata[:,14:14]'
>>>> w = Any[ 0.1*randn(1,13), 0 ]
>>>> niter = 25
>>>> lossest = zeros(niter)
>>>> for i=1:niter; train(w, [(x,y)]); lossest[i]=loss(w,x,y); end
>>>>
>>>>
>>>> Best regards,
>>>>
>>>> Patrik
>>>>
>>>

Reply via email to