Hello.

Thanks for the reply - I do indeed have mkl-service installed. I ran conda
update --all and am still getting the same results. If anyone has any other
suggestions I'm all ears.

Thank you,

On Tue, Jun 13, 2017 at 8:00 AM, Frédéric Bastien <
[email protected]> wrote:

> This is not normal.
>
> Did you install the conda package mkl- service ?
>
> Try to update numpy. It could also help.
>
> Le lun. 12 juin 2017 07:52, Aaron Snoswell <[email protected]> a
> écrit :
>
>> I'm working through the the DeepLearning.net tutorials using Windows 64
>> bit, Python 3.6 and Theano installed through conda.
>>
>> I was able to run the Classifying MNIST digits using Logistic Regression
>> <http://deeplearning.net/tutorial/logreg.html> demo fine, and got the
>> same results as listed in the tutorial, hitting 4 epochs/second (about
>> double the listed CPU performance in the tutorial). I then tried running
>> the MLP tutorial code <http://deeplearning.net/tutorial/mlp.html> (classify
>> MNIST digits using a simple MLP). During execution, the process gobbles up
>> memory continuously until I get a MemoryError and the python crashes.
>> Watching the task manager, I will occasionally see the memory usage drop -
>> I assume this is the garbage collector kicking in, but it happens rarely.
>>
>>
>> <https://lh3.googleusercontent.com/-4EYsaeVqr_w/WT5-SyEMmWI/AAAAAAAAEZw/z3aqQrLFVVcdVfqfnlLDvvS7n8WH8Qt9QCLcB/s1600/theano-running-memory.PNG>
>>
>>  I've tried adjusting the MLP 'batch_size' parameter;
>>
>>    - With a value of 1000 (therefore n_train_batches == 50) the code
>>    runs until the patience condition causes it to stop (no crash)
>>    - With the default of 20 (n_train_batches == 2500) the code gets to
>>    epoch 17 and crashes
>>    - With a value of 10 (n_train_batches == 5000) I only get to epoch 3
>>    before it crashes
>>
>> Is this behavior expected with the hardware specs of the laptop I'm
>> running on? I've attached my DxDiag results here, but I've got 20GB of ram
>> on this machine.
>>
>> Just trying to figure out if this crashing behavior is expected, or if
>> I'm seeing a memory leak of some sort.
>>
>> Thanks.
>>
>> --
>>
>> ---
>> You received this message because you are subscribed to the Google Groups
>> "theano-users" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to [email protected].
>> For more options, visit https://groups.google.com/d/optout.
>>
> --
>
> ---
> You received this message because you are subscribed to a topic in the
> Google Groups "theano-users" group.
> To unsubscribe from this topic, visit https://groups.google.com/d/
> topic/theano-users/Rz408i5rx2k/unsubscribe.
> To unsubscribe from this group and all its topics, send an email to
> [email protected].
> For more options, visit https://groups.google.com/d/optout.
>



-- 

Aaron Snoswell

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to