Hi Fred,

      What command would I use to uninstall theano.   I also installed the 
latest development version.    Does it need a different command to uninstall?

      Thanks.

Chi


From: [email protected] [mailto:[email protected]] On 
Behalf Of Frédéric Bastien
Sent: Tuesday, November 15, 2016 3:11 PM
To: theano-users
Subject: Re: [theano-users] TypeError: Cannot convert Type TensorType(float64, 
3D) (of Variable Subtensor{:int64:}.0) into Type TensorType(float64, (False, 
True, False)). You can try to manually convert Subtensor{:int64:}.0 into a 
TensorType(float64, (False, Tr...


To be safer, uninstall Theano a few time before and make sure you can't import 
it before reinstalling.

Le 15 nov. 2016 16:20, "Pascal Lamblin" 
<[email protected]<mailto:[email protected]>> a écrit :
On Tue, Nov 15, 2016, Chi Ku wrote:
> Hi Pascal,
>
>    How would I find out which version of Theano I have?

You can try to print theano.version.version, that should be generated when you 
call "pip install"

>    Can I use the following command to install the latest development version?
>
>       <sudo> pip install <--user> <--no-deps> 
> git+https://github.com/Theano/Theano

I think so.

>
>    Thanks.
>
> Chi
>
>
>
>
>
>
> -----Original Message-----
> From: [email protected]<mailto:[email protected]> 
> [mailto:[email protected]<mailto:[email protected]>] 
> On Behalf Of Pascal Lamblin
> Sent: Monday, November 14, 2016 12:37 PM
> To: [email protected]<mailto:[email protected]>
> Subject: Re: [theano-users] TypeError: Cannot convert Type 
> TensorType(float64, 3D) (of Variable Subtensor{:int64:}.0) into Type 
> TensorType(float64, (False, True, False)). You can try to manually convert 
> Subtensor{:int64:}.0 into a TensorType(float64, (False, Tr...
>
> Which version of Theano are you using?
> If you are using the 0.8.2 release, can you try the latest development 
> version?
>
> On Sat, Nov 12, 2016, [email protected]<mailto:[email protected]> wrote:
> >
> > I got the following errors when calling tensor.grad() to compute the
> > symbolic gradient of the finetune_cost of a hybrid DBN-RNN model.
> > I tried changing the way this expression is formed in several ways without
> > any success.   I need help from experts.   A tar file of the source code
> > and data is attached here.
> >
> > The pretraining code for file hybrid_array.py ran for 0.48m
> > ... getting the finetuning functions
> > Traceback (most recent call last):
> >   File "/usr/lib/python2.7/pdb.py", line 1314, in main
> >     pdb._runscript(mainpyfile)
> >   File "/usr/lib/python2.7/pdb.py", line 1233, in _runscript
> >     self.run(statement)
> >   File "/usr/lib/python2.7/bdb.py", line 400, in run
> >     exec cmd in globals, locals
> >   File "<string>", line 1, in <module>
> >   File "hybrid_array.py", line 494, in <module>
> >     test_DBN()
> >   File "hybrid_array.py", line 412, in test_DBN
> >     learning_rate=finetune_lr
> >   File "hybrid_array.py", line 253, in build_finetune_functions
> >     gparams = T.grad(self.finetune_cost, self.params)
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 561, in grad
> >     grad_dict, wrt, cost_name)
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1324, in _populate_grad_dict
> >     rval = [access_grad_cache(elem) for elem in wrt]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_gra    term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 973, in access_term_cache
> >     output_grads = [access_grad_cache(var) for var in node.outputs]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1279, in access_grad_cache
> >     term = access_term_cache(node)[idx]
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gradient.py", line
> > 1113, in access_term_cache
> >     input_grads = node.op.grad(inputs, new_output_grads)
> >   File
> > "/usr/local/lib/python2.7/dist-packages/theano/scan_module/scan_op.py",
> > line 2523, in grad
> >     outputs = local_op(*outer_inputs)
> >   File "/usr/local/lib/python2.7/dist-packages/theano/gof/op.py", line 611,
> > in __call__
> >     node = self.make_node(*inputs, **kwargs)
> >   File
> > "/usr/local/lib/python2.7/dist-packages/theano/scan_module/scan_op.py",
> > line 430, in make_node
> >     new_inputs.append(format(outer_seq, as_var=inner_seq))
> >   File
> > "/usr/local/lib/python2.7/dist-packages/theano/scan_module/scan_op.py",
> > line 422, in format
> >     rval = tmp.filter_variable(rval)
> >   File "/usr/local/lib/python2.7/dist-packages/theano/tensor/type.py", line
> > 233, in filter_variable
> >     self=self))
> > TypeError: Cannot convert Type TensorType(float64, 3D) (of Variable
> > Subtensor{:int64:}.0) into Type TensorType(float64, (False, True, False)).
> > You can try to manually convert Subtensor{:int64:}.0 into a
> > TensorType(float64, (False, True, False)).
> > Uncaught exception. Entering post mortem debugging
> > Running 'cont' or 'step' will restart the program
> > >
> > /usr/local/lib/python2.7/dist-packages/theano/tensor/type.py(233)filter_variable()
> > -> self=self))
> >
> >
> >
> > --
> >
> > ---
> > You received this message because you are subscribed to the Google Groups 
> > "theano-users" group.
> > To unsubscribe from this group and stop receiving emails from it, send an 
> > email to 
> > [email protected]<mailto:theano-users%[email protected]>.
> > For more options, visit https://groups.google.com/d/optout.
>
>
>
> --
> Pascal
>
> --
>
> ---
> You received this message because you are subscribed to a topic in the Google 
> Groups "theano-users" group.
> To unsubscribe from this topic, visit 
> https://groups.google.com/d/topic/theano-users/8uSF6ub-drA/unsubscribe.
> To unsubscribe from this group and all its topics, send an email to 
> [email protected]<mailto:theano-users%[email protected]>.
> For more options, visit https://groups.google.com/d/optout.
>
> --
>
> ---
> You received this message because you are subscribed to the Google Groups 
> "theano-users" group.
> To unsubscribe from this group and stop receiving emails from it, send an 
> email to 
> [email protected]<mailto:theano-users%[email protected]>.
> For more options, visit https://groups.google.com/d/optout.

--
Pascal

--

---
You received this message because you are subscribed to the Google Groups 
"theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to 
[email protected]<mailto:theano-users%[email protected]>.
For more options, visit https://groups.google.com/d/optout.
--

---
You received this message because you are subscribed to a topic in the Google 
Groups "theano-users" group.
To unsubscribe from this topic, visit 
https://groups.google.com/d/topic/theano-users/8uSF6ub-drA/unsubscribe.
To unsubscribe from this group and all its topics, send an email to 
[email protected]<mailto:[email protected]>.
For more options, visit https://groups.google.com/d/optout.

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to