Hi Yang,

can you share more about how to draw the graph and label parameter to 
debug? I met one some problem, but my usage is different:
I first trained a modelA, then I use this modelA just as part of an 
function during training modelB, and modelA should be fixed, but when I try 
to compile modelB, it dropped this error: DisconnetedInputError: xxxx by a 
non-differentiable operator: b.
But I even don't know where this b occurred, could you give some advice? 
Thanks. 

On Wednesday, January 20, 2016 at 9:24:21 AM UTC+8, Yang Xiang wrote:
>
> Thanks Daniel,
>
> I tried to draw out the whole graph and label all the parameters along the 
> path, and finally fixed the problem. The name of one parameter (when I 
> copied the paramters from the forward RNN to backward) was misspelled. 
>
> Finally I still have no idea about why the NIL params for grad would also 
> caused this problem, and I suggest this warning (exception) should be threw 
> along with the name of the disconnected paramter but not only some type 
> information. And one thing more I learned is that drawing the graph and 
> labeling the parameters is a good way for debugging this disconnect error.
>
> Yang
>
> 在 2016年1月17日星期日 UTC+8下午9:53:53,Daniel Renshaw写道:
>>
>> Are you saying you have code of this form:
>>
>> import theano.tensor as tt
>>
>> x = tt.matrix()
>> c = tt.sum(2 * x)
>> gs = tt.grad(c, [])
>>
>> i.e. an attempt to compute the gradient of some cost c with respect to... 
>> nothing, is generating the exception whose details you posted?
>>
>> If so we'll probably need to see what the cost computation is, can you 
>> share more code? Have you been able to reproduce the problem with simple 
>> code that can be executed without any external dependencies?
>>
>> Daniel
>>
>>
>>
>> On 16 January 2016 at 15:02, Yang Xiang <[email protected]> wrote:
>>
>>> Hi all,
>>>
>>> I encountered theano.gradient.DisconnectedInputError when I wrote my 
>>> code for an end-to-end process. I have a series of parameters to update. In 
>>> order to check which parameter caused the disconnect error, I removed them 
>>> from the function's parameters one by one. But after I removed all the 
>>> parameters (params=[]), this error was still there? What does this case 
>>> mean?
>>>
>>> The error report stated: theano.gradient.DisconnectedInputError: grad 
>>> method was asked to compute the gradient with respect to a variable that is 
>>> not part of the computational graph of the cost, or is used only by a 
>>> non-differentiable operator: <TensorType(float64, 4D)>
>>>
>>> Could anyone help?
>>>
>>> Thanks.
>>>
>>> Yang
>>>
>>> -- 
>>>
>>> --- 
>>> You received this message because you are subscribed to the Google 
>>> Groups "theano-users" group.
>>> To unsubscribe from this group and stop receiving emails from it, send 
>>> an email to [email protected].
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>
>>

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to