[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-18 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-503403627 @apeforest Note that the pending job has succeeded but for some reason it isn't updated here.

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-10 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-500517870 Oops, sorry for confusion. I meant the value of a NDArray. `auto gx = nnvm::NodeEntry{n};`, can I get/print the value of the NDArray

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-08 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-500114215 @apeforest , Thank You for explaining how to get the dump of the graph. Waiting the PR which simplifies that.

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-07 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-499973200 Oh thank you very much for explaining what is happening. I get it now. It makes sense, as passing argument as `head_grads` doesn't make any

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-05 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-499290399 > @kshitij12345 The computation graph for the second backward pass makes sense to me. As you can see there is only one output from the graph,

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-05 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-499271979 > @kshitij12345 I have some question about the equation `expected_head_grad = (grad_op(x) * head_grad_grads).asnumpy()` in your test. >

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-05 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-498950371 > So are you saying that by chain rule, the node is ograd(of log) * d (log(x)) / dx = ograd * reciprocal? Following the unary_bwd

[GitHub] [incubator-mxnet] kshitij12345 commented on issue #15120: [bug] fix higher grad log

2019-06-01 Thread GitBox
kshitij12345 commented on issue #15120: [bug] fix higher grad log URL: https://github.com/apache/incubator-mxnet/pull/15120#issuecomment-497965942 @apeforest @larroy please review This is an automated message from the Apache