You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In some cases, SoftmaxGrad is not optimized away and still part of the final graph, that happens in particular in RNN (and similar models), presumably when the softmax is done inside the scan.
A related problem is that the computation of the log-probability is not stabilized numerically, which is why we have to compute log(prop + eps) in the LSTM deep learning tutorial (https://github.com/lisa-lab/DeepLearningTutorials/blob/master/code/lstm.py#L336).
The text was updated successfully, but these errors were encountered:
In some cases, SoftmaxGrad is not optimized away and still part of the final graph, that happens in particular in RNN (and similar models), presumably when the softmax is done inside the scan.
A related problem is that the computation of the log-probability is not stabilized numerically, which is why we have to compute
log(prop + eps)
in the LSTM deep learning tutorial (https://github.com/lisa-lab/DeepLearningTutorials/blob/master/code/lstm.py#L336).The text was updated successfully, but these errors were encountered: