-
Notifications
You must be signed in to change notification settings - Fork 19.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BUG: Setting initial state on ConvLSTM2D with input variables #9761
Comments
Hello,
BUT, it seems that _standardize_args is called twice for ConvLSTM2D (once by ConvRNN2D impl and then via I think the solution is to remove the (redundant?)
from Can anybody confirm? |
I had the same problem with mizima, it would be really helpful if someone can confirm this hacky solution. |
We are hitting the same issue with Seq2Seq LSTM model based on https://arxiv.org/pdf/1409.3215.pdf. Issue is seen when creating converting model to tf estimator. Is there a fix planned anytime soon? |
…ut variables Deleted redundant call of `_standardize_args`
I am still running into this bug at the moment. I've tried removing those (seemingly redundant) standardize calls, but haven't had much luck with that method. If this does not work, is there any other method of implementing convolutional seq2seq models in Keras? |
I had met the same AssertionError, thanks for your suggestions, I solved the problems. |
This issue is still not addressed in the latest release. Also, although the proposed solution (below) fixes the model for single GPU training, I still face issues due to the initial_state when training on multiple GPUs using keras.utils.multi_gpu_model .
|
Do we have any update on this issue? In my seq2seq model based on convLSTM2DCell and convRNN2D, I have to pass the encoder_states as initial_state to the decoder (see code below). For now, this
leads to an Assertion error. Here is the function that tries to implement a seq2seq convLSTM2D model, for sequences of video frames as input.
|
Hello, I have the same problem too. Anybody who has a solution? num_input_features = (201,201,1) encoder_inputs = keras.layers.Input(shape=(None, num_input_features[0], num_input_features[1], num_input_features[2])) encoder = ConvLSTM2D(filters=8, kernel_size=(3, 3),input_shape=(None, 201, 201, 1), decoder_inputs = keras.layers.Input(shape=(None, num_output_features[0], num_output_features[1], num_output_features[2])) model = keras.models.Model(inputs=[encoder_inputs, decoder_inputs], outputs=decoder_outputs) Traceback (most recent call last): removing below does not work.. from ConvRNN2D.call. It fixes the AsserionError for me. |
A better fix would follow tf.keras issue.
|
I think the problem comes from line 307 - line 337 in convolutional_recurrent.py. I think a possible fix is not to update kwargs with "initial_state". So here i remove line 308: and things go well |
I meet the same problem with keras==2.3.1. |
Should this work, or is this an unsupported use case?
Thanks!
The text was updated successfully, but these errors were encountered: