From 8ee3a7ee40f528f5c81e59e5b391fd246ae6a235 Mon Sep 17 00:00:00 2001 From: megemini Date: Mon, 16 Dec 2024 14:25:41 +0800 Subject: [PATCH] [Fix] fastspeech2 0d (#3951) --- paddlespeech/t2s/models/fastspeech2/fastspeech2.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/paddlespeech/t2s/models/fastspeech2/fastspeech2.py b/paddlespeech/t2s/models/fastspeech2/fastspeech2.py index fcd54f0d2f4..91bfc540aa9 100644 --- a/paddlespeech/t2s/models/fastspeech2/fastspeech2.py +++ b/paddlespeech/t2s/models/fastspeech2/fastspeech2.py @@ -903,14 +903,14 @@ def _reset_parameters(self, init_enc_alpha: float, init_dec_alpha: float): # initialize alpha in scaled positional encoding if self.encoder_type == "transformer" and self.use_scaled_pos_enc: - init_enc_alpha = paddle.to_tensor(init_enc_alpha) + init_enc_alpha = paddle.to_tensor(init_enc_alpha).reshape([1]) self.encoder.embed[-1].alpha = paddle.create_parameter( shape=init_enc_alpha.shape, dtype=str(init_enc_alpha.numpy().dtype), default_initializer=paddle.nn.initializer.Assign( init_enc_alpha)) if self.decoder_type == "transformer" and self.use_scaled_pos_enc: - init_dec_alpha = paddle.to_tensor(init_dec_alpha) + init_dec_alpha = paddle.to_tensor(init_dec_alpha).reshape([1]) self.decoder.embed[-1].alpha = paddle.create_parameter( shape=init_dec_alpha.shape, dtype=str(init_dec_alpha.numpy().dtype),