-
Notifications
You must be signed in to change notification settings - Fork 525
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Model converted from PT to TF backend could not run with TF #3997
Labels
Comments
|
1 task
1 task
Fixed in #4007. |
njzjz
added a commit
to njzjz/deepmd-kit
that referenced
this issue
Jul 26, 2024
- [x] (Tomorrow) Test if it works for deepmodeling#3997. deepmodeling#3997 needs another fix in deepmodeling#4022 . <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit - **New Features** - Introduced a method to dynamically determine the output dimension of the descriptor, enhancing its functionality and interaction with other components. - Improved tensor dimensionality handling in tests to ensure compatibility with the new output dimension method. <!-- end of auto-generated comment: release notes by coderabbit.ai --> --------- Signed-off-by: Jinzhe Zeng <jinzhe.zeng@rutgers.edu> Co-authored-by: Jinzhe Zeng <jinzhe.zeng@rutgers.edu>
mtaillefumier
pushed a commit
to mtaillefumier/deepmd-kit
that referenced
this issue
Sep 18, 2024
- [x] (Tomorrow) Test if it works for deepmodeling#3997. deepmodeling#3997 needs another fix in deepmodeling#4022 . <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit - **New Features** - Introduced a method to dynamically determine the output dimension of the descriptor, enhancing its functionality and interaction with other components. - Improved tensor dimensionality handling in tests to ensure compatibility with the new output dimension method. <!-- end of auto-generated comment: release notes by coderabbit.ai --> --------- Signed-off-by: Jinzhe Zeng <jinzhe.zeng@rutgers.edu> Co-authored-by: Jinzhe Zeng <jinzhe.zeng@rutgers.edu>
Reopen. #4007 may not fix this issue, which needs more validation. |
#4320 should fix the issue. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Bug summary
I am now working on multi-task training with DeePMD-kit v3.0.0b0, and I get a header with
se_a
descriptor after freezing step. Then, I tried to usedp --pt convert-backend frozen_model.pth frozen_model.pb
(and without--pt
, getting the same result.) to get afrozen_model.pb
. But it could not be used when running Lammps with both v2.2.9 and v3.0.0b0, raising the following error:It seems something wrong when converting the model, and seems to be a bug.
DeePMD-kit Version
DeePMD-kit v3.0.0b0
Backend and its version
PyTorch v2.0.0.post200, TensorFlow v2.14.0
How did you download the software?
Offline packages
Input Files, Running Commands, Error Log, etc.
Running command:
or use
--pt
.And the Lammps error log is under below.
slurm-2623892.txt
Steps to Reproduce
Please use the following
frozen_model.pth
to freeze and use the following Lammps task to reproduce the bug.Further Information, Files, and Links
No response
The text was updated successfully, but these errors were encountered: