Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Release v0.4.0 #60

Merged
merged 1 commit into from
Jun 16, 2023
Merged

Release v0.4.0 #60

merged 1 commit into from
Jun 16, 2023

Conversation

kmaziarz
Copy link
Collaborator

This PR edits the CHANGELOG to mark the v0.4.0 release. After it's merged, I will tag the resulting commit, and then push the release to PyPI.

In particular, this release includes several changes that make molecule_generation compatible with the latest versions of rdkit and numpy.

@kmaziarz kmaziarz merged commit d243e6a into main Jun 16, 2023
5 checks passed
@kmaziarz kmaziarz deleted the kmaziarz/release-0.4.0 branch June 16, 2023 17:17
@thk9178
Copy link

thk9178 commented Jun 30, 2023

Thank you for publishing your great work.
In addition to VAE, I would like to supervised train my SMILES-property data, do you have a manual for the program?
If not, how can I train it with my property data?

@kmaziarz
Copy link
Collaborator Author

In addition to VAE, I would like to supervised train my SMILES-property data, do you have a manual for the program? If not, how can I train it with my property data?

Thank you for your question @thk9178, and sorry for the slow response. This package only provides the code to train the VAE, which would typically be done in a property-agnostic way. If you have some property data, there are essentially two routes:

  • Train a separate supervised property predictor, and use it to perform optimization in the latent space of a trained MoLeR model (for the optimization part itself you could use things like Bayesian Optimisation, Molecular Swarm Optimisation, Genetic Algorithms, etc).
  • Take MoLeR trained in a property-agnostic way and then fine-tune it further on samples with high property values. Then "property optimization" can be performed by sampling from the prior directly and ranking the results. We haven't explored this direction much though.

If you take the first route, then you need to train a supervised property prediction model, for which you'd have to look elsewhere. A simple starting point would be to just train a shallow MLP on molecular fingerprints.

@thk9178
Copy link

thk9178 commented Jul 14, 2023

In addition to VAE, I would like to supervised train my SMILES-property data, do you have a manual for the program? If not, how can I train it with my property data?

Thank you for your question @thk9178, and sorry for the slow response. This package only provides the code to train the VAE, which would typically be done in a property-agnostic way. If you have some property data, there are essentially two routes:

  • Train a separate supervised property predictor, and use it to perform optimization in the latent space of a trained MoLeR model (for the optimization part itself you could use things like Bayesian Optimisation, Molecular Swarm Optimisation, Genetic Algorithms, etc).
  • Take MoLeR trained in a property-agnostic way and then fine-tune it further on samples with high property values. Then "property optimization" can be performed by sampling from the prior directly and ranking the results. We haven't explored this direction much though.

If you take the first route, then you need to train a supervised property prediction model, for which you'd have to look elsewhere. A simple starting point would be to just train a shallow MLP on molecular fingerprints.

I'm so grateful for your help. I can definitely do that as you recommended.
I'll have to pray that MoLer's latent space shows a smooth gradient for my property data. 😁
Thank you very much for your kind reply.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants