Skip to content
This repository has been archived by the owner on Jun 18, 2024. It is now read-only.

Make bulk inference easier & output internal embeddings #1

Open
wants to merge 19 commits into
base: main
Choose a base branch
from

Commits on Oct 6, 2021

  1. Changes to make bulk inference easier:

    - load in features if already computed
    - allow computing features for multiple fastas in parallel
    - allow disabling Amber relaxation, b/c our sequences have X's
    Prasanna Vasudevan authored and Prasanna Vasudevan committed Oct 6, 2021
    Configuration menu
    Copy the full SHA
    1bee52e View commit details
    Browse the repository at this point in the history
  2. Output internal embeddings as part of result_model_*.pkl

    Prasanna Vasudevan authored and Prasanna Vasudevan committed Oct 6, 2021
    Configuration menu
    Copy the full SHA
    9c00671 View commit details
    Browse the repository at this point in the history

Commits on Oct 7, 2021

  1. Minor changes

    Prasanna Vasudevan committed Oct 7, 2021
    Configuration menu
    Copy the full SHA
    c98b294 View commit details
    Browse the repository at this point in the history
  2. Noticed that re-using compiled Jax model is 2-10x faster! So now,

    sorting fastas by seq len and padding to round up to nearest 50, so
    many consecutive sequences can be run by the same compiled Jax model.
    
    Note that different models (1 through 5) have different sizes
    (e.g. extra msa size), so for now you must choose exactly one model
    in run_docker.py. In the future, may want to refactor run_alphafold.py
    so the outer loop is over models instead of inputs.
    Prasanna Vasudevan committed Oct 7, 2021
    Configuration menu
    Copy the full SHA
    74beba9 View commit details
    Browse the repository at this point in the history

Commits on Oct 9, 2021

  1. Run inference on multiple GPUs in parallel.

    On a related note, tried running multiple inputs on a single GPU in separate child processes,
    but that was actually slower than doing it serially.
    Prasanna Vasudevan committed Oct 9, 2021
    Configuration menu
    Copy the full SHA
    0c23382 View commit details
    Browse the repository at this point in the history

Commits on Oct 11, 2021

  1. Refactor (WIP: need to test):

    - split up feature generation vs model running
    - run one model at a time, thru all inputs (for Jax compile speedup),
      and then rank models at the end.
    - don't run a model if output already exists
    Prasanna Vasudevan authored and Prasanna Vasudevan committed Oct 11, 2021
    Configuration menu
    Copy the full SHA
    cdaf2fa View commit details
    Browse the repository at this point in the history

Commits on Jan 27, 2022

  1. Add line in run_docker to remind users to mount repo dir to avoid Doc…

    …ker re-build on code changes
    Prasanna Vasudevan committed Jan 27, 2022
    Configuration menu
    Copy the full SHA
    803a11d View commit details
    Browse the repository at this point in the history

Commits on Jan 28, 2022

  1. Fix jax and jaxlib versions getting out of sync

    Prasanna Vasudevan committed Jan 28, 2022
    Configuration menu
    Copy the full SHA
    3b882f7 View commit details
    Browse the repository at this point in the history

Commits on Jan 29, 2022

  1. Don't fail if can't find a PDB .cif template file

    (workaround for now, in the interest of time)
    Prasanna Vasudevan committed Jan 29, 2022
    Configuration menu
    Copy the full SHA
    8edaafd View commit details
    Browse the repository at this point in the history

Commits on Jan 30, 2022

  1. Don't run a model if output already exists

    Prasanna Vasudevan committed Jan 30, 2022
    Configuration menu
    Copy the full SHA
    f459ad4 View commit details
    Browse the repository at this point in the history
  2. Minor change, forgot to commit earlier

    Prasanna Vasudevan committed Jan 30, 2022
    Configuration menu
    Copy the full SHA
    0d737df View commit details
    Browse the repository at this point in the history

Commits on Feb 5, 2022

  1. After refactor of run_alphafold, forgot to re-enable ranking of model…

    …s at the end. Do that now.
    Prasanna Vasudevan committed Feb 5, 2022
    Configuration menu
    Copy the full SHA
    ab59843 View commit details
    Browse the repository at this point in the history

Commits on Feb 16, 2022

  1. Average embeddings across amino acids and output to a separate, conve…

    …nient file
    Prasanna Vasudevan committed Feb 16, 2022
    Configuration menu
    Copy the full SHA
    cf48f3e View commit details
    Browse the repository at this point in the history
  2. Minor logging fix

    Prasanna Vasudevan committed Feb 16, 2022
    Configuration menu
    Copy the full SHA
    69a289a View commit details
    Browse the repository at this point in the history

Commits on Feb 21, 2022

  1. AlphaFold runner: skip model if result .pkl file OR just the small em…

    …beddings .pkl is outputted.
    
    Also, when ranking models, grab plddt from the small embeddings .pkl file, not the huge result .pkl file.
    Prasanna Vasudevan committed Feb 21, 2022
    Configuration menu
    Copy the full SHA
    fc79d58 View commit details
    Browse the repository at this point in the history

Commits on Feb 22, 2022

  1. Catch exception in run_model instead of exiting

    Prasanna Vasudevan committed Feb 22, 2022
    Configuration menu
    Copy the full SHA
    74d245e View commit details
    Browse the repository at this point in the history
  2. Fix outputting of embeddings / pLDDT

    Prasanna Vasudevan committed Feb 22, 2022
    Configuration menu
    Copy the full SHA
    3241dec View commit details
    Browse the repository at this point in the history
  3. Output full, per-token embeddings in addition to average-token embedd…

    …ings
    Prasanna Vasudevan committed Feb 22, 2022
    Configuration menu
    Copy the full SHA
    9e5dd19 View commit details
    Browse the repository at this point in the history

Commits on Mar 31, 2022

  1. Catch exception in generating features, rather than failing.

    Prasanna Vasudevan committed Mar 31, 2022
    Configuration menu
    Copy the full SHA
    48f9776 View commit details
    Browse the repository at this point in the history