-
Notifications
You must be signed in to change notification settings - Fork 460
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is there a way to find the max GPU memory watermark? How to run locally with minimal setup #17
Comments
The old K40 GPUs (12GB RAM) we have locally ran all but one (it was 900-1000aa) CASP FM target without issues with the official pipeline, so AF2 doesn't necessarily need very new GPUs. You might still want to poke at the python code in the Colab, as this will be a lot easier to supply your own MSAs to than the official pipeline. Ideally we want to make the Colabs also runnable on the command line, but haven't started working on that yet. |
It would be great to have an option for a command-line Colab: eager to run
it on 10E3-10E4 inputs. Thanks in advance.
…On Thu, Aug 5, 2021 at 9:35 AM Milot Mirdita ***@***.***> wrote:
The old K40 GPUs (12GB RAM) we have locally ran all but one (it was
900-1000aa) CASP FM target without issues with the official pipeline, so
AF2 doesn't necessarily need very new GPUs.
You might still want to poke at the python code in the Colab, as this will
be a lot easier to supply your own MSAs to than the official pipeline.
Ideally we want to make the Colabs also runnable on the command line, but
haven't started working on that yet.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#17 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AABGSN34BZ3MV4PXMM4ZNNTT3JEM7ANCNFSM5BTJXEPQ>
.
Triage notifications on the go with GitHub Mobile for iOS
<https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675>
or Android
<https://play.google.com/store/apps/details?id=com.github.android&utm_campaign=notification-email>
.
|
This is also mentioned in #20. It would be great to have either a local command-line interface or the local notebook version so that we can run on inputs with >1000 amino acids and predict complexes (dimers/trimers) of it. I'm not that familiar with all the steps involved in the code. Mostly use it as an end-to-end tool. I tried to localize the |
I ended up biting the bullet, getting a computer with a large enough NVIDIA
GPU and installing Alphafold2 on the Linux machine. The docker version of
the instructions was a pain to get sorted, so I did it with a non-docker
recipe here:
https://github.com/kalininalab/alphafold_non_docker
For other people without the means or access to their own computer with a
large enough NVIDIA GPU, I think it's still worth have a way to expose this
wonderful Colab notebooks to a more programmatic / serial way of
performing the tasks.There is so much work and experimentation going on in
this Colabs, that it's worth keeping an eye on them even for people who
have managed to locally deploy Alphafold2 on their computers.
…On Mon, Aug 23, 2021 at 9:41 AM Roden Luo ***@***.***> wrote:
Ideally we want to make the Colabs also runnable on the command line, but
haven't started working on that yet.
This is also mentioned in #20
<#20>. It would be great to
have either a local command-line interface or the local notebook version so
that we can run on inputs with >1000 amino acids and predict complexes
(dimers/trimers) of it.
I'm not that familiar with all the steps involved in the code. Mostly use
it as an end-to-end tool. I tried to localize the AlphaFold2_advanced
notebook. After solving several package issues, now stuck at No module
named 'colabfold'. I also see the database_path all go to googleapis
which will work fine on Colab but less smoothly on local I guess. I have a
local version of AlphaFold2 running fine. Would be much appreciated if you
can give some hints on how to localize the AlphaFold2_advanced notebook.
Thanks.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#17 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AABGSN2RP5Z62M5W2CK3YO3T6ICTVANCNFSM5BTJXEPQ>
.
Triage notifications on the go with GitHub Mobile for iOS
<https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675>
or Android
<https://play.google.com/store/apps/details?id=com.github.android&utm_campaign=notification-email>
.
|
We now have an internal version that runs on a cluster. The main issue still remains that the MMseqs2 API runs on one single server and will probably not scale to multiple research group submitting jobs. We are still preparing databases, scripts etc. so people could deploy their own server. However, to use MMseqs2 as we use it for ColabFold we do require that all databases are fully in RAM (currently requiring 535GB of RAM + some RAM for each worker process). We can change the local ColabFold version to work with MMseqs2's usual batch mode where the memory requirements are not as high. If you want to run a few thousand sequences please contact me directly (email, twitter etc). I can give you access to the local version. We still need to figure something out how to scale the API better though. |
Thanks! I have a local version of AlphaFold2 installed with docker on a server. (I met some problems during installation. And then I was trying to install the non_docker version on a cluster as well but later dropped it as the one on the server worked out fine after changing the Cuda version.) I have 4 I am facing the below error if I run them on Colab.
I'm trying to run the notebooks locally on the server now. The previously mentioned And now, the
However, all 4 GPUs and the RAM are available as shown below.
Any help would be very appreciated. Thanks |
The advanced notebook is under active development. I would avoid trying to deploy it locally (unless you are willing to track daily bug fixes and implement them yourself). For a more stable setup see alphafold2_mmseqs2 notebook. |
According to the README.md, the memory goes as follows:
I am interested in structures of around either (a) one single chain of 240-280aa or around (b) 2 different chains of ~120 + ~140aa. What would be the minimal GPU that would allow us to run this locally?
I am thinking that given our own custom MSAs, it wouldn't need to connect to MMSeqs2 or download the 2Tb of sequence data, thus going straight into running the prediction based on the MSA of internal data on the docker container?
Or am I missing something obvious that would still require Colab or something else remote?
The text was updated successfully, but these errors were encountered: