Issues with VEP Script #18
-
Title: Using Hugging Face Transformers Hello, I hope this post finds you well. I'm currently facing challenges running the VEP inference script. I've faced an error related to mixed precision training when attempting to run the script on GPU. As an alternative, I'm interested in exploring the possibility of running the code on CPU instead. This is the error I get: ValueError: FP16 Mixed precision training with AMP or APEX ( torch.distributed.elastic.multiprocessing.errors.ChildFailedError: I've successfully run the simple_example script, indicating that the library is working well in my environment. Thank you in advance for any guidance or insights you can provide! |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
Hi, sorry for your troubles! fp16 should certainly be an optional argument. For now, you can remove this line or set to False: Line 39 in 05b23c5 Additionally, in the notebook you might need to replace {torchrun_path} --nproc_per_node={n_gpu} with {python_path} (which could just be python )Let me know if you get it to work. |
Beta Was this translation helpful? Give feedback.
-
Hello gonzalo, |
Beta Was this translation helpful? Give feedback.
Hi, sorry for your troubles! fp16 should certainly be an optional argument. For now, you can remove this line or set to False:
gpn/gpn/msa/inference.py
Line 39 in 05b23c5
Additionally, in the notebook you might need to replace
{torchrun_path} --nproc_per_node={n_gpu}
with{python_path}
(which could just bepython
)Let me know if you get it to work.