Skip to content

Conversation

sean-smith
Copy link

@sean-smith sean-smith commented Aug 24, 2025

Fixes #821, #820, #824

@sean-smith sean-smith requested a review from a team as a code owner August 24, 2025 23:25
Copy link

github-actions bot commented Aug 24, 2025

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@sean-smith sean-smith changed the title Setup git dir in /workspace/llama31 Fixes for llama 3.1 training Aug 24, 2025
@sean-smith
Copy link
Author

recheck

@ShriyaRishab
Copy link
Contributor

ShriyaRishab commented Aug 26, 2025

Added comments on the issues, can you please revert the git based changes - the other 2 look good.

@sean-smith
Copy link
Author

Added comments on the issues, can you please revert the git based changes - the other 2 look good.

Done ✅

* Remove subpath from pretrain_llama.py
* Install toml package
* Adjust --gres=gpu:8 to number of user specified devices

Signed-off-by: Sean Smith <seasmith@nvidia.com>
mem="0",
exclusive=True,
gres="gpu:8",
packager=run.GitArchivePackager(subpath="large_language_model_pretraining/nemo", ref="HEAD"),
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you please revert packager=run.GitArchivePackager(subpath="large_language_model_pretraining/nemo", ref="HEAD"), as well?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I can however I don't understand who is able to run this. If you follow the instructions in the readme this will fail since the path is wrong.

Maybe if you move the Dockerfile to the root directory and build there then this will work, i.e.

cp Dockerfile ../..
cd ../..
docker build -t nemo .

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

LLama 3.1 toml package missing
2 participants