Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update llama-3 PEFT notebook to download model from NGC #9667

Merged
merged 8 commits into from
Jul 10, 2024

Conversation

shashank3959
Copy link
Contributor

  • Updates the llama-3 PEFT tutorial to download the model from NVIDIA NGC to skip conversion

@cuichenx cuichenx requested a review from jgerh July 10, 2024 20:13
Copy link
Collaborator

@cuichenx cuichenx left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This looks good to me

@jgerh
Copy link
Collaborator

jgerh commented Jul 10, 2024

I am providing additional copyedits to the file, ReadMe.rst, herein. I was not able to provide inline edits because part of the file was read-only.

Line 4 fix punctuation

Llama 3 <https://blogs.nvidia.com/blog/meta-llama3-inference-acceleration/>_ is an open-source large language model by Meta that delivers state-of-the-art performance on popular industry benchmarks.

Line 6 fix punctuation

Low-Rank Adaptation (LoRA) <https://arxiv.org/pdf/2106.09685>__ has emerged as a popular Parameter-Efficient Fine-Tuning (PEFT) technique that tunes a very small number of additional parameters as compared to full fine-tuning, thereby reducing the compute required.

Line 19 revise for clarity

NIM enables seamless deployment of multiple LoRA adapters (referred to as ‘multi-LoRA’) on the same base model. It dynamically loads adapter weights based on incoming requests at runtime. This flexibility allows handling inputs from various tasks or use cases without requiring a unique model for each individual scenario. For further details, consult the NIM documentation for LLMs.

Line 24 delete the following sentence, it is sufficient to retain the bullets

In order to proceed, ensure that you have met the following requirements:

Line 103/109 fix punctuation

  1. Prepare the LoRA model store.

Line 110/116 fix grammar

To ensure the model store is organized as expected, create a folder named llama3-8b-pubmed-qa and move your .nemo checkpoint there.

Line 122/128 revise for clarity

Ensure that the LoRA model store directory follows this structure: the model name should be a sub-folder containing the .nemo file.

Line 138/144 fix punctuation

  1. Set up NIM.

Line 170/176 fix grammar

The first time you run the command, it will download the model and cache it in $NIM_CACHE_PATH so subsequent deployments are even faster. There are several options to configure NIM other than the ones listed above. You can find a full list in the NIM configuration documentation.

Line 173/179 fix punctuation

  1. Start the notebook.

Line 175/181 fix grammar

From another terminal, follow the same instructions as the previous notebook to launch Jupyter Lab, and then navigate to this notebook.

Line 178/185 revise for clarity

You can use the same NeMo Framework docker container in which Jupyter Lab was previously installed.

Signed-off-by: Shashank Verma <shashank3959@gmail.com>
Signed-off-by: Shashank Verma <shashank3959@gmail.com>
Signed-off-by: Shashank Verma <shashank3959@gmail.com>
Copy link
Collaborator

@cuichenx cuichenx left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@pablo-garay pablo-garay merged commit 3ab0a2a into NVIDIA:main Jul 10, 2024
8 checks passed
marcromeyn pushed a commit that referenced this pull request Jul 11, 2024
* Update llama-3 PEFT notebook to download model from NGC

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link in llama-3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken code block in llama 3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Copy-edits to Llama-3 8B PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Minor formatting fixes

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

---------

Signed-off-by: Shashank Verma <shashank3959@gmail.com>
marcromeyn pushed a commit that referenced this pull request Jul 11, 2024
* Update llama-3 PEFT notebook to download model from NGC

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link in llama-3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken code block in llama 3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Copy-edits to Llama-3 8B PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Minor formatting fixes

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

---------

Signed-off-by: Shashank Verma <shashank3959@gmail.com>
maanug-nv pushed a commit that referenced this pull request Jul 14, 2024
* Update llama-3 PEFT notebook to download model from NGC

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link in llama-3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken code block in llama 3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Copy-edits to Llama-3 8B PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Minor formatting fixes

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

---------

Signed-off-by: Shashank Verma <shashank3959@gmail.com>
ertkonuk pushed a commit that referenced this pull request Jul 19, 2024
* Update llama-3 PEFT notebook to download model from NGC

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link in llama-3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken code block in llama 3 PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Copy-edits to Llama-3 8B PEFT tutorial README

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Fix broken link

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

* Minor formatting fixes

Signed-off-by: Shashank Verma <shashank3959@gmail.com>

---------

Signed-off-by: Shashank Verma <shashank3959@gmail.com>
Signed-off-by: Tugrul Konuk <ertkonuk@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants