This directory contains example scripts to train or fine-tune large scale models,
with the Sagemaker distributed model parallelism library.
When using one of the ipynb notebooks within the folders of this directory please
make sure to use the ./shared-scripts/
directory as the source directory when submitting a job.
For example, if one wanted to submit a llama finetune job on Sagemaker using the /llama_v2/smp-finetuning-llama-fsdp-tp.ipynb
notebook, they would have to copy that notebook within the ./shared-scripts/
directory to make sure it can access all the accompanied files.
After cloning this repository run the following command to setup a copy of the notebook associated with your desired model into the ./shared-scripts/
directory.
cp [RELATIVE PATH TO ipynb] shared-scripts/
Finally, upload the ./shared-scripts/
directory to a Sagemaker notebook to submit your training/finetuning job.