From b58eef036154eff510fafd0b330555fb2a858549 Mon Sep 17 00:00:00 2001 From: maxsamazon <129529238+maxsamazon@users.noreply.github.com> Date: Fri, 29 Aug 2025 15:18:12 +0100 Subject: [PATCH] Update 00-upload-training-data.md Removed link to wikitext data location. --- content/09-ml-on-parallelcluster/00-upload-training-data.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/content/09-ml-on-parallelcluster/00-upload-training-data.md b/content/09-ml-on-parallelcluster/00-upload-training-data.md index df48bdda..4177e371 100644 --- a/content/09-ml-on-parallelcluster/00-upload-training-data.md +++ b/content/09-ml-on-parallelcluster/00-upload-training-data.md @@ -12,7 +12,7 @@ In this step, you create an environment configuration script to train a Natural First, create an Amazon S3 bucket and upload the training data folder. This training folder will be accessed by the cluster worker nodes through FSx. 1. Open a terminal in your AWS Cloud9 instance. -2. Run the following commands to create a new Amazon S3 bucket. These commands also retrieve and store the [Wikitext 103 dataset](https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/) +2. Run the following commands to create a new Amazon S3 bucket. These commands also retrieve and store the Wikitext 103 dataset ```bash # generate a unique postfix @@ -21,7 +21,7 @@ echo "Your bucket name will be mlbucket-${BUCKET_POSTFIX}" aws s3 mb s3://mlbucket-${BUCKET_POSTFIX} # downloading data: -export URL="https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-103-v1.zip" +export URL="https://path/to/wikitext-103-v1.zip" export FILE="wikitext-103-v1.zip" wget $URL -O $FILE unzip $FILE