generated from amazon-archives/__template_MIT-0
-
Notifications
You must be signed in to change notification settings - Fork 1k
New Serverless Pattern - Selective file transfer between SFTP server and S3 using AWS Transfer Family #2267
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
julianwood
merged 7 commits into
aws-samples:main
from
biswanathmukherjee:biswanathmukherjee-feature-sftp-connector
May 20, 2024
Merged
Changes from all commits
Commits
Show all changes
7 commits
Select commit
Hold shift + click to select a range
fd33490
Initial checkin
15d5425
fixed review comments
46b9116
fixed review comments
a49c8fb
fixed review comments
e2850f7
Add final pattern file
bfreiberg 9489ab4
Update awstransfer-s3-sam.json
julianwood 299df14
Update example-pattern.json
julianwood File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,165 @@ | ||
# Bidirectional selective file transfer between remote SFTP server and Amazon S3 using AWS Transfer Family Connector | ||
|
||
This pattern shows how to setup an AWS Transfer Family SFTP connector to list files from the remote server and transfer specific files to Amazon S3 bucket. You can also transfer specific files from Amazon S3 bucket to the remote SFTP server. | ||
|
||
Learn more about this pattern at Serverless Land Patterns: https://serverlessland.com/patterns/awstransfer-s3-sam. | ||
|
||
Important: this application uses various AWS services and there are costs associated with these services after the Free Tier usage - please see the [AWS Pricing page](https://aws.amazon.com/pricing/) for details. You are responsible for any AWS costs incurred. No warranty is implied in this example. | ||
|
||
## Requirements | ||
|
||
* [Create an AWS account](https://portal.aws.amazon.com/gp/aws/developer/registration/index.html) if you do not already have one and log in. The IAM user that you use must have sufficient permissions to make necessary AWS service calls and manage AWS resources. | ||
* [AWS CLI](https://docs.aws.amazon.com/cli/latest/userguide/install-cliv2.html) installed and configured | ||
* [Git Installed](https://git-scm.com/book/en/v2/Getting-Started-Installing-Git) | ||
* [AWS Serverless Application Model](https://docs.aws.amazon.com/serverless-application-model/latest/developerguide/serverless-sam-cli-install.html) (AWS SAM) installed | ||
* [JQ](https://docs.aws.amazon.com/solutions/latest/dynamic-object-and-rule-extensions-for-aws-network-firewall/operation-and-customization.html#install-jq) should be installed. | ||
|
||
|
||
|
||
## Deployment Instructions | ||
|
||
1. Create a new directory, navigate to that directory in a terminal and clone the GitHub repository: | ||
``` | ||
git clone https://github.com/aws-samples/serverless-patterns | ||
``` | ||
|
||
2. Change directory to the pattern directory: | ||
``` | ||
cd serverless-patterns/awstransfer-s3-sam | ||
``` | ||
|
||
3. From the command line, run the below command to deploy the pattern: | ||
``` | ||
bash deploy.sh | ||
``` | ||
|
||
4. During the prompts: | ||
* Enter a stack name | ||
* Enter the desired AWS Region (e.g. us-east-1) | ||
|
||
5. The deployment script deploys both `template-sftp-server.yaml` and `template-sftp-connector.yaml`. Please make a note of the output both the deployments as they will be used during testing. | ||
|
||
|
||
## How it works | ||
|
||
Please refer to the architecture diagram below: | ||
|
||
 | ||
|
||
* The remote SFTP server is simulated using AWS Transfer Family SFTP Server for this pattern. In a real use case, this can be any remote SFTP server outside of AWS. | ||
* SFTP Connector is configured to connect to the remote server with Amazon S3 bucket using SFTP protocol. The authentication is done using SSH Key based handshake. | ||
* Amazon S3 bucket is used for file storage on the AWS side. | ||
* User can list files on the remote server and selectively transfer files from the remote server to the Amazon S3 bucket using AWS Transfer Family API or CLI commands. | ||
* User can also transfer files from Amazon S3 to the remote server using the AWS Transfer Family API or CLI commands. | ||
|
||
## Testing | ||
|
||
1. Use the endpoint to test the SFTP server with transferring a file using a client, the rest of this test steps are shown using OpenSSH. Please refer to [Transferring files over a server endpoint using a client](https://docs.aws.amazon.com/transfer/latest/userguide/transfer-file.html) for other options. | ||
|
||
2. Test the connection using using the below command from your command line. Please replace `SFTPTransferConnector` from the deployment output: | ||
```bash | ||
aws transfer test-connection --region {your-region} --connector-id {SFTPTransferConnector} | ||
``` | ||
It should give an output similar to below: | ||
```json | ||
{ | ||
"Status": "OK", | ||
"StatusMessage": "Connection succeeded" | ||
} | ||
``` | ||
|
||
3. Transfer `sample1.txt` and `sample2.txt` files to the remote SFTP server (similated) using the below commands. In this sample project replace `SSHPrivateKeyFileName` and `SSHPrivateKeyFileName` with `sftpuser`. Replace the value of `TransferServerEndpoint` from the deployment output: | ||
```bash | ||
sftp -i {SSHPrivateKeyFileName} {TransferServerUser}@{TransferServerEndpoint} | ||
pwd | ||
mkdir Remote | ||
cd Remote | ||
put sample1.txt | ||
put sample2.txt | ||
ls | ||
``` | ||
Confirm to proceed with the connection after the first command. | ||
|
||
5. List the files on the remote SFTP server using the below command. Please replace `SFTPTransferConnector` and `MyLocalS3Bucket` from the deploy output: | ||
|
||
```bash | ||
aws transfer start-directory-listing --region {your-region} --connector-id {SFTPTransferConnector} --remote-directory-path /Remote --output-directory-path /{MyLocalS3Bucket}/FromRemoteSFTPServer | ||
``` | ||
|
||
The command invokes an asynchronous API. The output of the command will be as follows: | ||
```json | ||
{ | ||
"ListingId": "273e5b33-xxxx-xxxx-xxxx-xxxxx9a507f53", | ||
"OutputFileName": "c-cxxxxxxxx-xxxxxx-xxxx-xxxx-xxxx-xxxxxa507f53.json" | ||
} | ||
``` | ||
|
||
6. Log into [Amazon S3 console](https://console.aws.amazon.com/s3). Open the `MyLocalS3Bucket` and navigate to `FromRemoteSFTPServer` folder. Check the content of the JSON file. It should look something like below: | ||
```json | ||
{ | ||
"files": [ | ||
{ | ||
"filePath": "/Remote/sample1.txt", | ||
"modifiedTimestamp": "2024-04-28T07:47:27Z", | ||
"size": 146 | ||
}, | ||
{ | ||
"filePath": "/Remote/sample2.txt", | ||
"modifiedTimestamp": "2024-04-28T07:47:46Z", | ||
"size": 146 | ||
} | ||
], | ||
"paths": [], | ||
"truncated": false | ||
} | ||
``` | ||
|
||
7. Transfer one of the files from the remote SFTP server to the Amazon S3 bucket using the following command: | ||
```bash | ||
aws transfer start-file-transfer --region {your-region} --connector-id {SFTPTransferConnector} --retrieve-file-paths /Remote/sample1.txt --local-directory-path /{MyLocalS3Bucket}/FromRemoteSFTPServer | ||
``` | ||
|
||
The output of the command should something like below: | ||
```json | ||
{ | ||
"TransferId": "e863xxxx-xxxx-xxxx-xxxx-xxxxa40c5ff9" | ||
} | ||
``` | ||
|
||
8. Log into [Amazon S3 console](https://console.aws.amazon.com/s3). Open the `MyLocalS3Bucket` and navigate to `FromRemoteSFTPServer` folder. You should be able to find the transferred `sample1.txt` file. | ||
|
||
9. Upload a file into the `MyLocalS3Bucket` bucket using the following command. Replace the `MyLocalS3Bucket` from the deploy output: | ||
```bash | ||
aws s3 cp sample3.txt s3://{MyLocalS3Bucket}/local/sample3.txt | ||
``` | ||
|
||
10. Transfer the `sample3.txt` file Amazon S3 `MyLocalS3Bucket` bucket to the remove SPT server using the following command: | ||
```bash | ||
aws transfer start-file-transfer --region {your-region} --connector-id {SFTPTransferConnector} --send-file-paths /{MyLocalS3Bucket}/local/sample3.txt --remote-directory-path /FromAmazonS3 | ||
``` | ||
|
||
11. Validate the file transfer by logging into the the remove SFTP server using the below commands: | ||
```bash | ||
sftp -i {SSHPrivateKeyFileName} {TransferServerUser}@{TransferServerEndpoint} | ||
ls | ||
cd FromAmazonS3 | ||
ls | ||
``` | ||
|
||
## Cleanup | ||
|
||
1. Delete the content in the Amazon S3 bucket using the following command. Please *ensure* that the correct bucket name is provided to avoid accidental data loss: | ||
```bash | ||
aws s3 rm s3://{MySFTPServerS3Bucket} --recursive --region {my-region} | ||
aws s3 rm s3://{MyLocalS3Bucket} --recursive --region {my-region} | ||
``` | ||
|
||
2. Delete the stack | ||
```bash | ||
bash undeploy.sh | ||
``` | ||
|
||
---- | ||
Copyright 2024 Amazon.com, Inc. or its affiliates. All Rights Reserved. | ||
|
||
SPDX-License-Identifier: MIT-0 |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,82 @@ | ||
{ | ||
"title": "Selective file transfer between SFTP server & Amazon S3 using AWS Transfer Family", | ||
"description": "This pattern shows how to use AWS Transfer Family to list and transfer specific files between an SFTP server and Amazon S3 bucket.", | ||
"language": "YAML", | ||
"level": "200", | ||
"framework": "SAM", | ||
"introBox": { | ||
"headline": "How it works", | ||
"text": [ | ||
"The remote SFTP server is simulated using AWS Transfer Family SFTP Server for this pattern. In a real use case, this can be any remote SFTP server outside of AWS.", | ||
"SFTP Connector is configured to connect to the remote server with Amazon S3 bucket using SFTP protocol. The authentication is done using SSH Key based handshake.", | ||
"Amazon S3 bucket is used for file storage on the AWS side.", | ||
"User can list files on the remote server and selectively transfer files from the remote server to the Amazon S3 bucket using AWS Transfer Family API or CLI commands.", | ||
"User can also transfer files from Amazon S3 to the remote server using the AWS Transfer Family API or CLI commands." | ||
] | ||
}, | ||
"gitHub": { | ||
"template": { | ||
"repoURL": "https://github.com/aws-samples/serverless-patterns/tree/main/awstransfer-s3-sam", | ||
"templateURL": "serverless-patterns/awstransfer-s3-sam", | ||
"projectFolder": "awstransfer-s3-sam", | ||
"templateFile": "template-sftp-server.yaml" | ||
} | ||
}, | ||
"resources": { | ||
"bullets": [ | ||
{ | ||
"text": "Getting started with AWS Transfer Family server endpoints", | ||
"link": "https://docs.aws.amazon.com/transfer/latest/userguide/getting-started.html" | ||
}, | ||
{ | ||
"text": "Configure SFTP connectors", | ||
"link": "https://docs.aws.amazon.com/transfer/latest/userguide/configure-sftp-connector.html" | ||
} | ||
] | ||
}, | ||
"deploy": { | ||
"text": [ | ||
"See the GitHub repo for detailed deployment instructions.", | ||
"bash deploy.sh" | ||
] | ||
}, | ||
"testing": { | ||
"text": [ | ||
"See the GitHub repo for detailed testing instructions." | ||
] | ||
}, | ||
"cleanup": { | ||
"text": [ | ||
"Delete the Amazon S3 input bucket content: <code>aws s3 rm s3://{MySFTPServerS3Bucket} --recursive --region {my-region}</code>", | ||
"Delete the Amazon S3 output bucket content: <code>aws s3 rm s3://{MyLocalS3Bucket} --recursive --region {my-region}</code>", | ||
"<code>bash undeploy.sh</code>" | ||
] | ||
}, | ||
"authors": [ | ||
{ | ||
"name": "Biswanath Mukherjee", | ||
"image": "https://d1rwvjey2iif32.cloudfront.net", | ||
"bio": "I am a Sr. Solutions Architect working at AWS India.", | ||
"linkedin": "biswanathmukherjee" | ||
} | ||
], | ||
"patternArch": { | ||
"icon1": { | ||
"x": 20, | ||
"y": 50, | ||
"service": "transfer", | ||
"label": "Transfer Family server" | ||
}, | ||
"icon2": { | ||
"x": 80, | ||
"y": 50, | ||
"service": "s3", | ||
"label": "S3 bucket" | ||
}, | ||
"line1": { | ||
"from": "icon1", | ||
"to": "icon2", | ||
"label": "" | ||
} | ||
} | ||
} |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,90 @@ | ||
#!/bin/bash | ||
|
||
# Take the stack name | ||
echo "Enter a stack name" | ||
read -r STACK_NAME | ||
|
||
# Take the desired AWS Region | ||
echo "Enter the desired AWS Region:" | ||
read -r AWS_REGION | ||
|
||
|
||
USER_NAME="sftpuser" | ||
|
||
# Generate key-pair | ||
# AWS Documentation: https://docs.aws.amazon.com/transfer/latest/userguide/configure-sftp-connector.html#format-sftp-connector-key | ||
ssh-keygen -t rsa -b 4096 -m PEM -f $USER_NAME -N "" | ||
|
||
# Check if the public key file exists | ||
if [ -f "$USER_NAME.pub" ]; then | ||
|
||
# Store the content of the public key in a variable | ||
PUBLIC_KEY=$(cat "$USER_NAME.pub") | ||
|
||
# Deploy template-sftp-server.yaml | ||
sam deploy \ | ||
--template-file template-sftp-server.yaml \ | ||
--stack-name "$STACK_NAME-1" \ | ||
--parameter-overrides "UserName=\"$USER_NAME\"" "SSHPublicKey=\"$PUBLIC_KEY\"" \ | ||
--capabilities CAPABILITY_IAM \ | ||
--region $AWS_REGION | ||
|
||
# Get the stack ID | ||
STACK_ID=$(aws cloudformation list-stacks --stack-status-filter CREATE_COMPLETE --query "StackSummaries[?contains(StackName, '$STACK_NAME-1')].StackId" --output text --region $AWS_REGION) | ||
|
||
# Check if the stack ID is empty | ||
if [ -z "$STACK_ID" ]; then | ||
echo "Stack not found. Exiting..." | ||
exit 1 | ||
fi | ||
|
||
# Get the stack outputs | ||
OUTPUTS=$(aws cloudformation describe-stacks --stack-name "$STACK_ID" --query "Stacks[0].Outputs" --output json --region $AWS_REGION) | ||
|
||
# Get a TransferServerId output value | ||
TRANSFER_SERVER_ID=$(echo "$OUTPUTS" | jq -r '.[] | select(.OutputKey == "TransferServerId") | .OutputValue') | ||
|
||
# Get a TransferServerEndpoint output value | ||
TRANSFER_SERVER_ENDPOINT=$(echo "$OUTPUTS" | jq -r '.[] | select(.OutputKey == "TransferServerEndpoint") | .OutputValue') | ||
|
||
# Get a TransferLoggingRoleArn output value | ||
TRANSFER_LOGGING_ROLE_ARN=$(echo "$OUTPUTS" | jq -r '.[] | select(.OutputKey == "TransferLoggingRoleArn") | .OutputValue') | ||
|
||
# Get a SSHPrivateKey in single line without double quotes | ||
# AWS Documentation: https://docs.aws.amazon.com/transfer/latest/userguide/sftp-connectors-tutorial.html | ||
FORMATTED_PK=$(jq -sR . < "$USER_NAME"| sed 's/^"//;s/"$//') | ||
|
||
# Wait for the server to be ready | ||
STATE="NOT_AVAILABLE" | ||
|
||
# Loop until the server is available | ||
while [ "$STATE" != "ONLINE" ]; do | ||
# Get the server state using the AWS CLI | ||
STATE=$(aws transfer describe-server --server-id "$TRANSFER_SERVER_ID" --query "Server.State" --output text) | ||
|
||
# Print the server state | ||
echo "Server state: $STATE" | ||
|
||
# Wait for 1 minute before checking again | ||
sleep 60 | ||
done | ||
|
||
# Print a message when the server is available | ||
echo "Server is online! Proceesing with the next steps..." | ||
|
||
# Get the TrustedHostKey from the TransferServer | ||
# AWS Documentation: https://docs.aws.amazon.com/transfer/latest/userguide/API_SftpConnectorConfig.html | ||
TRUSTED_HOST_KEY=$(ssh-keyscan $TRANSFER_SERVER_ENDPOINT) | ||
|
||
# Deploy template-sftp-connector.yaml | ||
sam deploy \ | ||
--template-file template-sftp-connector.yaml \ | ||
--stack-name "$STACK_NAME-2" \ | ||
--parameter-overrides "TransferServerEndpoint=\"sftp://$TRANSFER_SERVER_ENDPOINT\"" "UserName=\"$USER_NAME\"" "TransferLoggingRoleArn=\"$TRANSFER_LOGGING_ROLE_ARN\"" "SSHPrivateKey=\"$FORMATTED_PK\"" "TrustedHostKeys=\"$TRUSTED_HOST_KEY\"" \ | ||
--capabilities CAPABILITY_IAM \ | ||
--region $AWS_REGION | ||
|
||
else | ||
echo "Public key file not found. Exiting..." | ||
exit 1 | ||
fi |
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
How would a customer using Windows do this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Customer using Windows system can use any command line tool that support using using bash script. I used to use GitBash but there may other alternative options also:
https://stackoverflow.com/questions/26522789/how-to-run-sh-on-windows-command-prompt