Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Need to use swap memory for loading (sdxl turbo) model, But I can't set it in sagemaker #858

Open
Suprhimp opened this issue Apr 2, 2024 · 3 comments

Comments

@Suprhimp
Copy link

Suprhimp commented Apr 2, 2024

Hi, I found that I can't load my model to sagemaker because inside the docker, my model cause OOM while loading in side the sagemaker deploy docker.

I checked that with My EC2 instance inf2.xlarge works well in this environement. (logged with free -m)

this environment the works very well with sdxl turbo
스크린샷 2024-04-03 오전 2 04 45

this environment gives me OOM error.
스크린샷 2024-04-03 오전 2 15 26

Is there any setting that I can use swap memory? especially I want to know How can I allocate swap memory (sagemaker docker env can't use --privilegeflag).

@jeffhataws
Copy link
Contributor

Hi @Suprhimp, will you help provide more information:

  • Version of docker image
  • Are you using SageMaker Notebooks or SageMaker Studio?

@Suprhimp
Copy link
Author

https://github.com/aws/deep-learning-containers/blob/master/huggingface/pytorch/inference/docker/1.13/py3/sdk2.15.0/Dockerfile.neuronx

I used this docker image for sagemaker endpoint

I finally give up to use sdxl with neuronx in sagemaker environment 😂

But I build my backend with ec2.

@aws-taylor
Copy link
Contributor

Hello @Suprhimp,

Since this appears to be an issue with SageMaker itself, I suggest you reach out to https://repost.aws/tags/questions/TAT80swPyVRPKPcA0rsJYPuA?view=all.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants