Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

New version 1.3.0 leads to "too many open files" while scanning bigger images #2819

Closed
willem-delbare opened this issue Apr 28, 2024 · 1 comment · Fixed by #2823
Closed
Assignees
Labels
bug Something isn't working

Comments

@willem-delbare
Copy link

willem-delbare commented Apr 28, 2024

What happened:
Images that used to scan successfully under < 1.3.0 are no longer scannable by Syft 1.3.0

failed to run task: unable to catalog file digests: failed to process file "": digests-cataloger unable to observe contents of *****: open /tmp/stereoscope-/oci-registry-image-/sha256:.tar: too many open files

What you expected to happen:
Scan finishes without crash

Steps to reproduce the issue:
Scan any (bigger) image related to ML/AI (eg https://hub.docker.com/r/pytorch/pytorch/tags).

Environment:

  • Output of syft version: 1.3.0
  • OS (e.g: cat /etc/os-release or similar): Linux
@willmurphyscode
Copy link
Contributor

Thanks very much for the report @willem-delbare!

Pull #2823 should fix the issue - on my machine, it reduces the number of simultaneous file descriptors needed by syft to scan pytorch/pytorch:latest by ~50%.

I'm also working to put better linting in place, so that we're less likely to leak file handles in the future. #2825 is the start of this. Issue #2826 tracks adding additional linting so that we don't have one of these leaks again.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
Archived in project
Development

Successfully merging a pull request may close this issue.

2 participants