Skip to content

Optimizing PolarisFS for Zarr files up to 10GB #74

@cwognum

Description

@cwognum

Context

To create signed URLs for the directory-based file format called Zarr, Polaris implements a custom Zarr storage class called PolarisFS. We have a first version of this subsystem implemented, but performance is not great yet.

Description

This ticket is concerned with profiling and optimizing the current implementation such that we can read and write larger datasets. Using the internet in the Valence office in Montreal and using Cloudflare's speed test, I found that we have an up- and download speed of about 210 Mbps. This would mean we can up- and download 10GB files in about 6.5 minutes. The goal of this ticket is to be able to read and write a Zarr archive at most three times as slow as when we would've uploaded / downloaded the archive directly. (In other words, using Polaris in the Valence office in Montreal, the upload should take at most ~19.5 min).

Acceptance Criteria

  • We can read a 10GB Zarr archive in 13 minutes.
  • We can write a 10GB Zarr archive in 13 minutes.
  • The log details which optimizations were tried and what their effect was.

Links

Metadata

Metadata

Assignees

Labels

enhancementNew feature or requestfeatureAnnotates any PR that adds new features; Used in the release process

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions