From be4275603f9d319ad3079a72ee9468893acdbe1e Mon Sep 17 00:00:00 2001 From: Morten Piibeleht Date: Fri, 4 Jul 2025 15:03:56 +0300 Subject: [PATCH 1/3] test: do not write temp file to repo --- test/datasets-large-live.jl | 22 ++++++++++++++++------ 1 file changed, 16 insertions(+), 6 deletions(-) diff --git a/test/datasets-large-live.jl b/test/datasets-large-live.jl index 37570f86c..6e7126044 100644 --- a/test/datasets-large-live.jl +++ b/test/datasets-large-live.jl @@ -1,24 +1,34 @@ @info "Uploading test data with prefix: JuliaHubLargeTest_$(TESTID)" dataset_name = "JuliaHubLargeTest_$(TESTID)_Blob" +large_data_file = joinpath(TESTDATA, "large.dat") try - open("testdata/large.dat", "w") do io + lf_ds, lf_filesize = mktemp() do path, io chunk = ones(UInt8, 1024^2) # 210 MB forces multipart upload in rclone for i = 1:210 write(io, chunk) end + close(io) + # Upload the file + r = JuliaHub.upload_dataset( + dataset_name, path; + description="some blob", tags=["x", "y", "z"], + auth + ) + r, filesize(path) end - JuliaHub.upload_dataset(dataset_name, - "testdata/large.dat"; - description="some blob", tags=["x", "y", "z"], - auth) + @test lf_ds isa JuliaHub.Dataset + @test lf_ds.name == dataset_name + datasets = JuliaHub.datasets(; auth) blob_dataset = only(filter(d -> d.name == dataset_name, datasets)) - @test blob_dataset.size == filesize("testdata/large.dat") + @test blob_dataset.size == lf_filesize finally try JuliaHub.delete_dataset(dataset_name; auth) catch err @warn "Failed to delete dataset '$(dataset_name)'" exception = (err, catch_backtrace()) end + # Also clean up the + rm(large_data_file; force=true) end From 86430349f5f9d893fb6d8aee7aaaba4cfaada9bc Mon Sep 17 00:00:00 2001 From: Morten Piibeleht Date: Fri, 4 Jul 2025 15:04:49 +0300 Subject: [PATCH 2/3] fix --- test/datasets-large-live.jl | 3 --- 1 file changed, 3 deletions(-) diff --git a/test/datasets-large-live.jl b/test/datasets-large-live.jl index 6e7126044..8925d3fc0 100644 --- a/test/datasets-large-live.jl +++ b/test/datasets-large-live.jl @@ -1,6 +1,5 @@ @info "Uploading test data with prefix: JuliaHubLargeTest_$(TESTID)" dataset_name = "JuliaHubLargeTest_$(TESTID)_Blob" -large_data_file = joinpath(TESTDATA, "large.dat") try lf_ds, lf_filesize = mktemp() do path, io chunk = ones(UInt8, 1024^2) @@ -29,6 +28,4 @@ finally catch err @warn "Failed to delete dataset '$(dataset_name)'" exception = (err, catch_backtrace()) end - # Also clean up the - rm(large_data_file; force=true) end From 50cd1853f4654a787e8b845add6c194c3e89af0a Mon Sep 17 00:00:00 2001 From: Morten Piibeleht Date: Fri, 4 Jul 2025 15:07:17 +0300 Subject: [PATCH 3/3] format --- test/datasets-large-live.jl | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test/datasets-large-live.jl b/test/datasets-large-live.jl index 8925d3fc0..8e282c309 100644 --- a/test/datasets-large-live.jl +++ b/test/datasets-large-live.jl @@ -12,7 +12,7 @@ try r = JuliaHub.upload_dataset( dataset_name, path; description="some blob", tags=["x", "y", "z"], - auth + auth, ) r, filesize(path) end