Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

4813 allow duplicate files #6924

Merged
merged 83 commits into from Aug 3, 2020
Merged

4813 allow duplicate files #6924

merged 83 commits into from Aug 3, 2020

Conversation

sekmiller
Copy link
Contributor

@sekmiller sekmiller commented May 20, 2020

What this PR does / why we need it:
This will allow users to upload multiple files with the same checksum value to a given dataset. On upload either via the interface or api the user will be warned that a duplicate file exists in the dataset (with the current path/label of the duplicate file.) At that point they can directly delete the newly uploaded file - if the upload is via the UI.

Which issue(s) this PR closes:
#4813 - allow files with the same MD5/checksum to exist in the same dataset

Closes #4813
Closes #6468

Special notes for your reviewer:
Really wanted to take a stick of dynamite to AddReplaceFileHelper, but ended up working with it as it exists. Also fixed an issue with the editFileMetadata api where if you weren't updating the file's label you'd get an duplicate file name error. This was causing a failure in the .testForceReplaceAndUpdate Test

Suggestions on how to test this:
various scenarios uploading a duplicate file including replace
document outlining upload use cases and expected messaging

Does this PR introduce a user interface change?:
Introduces a popup on upload of a duplicate file, which warns the user and allows them to immediately delete the newly uploaded file

Is there a release notes update needed for this change?:
We could note that duplicate files within a dataset are now allowed as a new feature

Additional documentation:

@coveralls
Copy link

coveralls commented May 20, 2020

Coverage Status

Coverage decreased (-0.04%) to 19.562% when pulling 57ab613 on 4813-allow-duplicate-files into a6f580f on develop.

@djbrooke djbrooke moved this from Code Review 🦁 to IQSS Team - In Progress 💻 in IQSS/dataverse (TO BE RETIRED / DELETED in favor of project 34) May 20, 2020
@sekmiller sekmiller moved this from IQSS Team - In Progress 💻 to Code Review 🦁 in IQSS/dataverse (TO BE RETIRED / DELETED in favor of project 34) May 20, 2020
@sekmiller sekmiller removed their assignment May 20, 2020
@scolapasta scolapasta assigned scolapasta and sekmiller and unassigned scolapasta May 28, 2020
@djbrooke djbrooke moved this from Code Review 🦁 to IQSS Team - In Progress 💻 in IQSS/dataverse (TO BE RETIRED / DELETED in favor of project 34) May 29, 2020
@sekmiller
Copy link
Contributor Author

I updated the messaging based on discussion during the Design Meeting.

@sekmiller sekmiller removed their assignment Jul 30, 2020
@sekmiller sekmiller removed their assignment Jul 30, 2020
@scolapasta scolapasta removed their assignment Jul 30, 2020
@kcondon kcondon self-assigned this Jul 31, 2020
@kcondon
Copy link
Contributor

kcondon commented Jul 31, 2020

So, I tested all the rules and used cases as best I could. I think they all work except on upload when paths are involved for duplicates that have not yet been saved.

  1. Upload file2.txt twice, edit paths to be c,d. Cannot save, says duplicate filenames.
    This works if we upload a zip with dupe filenames in different paths.

@kcondon kcondon removed their assignment Jul 31, 2020
@kcondon kcondon moved this from QA 🔎✅ to IQSS Team - In Progress 💻 in IQSS/dataverse (TO BE RETIRED / DELETED in favor of project 34) Jul 31, 2020
@sekmiller sekmiller moved this from IQSS Team - In Progress 💻 to QA 🔎✅ in IQSS/dataverse (TO BE RETIRED / DELETED in favor of project 34) Aug 3, 2020
@sekmiller sekmiller moved this from QA 🔎✅ to IQSS Team - In Progress 💻 in IQSS/dataverse (TO BE RETIRED / DELETED in favor of project 34) Aug 3, 2020
@sekmiller
Copy link
Contributor Author

I was able to upload duplicate files and edit their paths and successfully save them
Screen Shot 2020-08-03 at 9 45 58 AM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
No open projects
7 participants