Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Duplicate dataset, duplicate files deposited into different dataverses, same MD5, same UNF for tabular, different DOIs should not be allowed to complete upload #2621

Closed
sbarbosadataverse opened this issue Oct 2, 2015 · 5 comments

Comments

@sbarbosadataverse
Copy link

sbarbosadataverse commented Oct 2, 2015

https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/GYWNQG
https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/ZTPW0Y

These are the same datasets, deposited into different dataverses and they were allowed to be "uploaded and published" in the UI

need a filter to prevent this from happening.

@landreev
Copy link
Contributor

landreev commented Oct 2, 2015

Says who??
Why?
We've always allowed the same files to be parts of different studies and dataverses... The goal was not to allow duplicates within a dataset - and that we are already doing.

@landreev landreev assigned scolapasta and unassigned landreev Oct 2, 2015
@sbarbosadataverse
Copy link
Author

@landreev
These datasets are identical data and using two different DOIs. While we allow linking, we should not have two copies of the same file to be used under different DOIs...never mind this is the same exact dataset---so while I understand that people can share files---this case has some uniqueness to it. You can assign it to Gustavo as you did but it still needs discussion

@sbarbosadataverse
Copy link
Author

I have updated the issue to capture the entire case

@sbarbosadataverse sbarbosadataverse changed the title Duplicate files deposited into different dataverses, same MD5, same UNF for tabular, should not be allowed to complete upload Duplicate dataset, duplicate files deposited into different dataverses, same MD5, same UNF for tabular, different DOIs should not be allowed to complete upload Oct 2, 2015
@mercecrosas mercecrosas modified the milestones: 4.3, In Review Nov 30, 2015
@scolapasta scolapasta removed their assignment Jan 27, 2016
@scolapasta scolapasta modified the milestone: Not Assigned to a Release Jan 28, 2016
@pdurbin
Copy link
Member

pdurbin commented Jun 25, 2017

See "4100+ of these failures ARE ACTUALLY THE SAME FILE" at #3675 (comment) ... copies of the same file across many, many datasets. I guess we allow this.

@pdurbin
Copy link
Member

pdurbin commented Jun 27, 2017

@sbarbosadataverse can we close this issue? The business logic is by design.

@pdurbin pdurbin added the User Role: Curator Curates and reviews datasets, manages permissions label Jul 5, 2017
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

7 participants