Skip to content

consider computational complexity #321

@VladimirAlexiev

Description

@VladimirAlexiev

It would be nice for each new feature (or more realistically a bundle of features, i.e. Profile) to have some idea about its implementation and execution complexity.

Consider this scenario:

  • a database of 1, 10 or 100B triples (data at rest), which are assumed valid (eg parts have been validated, parts are from a trusted valid source)
  • a transaction of 1, 10 or 100M triples (data in motion)
  • a shapes graph of 1 or 10k shapes. Shapes refer to both data in motion, and also data at rest.

How do you validate this scenario in a reasonable time?
That's a difficult question.

  • Many implementations use in-memory models ("give me data files, give me shape files, I load them in memory and output a validation report").
    • I think all JS implementations are like this. Which is not efficient in the large-data scenario described above (@bergos , I'd love it if you prove me wrong!)
  • SPARQL opens up a "door" towards "unknown" complexity. At least I'm not aware of work on assessing the practical complexity of SPARQL queries

Relevant issues:

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions