Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

analysisTargets doesn't seem to scale well #28

Closed
zlandau opened this issue Sep 1, 2015 · 4 comments
Closed

analysisTargets doesn't seem to scale well #28

zlandau opened this issue Sep 1, 2015 · 4 comments
Milestone

Comments

@zlandau
Copy link

zlandau commented Sep 1, 2015

It sounds like analysisTargets should specify every file that has been looked at (which also includes hashing of these files). That seems unnecessarily expensive for very large codebases, especially ones that are in revision control. I imagine a common scenario would be to run the analysis of all files in a specific code base, in which something like an overall file tree version (like a commit hash when under revision control) would suffice.

Perhaps analysisTargets could allow different methods of specifying which files the analysis was run on, with listing of all files just being one of the options?

@codecurmudgeon
Copy link

Yes, for example a project name, or scm tag.

On Tue, Sep 1, 2015 at 11:25 AM, Zachary P. Landau <notifications@github.com

wrote:

It sounds like analysisTargets should specify every file that has been
looked at (which also includes hashing of these files). That seems
unnecessarily expensive for very large codebases, especially ones that are
in revision control. I imagine a common scenario would be to run the
analysis of all files in a specific code base, in which something like an
overall file tree version (like a commit hash when under revision control)
would suffice.

Perhaps analysisTargets could allow different methods of specifying which
files the analysis was run on, with listing of all files just being one of
the options?


Reply to this email directly or view it on GitHub
#28.

Regards,

Arthur "Code Curmudgeon" Hicken
Office: 626-275-2445 (Los Angeles)
Cell: 909-343-0936
Google voice: 978-49-OWASP (in theory rings my mobile and my desk... in
theory)

@michaelcfanning
Copy link
Contributor

this makes sense.

@ghost ghost added this to the future milestone Mar 7, 2016
@ghost
Copy link

ghost commented Mar 7, 2016

@zlandau, @codecurmudgeon: I'm going to come back and make this comment clearer. Right now it's a placeholder for me.

#107 says that direct producers will express result locations as a simple URI, and that they should provide more information about the files referenced by those URIs in the "physicalLocationInfo" property of "runLog". physicalLocationInfo is not required, it's only needed at all for files that are mentioned in results.

We'll recommend they fill out metadata for every file mentioned in a result, but they might not be able to.

In a future "compliance" profile, we'll examine the problem of exhaustively identifying all analysis targets.

@ghost
Copy link

ghost commented Apr 27, 2016

Good news! It is no longer necessary to enumerate every analysis target. The run.files property need only mention the files in which results are produces -- and in fact it doesn't even have to mention all of those.

@ghost ghost closed this as completed Apr 27, 2016
This issue was closed.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants