-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Test PhoSim 3.7 with subset of DC1 inputs #19
Comments
Here are some selected sensor visits from DC1, extracted from a DC1 metadata file attached here The obsId entries below refer to minion_1016
Note that for DC1 the instance catalogs were modified so that stars brighter than 10th magnitude were limited to 10th magnitude. This essentially eliminated the dependence of CPU time on the magnitude of the brightest star in the instance catalog. Before this change, Tom did run some sensor visits with the original stellar magnitudes, which ranged as bright as magnitude ~5. These had a very large effect on the CPU times; with the largest effects actually on sensors adjacent to the bright stars. This page has some notes from from a test simulation of an entire focal plane image that Tom ran about a year ago, with a then-current version of phoSim. (October 30: I updated the table to include a column with the total number of sensors in the corresponding DC1 visit.) |
What should we do about the magnitude 10 cut for these tests? Is the thought this can now be removed? |
Jim/Seth:
Tom Glanzman and Glenn Sembroski did a lot of testing over the past few months and have been putting the results here:
https://docs.google.com/document/d/1f5aoP9ND62NNwdfcex21zYcdnY3VFTZkLgRyYP4yPy8/edit?usp=sharing
Its a little complicated in that we realized its important to fill nodes fully with many jobs to get the true timings, but you can see from these tests that production rates of ~25 to ~40 times DC1 rates should be realistic. Both Tom G and Salman are adding more text to that document before it gets done.
What should we do about the magnitude 10 cut for these tests? Is the thought this can now be removed?
Chris, not sure about this. It will be much faster, but still might want to avoid these at some level.
john
|
Thank you, John. That Google document mentions DC1 simulations of bright stars in one place: "Second, there was a rare but very significant bright star optimization bug which caused some of the brighter stars (10-12 magnitude) to be simulated unoptimized when they fell in certain places off the chip." For DC1, once the magnitudes of the brightest stars were limited to 10, the reported CPU times of the sensor visits did not depend strongly on the magnitude. Here's a plot from the first Confluence page linked to my comment above (sensor visits with moonAlt > 0 are in red, and moonAlt < 0 are in blue) The brightest star magnitude is the brightest star in the trimcat file for that sensor visit, regardless of whether it is off chip. I think that you have fixed the bright star optimization bug in v3.7, but I'm fairly sure that we do not have recent tests of the impact of bright stars on the execution time. The brightest stars (magnitudes less than 10) have (or had) kind of a compounded effect because they end up in the trimcat files for many sensors in the visit. It might be worth retrying the same visit that Tom ran in the second link in my comment above, if the various inputs are still available. |
Seth-
yes, En-Hsin debugged this problem by taking that specific catalog from Tom that took very, very long. it was also a strange bug that didn’t always trigger for all bright stars.
also, the background is drastically faster with the examples/quickbackground, so bright stars matter more. but the combination of the bug fix and the quickbackground is faster by about an order of magnitude.
the one thing i don’t know is whether to cut off the catalog at 10 or just to do the full catalog with several 7-10 mag objects in them typically. we need to simulate some full fields to be sure it is safe, because all of the ghosts and everything.
are you interested in studying this?
john
|
The first plot on this page shows the locations of the brightest stars in the obsId 1668469 (from minion_1016?) visit together with the CPU time required for the individual sensor visits. The brightest stars were magnitude ~<7 and each was surrounded by an array of sensor locations for which the simulations did not finish within the 120 CPU hour limit of the SLAC batch farm. Fainter stars had smaller regions of influence on the CPU times. I think that re-running this visit, if the various inputs are still available, would be a good test of the feasibility of including realistic magnitude distributions for the stars, e.g., whether threads simulating bright stars influence the overall efficiency. I'm afraid that I don't know enough to be able to do this; even if I knew how to run it in the pipeline system at SLAC, it would better be run at NERSC. I think that @TomGlanzman is the only one who knows how it all works. If it is interesting, I could make some individual runs with, e.g., a single bright star at various distances from a single sensor. That would not answer questions about overall throughput/efficiency, though. |
I've updated the table above to include the total number of sensors in the selected DC1 visits. None of them corresponds to a full focal plane, but I think most have more than enough for a reasonable test of current processing efficiency in the pipeline(s). |
@sethdigel Seth, what is the plan on this now? Is @TomGlanzman going to run some more tests? Do we have an estimated timeline for when this will happen? Thanks!! |
My understanding is that Tom has it on his to-do list to (re)run a sample of DC1 visits. I don't know the schedule. I think that he was working on getting a prototype DC2 simulation workflow going (involving Pegasus/HTCondor at NERSC) to run these. |
Hi @TomGlanzman. Can you let us know what the plan for these are? Is this tied into the new workflow as @sethdigel mentions? |
There have been many (sometimes conflicting) requests related to getting the DC2 phosim production running. The ultimate goal is, as I understand it, to have a Pegasus/HTCondor workflow handle this project at NERSC running on KNL. The immediate goal, however, is to resurrect and clone the DC1 phosim workflow (using the SLAC Pipeline workflow engine) and get that running within a DC2 context. This will provide us with important performance benchmarks and, assuming proper DC2 catalogs are available, a prototype DC2 dataset. I am hoping to complete the commissioning of the new workflow this week. |
Akhtar has been running some DC1 on the OSG, if you want him to do more of these (cc’d here).
john
|
There is a third option: to use the set up that Glen has used at NERSC
(John?)
I don't think there are "conflicting" requests, if so we should capture
those somewhere and sort out the conflicts. I also don't think we have
made the final decision that we will use Pegasus for production before
we have at least one case where we have demonstrated that it works well.
So the "conflicting" requests are most likely because of timeline issues
-- if we want to stick with the DC2 timing plan (and therefore have a
test set for validation) but the Pegasus work is not ready, we have to
use a Plan B for the small tests. Sometimes things have to be worked on
in parallel if sequentially they are not ready.
Scott is working hard on getting the CatSim catalogs ready for ingestion
into PhoSim -- some unforeseen problems came up there, but that's ok,
that's the whole reason that we are doing this downscaled test in the
first place to iron out problems with small data.
…On 11/13/17 11:10 AM, Tom Glanzman wrote:
There have been many (sometimes conflicting) requests related to
getting the DC2 phosim production running. The ultimate goal is, as I
understand it, to have a Pegasus/HTCondor workflow handle this project
at NERSC running on KNL. The immediate goal, however, is to resurrect
and clone the DC1 phosim workflow (using the SLAC Pipeline workflow
engine) and get that running within a DC2 context. This will provide
us with important performance benchmarks and, assuming proper DC2
catalogs are available, a prototype DC2 dataset. I am hoping to
complete the commissioning of the new workflow this week.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#19 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AMQ9jKGMdOSYloG8ghyN1rU9D2FWZWTzks5s2HgFgaJpZM4QFF4W>.
|
Yeah, Glenn way is a good idea too, since it would be at NERSC. Glenn could either do it or teach Tom. should be quick.
john
|
Hi John,
If Glenn could do this (meaning he has time) that would be awesome! This
way Tom can continue to focus on Pegasus. If Glenn can also put together
some notes on how his set-up works or maybe there is something already
that can be looked at (sorry, I did not follow this in DC1) and he can
point people too for the future that would be super-useful as well.
Scott is still putting together the small test set coming out of CatSim
but he was optimistic that this should be done in the next couple of days.
Thanks!!
…On 11/13/17 11:38 AM, johnrpeterson wrote:
Yeah, Glenn way is a good idea too, since it would be at NERSC. Glenn
could either do it or teach Tom. should be quick.
john
On Nov 13, 2017, at 12:21 PM, Katrin Heitmann
***@***.******@***.***>> wrote:
There is a third option: to use the set up that Glen has used at NERSC
(John?)
I don't think there are "conflicting" requests, if so we should capture
those somewhere and sort out the conflicts. I also don't think we have
made the final decision that we will use Pegasus for production before
we have at least one case where we have demonstrated that it works well.
So the "conflicting" requests are most likely because of timeline issues
-- if we want to stick with the DC2 timing plan (and therefore have a
test set for validation) but the Pegasus work is not ready, we have to
use a Plan B for the small tests. Sometimes things have to be worked on
in parallel if sequentially they are not ready.
Scott is working hard on getting the CatSim catalogs ready for ingestion
into PhoSim -- some unforeseen problems came up there, but that's ok,
that's the whole reason that we are doing this downscaled test in the
first place to iron out problems with small data.
On 11/13/17 11:10 AM, Tom Glanzman wrote:
>
> There have been many (sometimes conflicting) requests related to
> getting the DC2 phosim production running. The ultimate goal is, as I
> understand it, to have a Pegasus/HTCondor workflow handle this project
> at NERSC running on KNL. The immediate goal, however, is to resurrect
> and clone the DC1 phosim workflow (using the SLAC Pipeline workflow
> engine) and get that running within a DC2 context. This will provide
> us with important performance benchmarks and, assuming proper DC2
> catalogs are available, a prototype DC2 dataset. I am hoping to
> complete the commissioning of the new workflow this week.
>
> —
> You are receiving this because you commented.
> Reply to this email directly, view it on GitHub
> <#19 (comment)>,
> or mute the thread
>
<https://github.com/notifications/unsubscribe-auth/AMQ9jKGMdOSYloG8ghyN1rU9D2FWZWTzks5s2HgFgaJpZM4QFF4W>.
>
—
You are receiving this because you commented.
Reply to this email directly, view it on
GitHub<https://secure-web.cisco.com/1EpzbXSk5jQk9B_Bw4-FvJE6qCEkr2Df3pG0wgN2eyWD3sp_PtK8EbGffv0qOiiC4NXimgPmklWPdprLzIx0QEk3_S6U3qBmrTnNRlADRS0u48atdNNYu-Up9Baynm3JNzjNzfDJXk9W_2lqI7ji3GOhNc1t-gcrFNGXKIfwUV9bEGnfRrbzU-SAPakA1OBq9RAkLR_W-Fb3-PyUMhTjWPQEbgkzGLLs4Yrt-mL5AMdUo7Og5-v0hNDUl0etVHeChmx_EBL8qLiAp2JT8coN2HCXVcRM4dBVtQ0oW6a73y7Ami96Gsin-z8wKSex5JGprji58q3bx_pU5nk3Xdb93-m1P7lMu8ZcUoq48FmIWF4o/https%3A%2F%2Fgithub.com%2FLSSTDESC%2FDC2_Repo%2Fissues%2F19%23issuecomment-343991798>,
or mute the
thread<https://secure-web.cisco.com/1O8CQX1BWVGNXKpPzazCiLm1BamD50DqGcFmEq-pjQ9KBhgvvFDRI4--0ao9EYnVe8uDSqECmq_uKPnCGTHsRqLFEhZjNW6Th5NwJ3EefhvaQS66wfE0hU3pevQFlA3G4Z2SnRRRPR6sk2IV_9t8OtfNgI20yyw8HbSYQwPItUPBgIBi7OCKGYVbbWAdr27fKEt-3OzJv4suB8okc7NWAJJvVlGiKAUiq-DZN8jOEX__-eIQLIcu2nO8Khdf41ALih0kDOr8tUpDlgnVL6Lq3hVSQjAs9tfPQGhdIAx6OGz_jVCW0Y_BwHAxcTcOK-xqGmf7lXjJH93WUWw05uvXq1CLv8GeVqPR515O9UrYxUqo/https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FAJbT8hg21hEzz9RX_UfNOavpOUE3bg-Sks5s2HqhgaJpZM4QFF4W>.
———————————
John R. Peterson
Assoc. Professor of Physics and Astronomy
Department of Physics and Astronomy
Purdue University
525 Northwestern Ave.
West Lafayette, IN 47906
(765) 494-5193
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#19 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AMQ9jMcoOIBGeRuQNMEybc27hp7v8hA7ks5s2H6IgaJpZM4QFF4W>.
|
Tom, could you send the settings file for DC1? i think we want to add the quickbackground to those, but probably want to keep the same settings otherwise.
john
|
Hi All, There seems to be a mixture of things being discussed now. This issue as we originally wrote it is not to test DC2 production or any other issues related to catalogs etc etc. It is to test 3.7 with the new settings with some of the same inputs from DC1 and careful control so we can have the real scaling numbers. It would be fine to run this in the old workflow on KNL but running it in the new one would be OK too. I would still like this to happen independent of the other things being discussed. It would be good to make new issues for other requests so we can track things in a focused way. |
Chris, the scaling numbers aren’t really necessary as we have those. The important this we need is to: 1) have workflow-type tools fully functioning and ready for DC2, and 2) to test DC2 production related to any catalog issues as you said. I don’t see what numbers we would need related to performance that we don’t already have.
john
|
I am not worried about scaling (agree with John) but we do want to check that the quick background method is acceptable. I would like to see an objective set of tests for this (will chat separately with John). |
I sent this file to John via email, but want to publish it here also for comment. This command file is a clone of that used for DC1 with the addition of the 'quickbackground' and sourceperthread directives. Is this a good start point for DC2?
|
@TomGlanzman Thanks. For the real DC2 we need to check the selected options are turned on. So for example if we wanted BF, we might need 'chargesharing 1' if cleardefects is clearing that too. Also, I can't remember: are we setting a fixed sky background with the zenith_v command? I remember that there was some bug in twinkles that required us to do that and then it was decided it should be removed. Was that this command or something different? |
@cwwalter If any of the "selected options" are known today, I would really like to include them asap to generate some benchmarks. Yes, 'cleardefects' does clear the 'chargesharing', so if the latter is desired, it would need to be reactivated explicitly. Don't recall the story behind zenith_v - perhaps someone listening in can comment? |
zenith_v was in there to not make the background variable between different between runs. it was always realistic and not a bug, but we didn’t have the time correlations that you naturally have when you do a sequence of exposures in one night.
in v3.7, the time correlations are there, so this command is not necessary anymore.
however, it is unclear what you mean by benchmarks. if you are checking science quality then don’t include this. if you are doing timing tests (which we do not need to do anymore as i argued), then you could keep it there.
john
|
On Nov 13, 2017, at 2:30 PM, Salman Habib <notifications@github.com<mailto:notifications@github.com>> wrote:
I am not worried about scaling (agree with John) but we do want to check that the quick background method is acceptable. I would like to see an objective set of tests for this (will chat separately with John).
its accurate to below 1% in normalization and the large scale patterns are still there (vignetting & various backgrounds).
|
If realistic time correlations are now in phosim, I agree we can take out {{zenith_v}}. The problem before was that simulating back to back observations using different phosim processes could cause the sky background to change by a few magnitudes. |
@johnrpeterson do you have some sort of validation data that shows this? Or is this essentially a by eye test? |
yeah, here you go:
[cid:B0D8CD49-2E7E-4BF7-81D7-7427C96AE74F@itap.purdue.edu]
|
Hi John, I think you were trying to attach an image (or a link)? It didn't work. Can you try again? BTW, would you consider going to the GH website to enter your responses? Because of something in the Purdue security mail scanner it is turning each of your answers into a big mess. Go ahead and look at what the last entry looked like on the web site. Rachel and I have been editing the comments to remove the extra material but it would be great if we could bypass that. Thanks! -Chris |
@johnrpeterson I think this is great. What we should do is write up a small document stating how you go from the full background to the acceptable quick background approximation. And what the validation tests are. |
@johnrpeterson - just to expand on @salmanhabib 's request, when you do that, can you please clarify a little what is being plotted in each panel on this figure so we know what we should be taking away from it? (I have a guess as to what you're doing in this validation test, but the labels are not quite obvious enough that I am 100% positive of my interpretation.) |
@johnrpeterson and one more request, when you plot those distributions (like the lower right), will you also plot the ratio so we can see fractionally how well they agree? |
OK this is under active discussion right now but I suspect BF will be on. Fringing likely won't be on but might be. |
Rachel, for the validation test: this is a flat near the edge of the field, so there is a lot of vignetting. the three panels are the image with the different level of approximations. the histogram is the relative flux in the pixels. Chris, with quick background the sensor physics will be averaged over for the background & dome light on small spatial scales. so you cannot see the effects in the background, but could see them in the astrophysical sources. so this makes turning on "tree rings", "BF", "fringing", for example, a little bit strange and inconsistent. not saying you couldn't do this, but would have to think through what DM is doing with these things. you could keep these on in the default optimization and there is no issue for that. |
Hi John, Thanks. As you know, I'm discussing now with DM people about what corrections are currently possible etc. I'm not sure I am following completely what you are saying so I want to make sure I understand (as we are working with the SAWG to check some related items in some tests now). Are you saying with the quick background the photon bunching optimization is large enough that you won't expect to see BF, tree rings etc in background sky photons, but that you expect the effect will still be there for any astrophysical source and they will be summed? |
yes, basically. its large scale patterns that are preserved, but not the smaller-scale sensor details. [but not sure what you mean by "summed"] so lets take each one:
|
Sorry.. I just meant the background with no effect and the sources with an effect will sum to the total electron level. This is what we are testing with SAWG because the BF correction kernel is extracted using the flats and we wanted to quantify the residual correlation size relative to the BF. So we will consider this. Thanks. |
definitely that is fine with regular optimization. for the others it really depends on what DM does. so we have to look into it. with fringing, for example, its not common to even correct photometry for fringing effects, but it is common to remove fringing from the background. with tree rings, historically they have never been corrected correctly, so no one knows this ,but it wouldn't be that important to correct the background for tree rings, but it would be important to correct the astrometry of sources. same with what you say about brighter-fatter, you'd use the flats to infer what to do about the sources, however probably not correct the flats, I'm guessing. so its safe to say that none of the physics would be completely "reversed", but its important to be clear about what observable consequence DM is capable of correcting now. also, with BF, you could make some special flats with normal optimization and then you could apply the BF corrections to images in the DC with quick background later. |
I would like to suggest a few limited tests on one FOV (or some reasonable fraction thereof), where we compare the full-on background, the approximate background, and runs with fixed sources including all the background options. This way, we will have a data set where at least some of these questions can be directly addressed quantitatively. We can do this at Argonne. |
As of 20:43 Sun 11/19/2017, all of the runs mentioned below are complete. Update: A set of 11 visits using (mostly) DC1 visits and phosim configuration but with quickbackgrounds enabled is nearing completion. The first three visits are the first three DC1 visits. The 4th visit is not DC1 (rejected because it took too many hours to complete) and is a performance check. The final seven (7) visits are those suggested by Seth (above). These visits are summarized in this spreadsheet: https://docs.google.com/spreadsheets/d/17-EUgaDMtfgQLt84WquxeAjycqJaqHJZS_jZqrQXOVw/edit?usp=sharing The remaining sensor-visits should complete before midnight tonight. As I will be on vacation all next week, please feel free to take a look at the data once they are available (by checking the workflow web page: http://srs.slac.stanford.edu/Pipeline-II/exp/LSST-DESC/task.jsp?refreshRate=60&task=49412792). |
Getting back to the issue of the phoSim command file to use for DC2, I am trying to keep notes on how the discussion is evolving. My current guess, based on DC1 with recent updates follows. Note that it does not include all of @cwwalter list as I am not certain which combination of phoSim commands corresponds to his list - perhaps @johnrpeterson can help with the association? ` # Disable sensor effects # Enable brighter-fatter # Disable dirt # Quick background # Number of sources handled by a single thread |
The actual commands.txt file (listed above) is in github and currently here: https://github.com/TomGlanzman/DC2-phoSim-2/blob/master/NERSC/commands.txt Please comment on the content in this issue. |
I propose that we discuss the content of this file the same way we should handle any other code: make a branch where the new version of the file is introduced, issue a pull-request for that branch, and in the pull-request people can comment directly on the relevant lines with proposals for changes. Since PRs (assuming the repo is configured properly) can only be merged after review, this provides control that the reviewers agree that the proposed changes are acceptable. Since this is a file that would be used for DC2 productions, I think this commands.txt file should be in the DC2_Repo repository. |
Here is the PR to discuss the contents of the PhoSim commands/physicsOverride file: |
Since we have shifted our testing to the DC2 era, I will close this issue. The most important point, checking out background implementations in more detail has been move to #113. |
This will be done on KNL with a couple of "typical" visits, both with default and new "quick background" settings for comparison to DC1 production runs to derive expected scalings for DC2.
We should also do a couple of long runs from DC1, i.e., where the background light was high or there was a off-sensor bright star to get a sense of the long execution time tails.
The text was updated successfully, but these errors were encountered: