-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RuntimeError: epics failed to respond #273
Comments
Currently, this is seeming very similar to slaclab/pysmurf#713 , which was "solved" by an update to the smurf-streamer but perhaps I marked as closed prematurely. |
just for reference, seems like this generated a core dump (core_1660699519_python3_11091_11979_1001_1000) with the following backtrace:
I will look more into it but if you find a way to semi-reliably reproduce it that would be very helpful |
But definitely seems like an issue with the streamer, possibly a race condition or something |
Sent this to Daniel in Slack, but if you run this command it will enable more debugging lots in the smurf-streamer dockers...
which could provide some good info related to what's going wrong |
Hi Daniel, I was able to debug this a bit after seeing it in some SAT1 tests. I added some queue limits in this PR of the smurf-streamer, which seems to have fixed a lot of the issues we were seeing when operating multiple slots on the SAT. If you want to upgrade you can use the docker tag Since this has fixed things on SAT1 I'm going to close this issue, but feel free to re-open if you upgrade and still see crashes. |
Happened with smurf-streamer version v0.4.1-3-g728183a . I was only doing things on one slot at the time. I don't see any out of the ordinary in the smurf streamer log or in core dumps. I can't communicate with the board now, just getting "epics failed to respond" errors. Original crash message:
|
What were you doing when it crashed? |
I was running https://github.com/simonsobs/readout-script-dev/blob/master/ddutcher/ufm_biasstep_sodetlib.py , it should have been running bias steps at the time it crashed. The last messges in stdout before the timeout were
|
Interesting... this could be the same issue but I don't see a core-dump file on your system. It seems like your smurf-server, being one of the first ones issued, is also under-spec'ed compared to the ones we're using on the SAT, so it kind of makes sense you're seeing this the most often. We were seeing it more frequently on our system that was having RAM issues. Replacing it with an official one might alleviate this issue... Apart from replacing your server there are a few things we can probably try that might help:
|
Got this error when running uxm_setup, during the estimate_phase_delay portion. The full traceback is below, though I know users often enounter this error in various places, so this can be a catch-all thread.
In this particular instances, there was no associated error in the smurf-streamer docker logs, and I could still communicate with the board via the pysmurf-ipython session, and I could just restart the uxm_setup script with no hammering required.
The text was updated successfully, but these errors were encountered: