-
Notifications
You must be signed in to change notification settings - Fork 149
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Large transfers failing with XCache (#2) #2168
Comments
What are you using to transfer these files? |
Well, good question. Those are CMS MC production jobs. I assume, they are just opening and streaming the files within CMSSW. |
At least the "Operation Interrupted" type of error might go away with 5.6.6 or later (see #2169). I would recommend to try again with XRootD 5.6.7, and if this is still an issue, we need more information on how to reproduce the problem to be able to investigate further the underlying cause (crash dump and/or full debug logs of a failed operation). |
Sure, sorry, I was a little busy lately 😅 Some more info: I will keep an eye on it and keep you updated! |
After talking with you yesterday, it would appear you are using a 2MB page size in Xcache. The problem is that a 2MB page size incompatible with a readv() since the maximum size is actually 2MB-16. I'd suggest a somewhat smaller page size. You might still get an error if Xcache wants to read too many large pages and you will exceed the total transfer limit for a readv. So, make sure the read ahead count is not excessive. Of course, we never discussed what is actually driving such a large page size in he first place. We can do that. |
Hi all,
I am currently observing this issue: #1893 again when transfering many CMS PREMIX files (~20GB) via XCache.
E.g.:
On top of that, I sometimes see this issue, not sure if it is connected:
I am using v5.6.4 on CentOs7.
Does anyone have an idea what can cause this or how to fix it?
Thanks
Robin
The text was updated successfully, but these errors were encountered: