New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WIP: Cirrus: Update to XFS root in Fedora VM images #10201
Conversation
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: cevich The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
@rhatdan I have some...errr...results 😁 Both F33 and F34 are running with XFS, and in all cases do not appear to run significantly faster compared to EXT4. Though I'll assume that deeper inspection/debugging would find some/many low-level inefficiencies that need fixing. The test failures are quite a bit more worrying, since they might describe real-world problems that need fixing. The other alternative is test/environment setup bugs (i.e. not "real world" reproducible) but these too may be worth fixing. In all cases it would be good to put some expert eyeballs on these failures. Especially since switching to XFS appears to cause checkpoint/restore to outright hang 😖
Please let me know how I can help. |
Could you check to see if reflink is enabled?
|
|
If /tmp is on the / OS then reflinks are enabled and look like they are enabled by default, which is goodness. |
Yep, so either our bits aren't taking advantage as expected, or the affect isn't very substantial, or both. Still, I'm a bit more concerned about the test failures I mentioned above. They all strike me as things which shouldn't fail due to underlying FS changing. Albeit, I also don't know the details, maybe it makes sense to someone? |
Yes I have no idea what is causing the timeouts. |
The network-create failures are even more strange: I wonder if maybe these are all race-conditions in the tests which are simply being exposed by timing changes. If so, the failures are pretty consistent. I got basically the same failures running these jobs multiple times and with BTRFS also 😕 |
Signed-off-by: Chris Evich <cevich@redhat.com>
rebased on master |
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: cevich The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
@cevich: PR needs rebase. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Closing this for now, since there's a LOT of additional complexity in the image-build workflow for apparently minimal performance gain. Happy to revisit/re-open this at a later date if there is a need. |
COMPLETELY experimental; DO NOT MERGE
Ref: containers/automation_images#69