octopus: osdc: add timeout configs for mons/osds#37530
Merged
yuriw merged 8 commits intoceph:octopusfrom Oct 29, 2020
Merged
Conversation
3 tasks
3c8da8c to
7e1fb65
Compare
Member
Author
|
tested with #37529. Just need to catch the expected exception. |
af6902f to
13f2292
Compare
Member
Author
|
retest this please |
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit 0feabb4)
13f2292 to
d437991
Compare
d437991 to
fbdf16a
Compare
3e70ca0 to
079286c
Compare
Have the Objecter track the rados_(mon|osd)_op_timeout configs so that it can be configured at runtime/startup. This is useful for the MDS/ceph-fuse so that we can avoid waiting forever for a response from the Monitors that will never come (statfs on a deleted file system's pools). Also: make these configs take a time value rather than double. This is simpler to deal with in the code and allows time units to be used (e.g. "5m" for 5 minutes). Fixes: https://tracker.ceph.com/issues/47734 Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit a8a2374) Conflicts: src/client/Client.cc src/librados/RadosClient.cc src/mds/MDSRank.cc src/mgr/MgrStandby.cc src/mon/MonClient.h src/neorados/RADOSImpl.cc src/osd/OSD.cc src/osdc/Objecter.cc src/osdc/Objecter.h src/test/mon/test_mon_workloadgen.cc src/tools/cephfs/MDSUtility.cc Notes: different Objecter cons arguments. Added conf obs for RadosClient.
Otherwise we have unnecessary timeout waits. Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit bc25bd7) Conflicts: qa/tasks/cephfs/test_admin.py Notes: delete_all_filesystems method moved
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit 2432871) Conflicts: qa/tasks/cephfs/fuse_mount.py
The mount.cleanup method will remove the mount point. This `rm -rf` will always fail (with exit status 0). Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit 8e7a890) Conflicts: qa/tasks/cephfs/fuse_mount.py Notes: convert to cleanup call.
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit f8f607d) Conflicts: qa/tasks/cephfs/mount.py Notes: skip as cleanup is abstract.
Now that the osdc Objecter obeys updates to these configs, let's use them to avoid having them block forever on operations that may never complete (or should complete in a timely manner). Fixes: https://tracker.ceph.com/issues/47734 Signed-off-by: Patrick Donnelly <pdonnell@redhat.com> (cherry picked from commit d060c9a)
Otherwise, the umount process will fail because the mount still exists
when the mountpoint cleanup (rmdir) is started.
See:
2020-10-04T22:08:24.448 INFO:teuthology.nuke.actions:Clearing teuthology firewall rules...
2020-10-04T22:08:24.449 INFO:teuthology.orchestra.run.smithi063:> sudo sh -c 'iptables-save | grep -v teuthology | iptables-restore'
2020-10-04T22:08:24.464 INFO:teuthology.orchestra.run.smithi189:> sudo sh -c 'iptables-save | grep -v teuthology | iptables-restore'
2020-10-04T22:08:24.482 INFO:teuthology.nuke.actions:Cleared teuthology firewall rules.
2020-10-04T22:08:24.483 INFO:teuthology.orchestra.run:Running command with timeout 900
2020-10-04T22:08:24.483 INFO:teuthology.orchestra.run.smithi063:> (cd /home/ubuntu/cephtest && exec stat --file-system '--printf=%T
2020-10-04T22:08:24.483 INFO:teuthology.orchestra.run.smithi063:> ' -- /home/ubuntu/cephtest/mnt.0)
2020-10-04T22:08:34.550 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:08:34.553 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:09:04.592 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:09:04.596 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:09:34.727 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:09:34.730 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:10:04.815 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:10:04.818 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:10:34.876 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:10:34.880 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:11:04.923 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:11:04.926 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:11:34.996 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:11:35.000 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:12:05.064 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:12:05.067 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:12:35.202 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:12:35.205 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:13:05.316 INFO:teuthology.orchestra.run.smithi063:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:13:05.318 INFO:teuthology.orchestra.run.smithi189:> sudo logrotate /etc/logrotate.d/ceph-test.conf
2020-10-04T22:13:24.520 INFO:teuthology.orchestra.run.smithi063.stderr:stat: cannot read file system information for '/home/ubuntu/cephtest/mnt.0': Connection timed out
2020-10-04T22:13:24.521 DEBUG:teuthology.orchestra.run:got remote process result: 1
2020-10-04T22:13:24.522 INFO:tasks.cephfs.fuse_mount:mount point does not exist: /home/ubuntu/cephtest/mnt.0
2020-10-04T22:13:24.640 INFO:teuthology.orchestra.run:Running command with timeout 300
2020-10-04T22:13:24.641 INFO:teuthology.orchestra.run.smithi063:> (cd /home/ubuntu/cephtest && exec rm -rf /home/ubuntu/cephtest/mnt.0)
2020-10-04T22:13:24.688 INFO:teuthology.orchestra.run.smithi063.stderr:rm: cannot remove '/home/ubuntu/cephtest/mnt.0': Is a directory
2020-10-04T22:13:24.688 DEBUG:teuthology.orchestra.run:got remote process result: 1
From: /ceph/teuthology-archive/pdonnell-2020-10-04_21:51:57-fs-wip-pdonnell-testing-20201004.051319-octopus-distro-basic-smithi/5494771/teuthology.log
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
079286c to
e6652fa
Compare
Member
Author
|
This is ready for wider testing in regular octopus QA. |
3 tasks
Member
Author
Contributor
Contributor
Contributor
|
passed rbd approved by @dillaman |
yuriw
approved these changes
Oct 29, 2020
Contributor
yuriw
left a comment
There was a problem hiding this comment.
Reviewed-by: Patrick Donnelly pdonnell@redhat.com
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
https://tracker.ceph.com/issues/47942