Nathan Cutler

Results 99 comments of Nathan Cutler

@jschmid1 tells me this is because the `ceph.restart` orchestration requires a cluster with `mds` and `rgw` roles deployed. The above error occurs when these roles are absent.

Yes, the CI can now run these tests, but reopening the issue to track the problematic error handling.

This could be resolved by implementing a validate runner for the functests and triggering it in init.sls (similar to how it is triggered by the stage orchestrations).

I've been running a lot of tests today in my private teuthology cluster in OVH region GRA1, and I'm not seeing this failure there at all. The only Stage 3...

It happens in my private GRA1 teuthology instance, too: * http://149.202.174.223:8081/ubuntu-2018-09-11_10:26:43-suse:tier1-ses5---basic-openstack/287/

Happening in the SES6 Staging CI: http://10.86.0.53:8081/ubuntu-2018-09-14_22:18:46-suse-ses6---basic-openstack/

Dropping priority because it doesn't happen on every CI run.

Happening in DeepSea master PR CI: http://ceph-ci.suse.de:8080/job/deepsea-pr/170/testReport/teuthology/deepsea_tier2_health-ok/_ceph_cm_salt_yaml_cluster__1node_yaml_4disks_yaml__deepsea_yaml_distros_sle_15_0_yaml_tasks_custom_profile_yaml_/

AFAICT this is *not* currently happening in SES5. Only in master/SES6.

After Stage 0 completes, `zypper ps -s` shows that systemd-udevd has been updated. It's not inconceivable that an updated, but not restarted, systemd-udevd might cause issues with ceph-disk, and that...