Tuesday, 2021-04-20

clarkbI need to call it here and help with dinner. It seems like we can safely use ssh-keyscan -4 or 127.0.0.1 in place of localhost for now if we need a workaround00:01
openstackgerritIan Wienand proposed zuul/nodepool master: Remove statsd args to OpenStack API client call  https://review.opendev.org/c/zuul/nodepool/+/78686200:19
*** hamalq has quit IRC01:27
openstackgerritIan Wienand proposed zuul/zuul-jobs master: collect-container-logs: don't copy on failure  https://review.opendev.org/c/zuul/zuul-jobs/+/78701902:10
*** ajitha has joined #zuul02:27
*** evrardjp has quit IRC02:33
*** evrardjp has joined #zuul02:33
*** sam_wan has joined #zuul03:15
*** ykarel has joined #zuul03:33
ianwhttps://github.com/docker/for-linux/issues/123303:59
ianwSince 20.10.6 it's not possible to run docker on a machine with disabled IPv6 interfaces03:59
ianwoh, sorry, ignore that, it's more about kernel disabled ipv604:00
openstackgerritIan Wienand proposed zuul/nodepool master: nodepool-functional : ignore errors copying logs  https://review.opendev.org/c/zuul/nodepool/+/78706504:18
*** vishalmanchanda has joined #zuul04:23
*** saneax has joined #zuul04:26
*** ianychoi__ has joined #zuul04:48
*** sanjayu_ has joined #zuul04:48
*** saneax has quit IRC04:50
*** ianychoi_ has quit IRC04:50
openstackgerritMerged zuul/zuul master: Use ssh-keyscan -4 in quick-start  https://review.opendev.org/c/zuul/zuul/+/78698804:56
*** jfoufas1 has joined #zuul05:11
openstackgerritIan Wienand proposed zuul/nodepool master: Account for resource usage of leaked nodes  https://review.opendev.org/c/zuul/nodepool/+/78582106:20
openstackgerritIan Wienand proposed zuul/nodepool master: QuotaInformation : abstract resource recording  https://review.opendev.org/c/zuul/nodepool/+/78709306:20
*** bhavikdbavishi has joined #zuul06:39
*** bhavikdbavishi has quit IRC06:43
*** jcapitao has joined #zuul06:51
openstackgerritIan Wienand proposed zuul/nodepool master: nodepool-functional : ignore errors copying logs  https://review.opendev.org/c/zuul/nodepool/+/78706506:52
*** harrymichal has quit IRC07:23
ianwi've run out of time to figure out why the nodepool functional jobs aren't working.  it's something to do with the docker install07:34
ianwhttps://review.opendev.org/c/zuul/nodepool/+/787065 is a change that makes the logs be captured correctly07:34
*** rpittau|afk is now known as rpittau07:35
ianwhttps://98b486d8b4772b9c610a-a795987d7c31db10daa7f58e24ee8596.ssl.cf1.rackcdn.com/787065/2/check/nodepool-functional-container-openstack-siblings/2d652ea/syslog07:35
ianwis a sample syslog07:35
ianwthe relevant part seems to be07:35
ianwApr 20 07:29:15 ubuntu-bionic-inap-mtl01-0024121870 systemd[1]: Starting Docker Application Container Engine...07:35
ianwApr 20 07:29:15 ubuntu-bionic-inap-mtl01-0024121870 dockerd[5153]: time="2021-04-20T07:29:15.607437905Z" level=info msg="Starting up"07:35
ianwApr 20 07:29:15 ubuntu-bionic-inap-mtl01-0024121870 dockerd[5153]: failed to load listeners: no sockets found via socket activation: make sure the service was started by systemd07:35
ianwApr 20 07:29:15 ubuntu-bionic-inap-mtl01-0024121870 systemd[1]: docker.service: Main process exited, code=exited, status=1/FAILURE07:35
ianwthere is no obvious solution when googling that, but there are several hits07:35
ianwif nobody else figures it out, i guess i'll get back to it tomorrow!07:35
openstackgerritDaniel Blixt proposed zuul/zuul-jobs master: WIP: Make build-sshkey handling windows compatible  https://review.opendev.org/c/zuul/zuul-jobs/+/78066207:44
*** tosky has joined #zuul07:45
openstackgerritDaniel Blixt proposed zuul/zuul-jobs master: Make build-sshkey handling windows compatible  https://review.opendev.org/c/zuul/zuul-jobs/+/78066207:45
*** jpena|off is now known as jpena07:58
openstackgerritAndy Ladjadj proposed zuul/zuul master: [reporter][elasticsearch] fix the timestamp when the system has a different timezone by forcing the UTC timezone  https://review.opendev.org/c/zuul/zuul/+/78644408:03
*** mgoddard has quit IRC08:08
openstackgerritMerged zuul/zuul master: Add nodepool.external_id to inventory  https://review.opendev.org/c/zuul/zuul/+/78673808:26
openstackgerritDaniel Blixt proposed zuul/zuul-jobs master: Make build-sshkey handling windows compatible  https://review.opendev.org/c/zuul/zuul-jobs/+/78066209:23
*** ykarel is now known as ykarel|lunch09:40
*** vishalmanchanda has quit IRC09:52
*** nils has joined #zuul09:58
*** rpittau is now known as rpittau|bbl10:04
*** harrymichal has joined #zuul10:08
*** harrymichal has quit IRC10:26
*** ykarel|lunch is now known as ykarel10:47
*** vishalmanchanda has joined #zuul10:49
openstackgerritDaniel Blixt proposed zuul/zuul-jobs master: Make build-sshkey handling windows compatible  https://review.opendev.org/c/zuul/zuul-jobs/+/78066211:09
*** jcapitao is now known as jcapitao_lunch11:16
*** jpena is now known as jpena|lunch11:31
*** rlandy has joined #zuul11:46
*** rlandy is now known as rlandy|rover11:46
*** jcapitao_lunch is now known as jcapitao11:51
*** ykarel_ has joined #zuul12:20
*** ykarel has quit IRC12:21
*** jpena|lunch is now known as jpena12:33
*** rpittau|bbl is now known as rpittau12:37
*** ykarel__ has joined #zuul13:16
*** ykarel_ has quit IRC13:18
*** vishalmanchanda has quit IRC13:39
*** sean-k-mooney has quit IRC13:45
tobiashcorvus: commented on 78674413:46
tobiashapart from that lgtm13:46
openstackgerritTobias Henkel proposed zuul/zuul master: Don't check for existing refs in isUpdateNeeded  https://review.opendev.org/c/zuul/zuul/+/78720113:47
*** vishalmanchanda has joined #zuul13:47
tobiashcorvus: and I've added a slight optimization on top ^13:47
corvustobiash: thanks! i'll take a look at the getfiles usage14:02
clarkbcorvus: was a change pushed up to force ipv4 on the ssh keyscan in quickstart?14:38
clarkboh ya I see it now14:38
*** sanjayu_ has quit IRC14:39
fungiclarkb: however the nodepool jobs are also broken wrt docker around the same timeframe, making in increasingly suspicious the change could have been in docker itself?14:41
clarkbya maybe docker changed how it does the port forwards14:42
clarkbcorvus: fungi ianw https://github.com/moby/moby/pull/42205 that looks suspicious for our port forwarding problems14:54
clarkb"Fix a regression in docker 20.10, causing IPv6 addresses no longer to be bound by default when mapping ports" is how the release notes refer to that. Maybe we weren't getting an ipv6 before and that caused a different error code to ssh-keyscan when connecting, one which caused it to keep looping and try 127.0.0114:55
corvusclarkb: that sounds plausible14:55
clarkbbut now they "fixed it" which gives minimal ipv6 connectivity somewhere and that makes ssh-keyscan sad14:55
openstackgerritClark Boylan proposed zuul/nodepool master: Test if docker packages start on focal  https://review.opendev.org/c/zuul/nodepool/+/78721014:59
clarkbthat change is for more info gathering. I'm curious if docker struggles on focal the asme way as bionic15:00
avassany tl;dr what the docker/keyscan issue is?15:01
funginot yet, but seems like we might be zeroing in on it15:02
clarkbavass: the keyscan issue is that ssh-keyscan tries to connect to ::1[:29418] to keyscan gerrit but fails as there isn't anything listening there15:02
clarkbavass: I'm beginning to suspect https://github.com/moby/moby/pull/42205 is related as they "fixed" ipv6 port forwarding in the most recent release15:02
fungibecause gerrit's listening on 0.0.0.0:29418 explicitly15:02
clarkbmy hunch is that prior to this release ipv6 wasn't even attempted ebacuse it wasn't configured at all15:02
fungibut yeah, maybe the container is forwarding ::1[:29418] in somehow now?15:03
avasshuh15:03
avassshould be possible to test that by installing an older docker release without that change15:04
clarkbyup15:04
clarkb20.10.5 if you can sort out how to install it15:05
avassdoes zuul-jobs expose any docker version variable? https://docs.docker.com/engine/install/ubuntu/15:06
avassotherwise, maybe it should?15:06
clarkbno I think it is just use upstream or use distro version15:08
avassI think it should be possible to include_role and override _docker_upstream_distro__packages to use a specific version15:08
clarkbI've just noticed the failing nodepool job appears to install docker.io first (the distro package), then uninstall docker.io and install docker-ce15:09
clarkbI wonder if there is some sort of leak between docker.io and docker-ce that is breaking docker-ce15:09
clarkbI need to eat some breakfast but I'll start pulling on that thread next15:09
*** bhavikdbavishi has joined #zuul15:10
*** jfoufas1 has quit IRC15:16
openstackgerritAlbin Vass proposed zuul/nodepool master: Test docker with 20.10.5  https://review.opendev.org/c/zuul/nodepool/+/78721715:20
avassclarkb: I wonder if that ^ works15:20
openstackgerritClark Boylan proposed zuul/nodepool master: Test if not installing docker.io fixes docker-ce installs  https://review.opendev.org/c/zuul/nodepool/+/78722015:25
*** ykarel__ is now known as ykarel15:41
clarkbcorvus: I think that ^ does fix the issue. However, that was done to support zookeeper tls in nodepool15:42
*** holser has joined #zuul15:42
clarkbthe unittest jobs haev passed though so maybe that was vestigal15:42
clarkbIf that passes I can write up a better commit message and then flip the order of the two changes there (one fixes logging the other fixes the job)15:43
avassclarkb: docker.io is removed before installing docker-ce in ensure-docker for upstream installation15:44
clarkbavass: yup, and then when docker-ce is installed it doesn't work15:44
clarkbmy hunch is that uninstalling docker.io without purging it is causing somethign to leak to the docker-ce install that breaks docker-ce15:44
clarkbbut if we don't need to install docker.io in the first place we can just avoid that extra step entirely15:45
avassmakes sense15:45
corvusi wonder why we use docker-ci15:45
corvusdocker-ce15:45
avasscorvus: I was gonna question why you're using docker.io :)15:45
corvuswe might be able to flip the fix the other way around; but i don't think it's very important15:45
corvusavass: bindep+distro packages is simpler for devs15:45
corvusit's great that we have an ensure-docker role in zuul-jobs, but it's even better if we don't have to use it :)15:46
clarkbcorvus: ensure-docker in zuul0jobs defaults to upstream iirc15:46
corvusi think we wrote it when distro dockers were ancient; maybe we can stop using it now15:47
corvusclarkb: right, whereas docker.io is an ubuntu package15:47
clarkbyup15:47
clarkbwe could set use_upstream_docker: false in nodepool and zuul project defs15:48
corvusso to be clear: my point is that it seems the simplest system would be to use docker.io in bindep and not use ensure-docker in jobs.  that would be simpler, faster zuul jobs in an environment that's more easily reproducible on dev workstations.  but also, i don't think this is very important and we should merge whatever fixes this asap.15:49
clarkb(it does appear to default to true)15:49
corvusjust that after it's fixed, maybe we can try dropping ensure-docker and switching to rely on bindep15:49
clarkbcorvus: I think you still want ensure-docker, but don't want it to install upstream in your scenario15:49
corvusclarkb: why?15:49
clarkbbecause ensure-docker does a few other things with config iirc15:49
avassuse-buildset-registry etc15:49
avassthough that could just be role: use-buildset-registry instead :)15:50
clarkblooks like it modifies mtus if necessary, puts zuul in the docker group, and configures docker proxy15:50
clarkbavass: ^ no that stuff15:50
corvusokay, that's worth keeping around then :)15:50
avassclarkb: yup just saw that15:51
avassclarkb, corvus: could even do include_role: ensure-docker, tasks_from: docker-setup and then role: use-buildset-registry15:52
clarkbavass: ya though telling esnure-docker to use the distro package seems simpler?15:52
clarkbit will notice the package is already there and move on15:52
corvusavass: true, but we've lost the "simplicity" aspect of this15:52
corvusi don't think it's worth changing anything if we're not removing the ensure-docker role.  i don't really care where the packages come from :)15:53
fungiwe do something similar with ensure-pip i think15:53
fungiuse distro package by default, install with get_pip.py as an alternative15:53
avassclarkb: heh, but think of the seconds you can shave off from the job ;)15:53
corvusi just saw a potential opportunity to replace a role with a bindep entry, but if that's not possible, let's not worry about it.15:53
clarkbcorvus: ok, I can modify the change to have a better commitm essage and flip the order with its parent so they can both merge15:54
clarkbwant to make sure the tests pass though15:54
corvus++15:54
fungiif the ensure-docker role skipped installing when docker was already present, that would presumably sidestep a lot of this?15:54
clarkbfungi: that would be a non backward compatible change though15:54
fungiright15:54
clarkb(granted the current situation doesn't work)15:54
clarkbbut probably better for things to fail and people to evaluate their options than to change the assumption they haev worked with15:55
fungihowever there could be a var to tell ensure-docker to only install when there wasn't already a docker command in the path or something15:55
*** ykarel is now known as ykarel|away15:56
fungithat said, our usual approach with default behavior for ensure-.* roles is that they should make sure the tool is installed, by installing only if necessary15:57
clarkbya I think the assumption was that distro docker was unusable (until more recently) which meant if you didn't ask for it then we'd force the other thing15:58
*** jcapitao has quit IRC16:10
openstackgerritClark Boylan proposed zuul/nodepool master: Stop installing docker via bindep  https://review.opendev.org/c/zuul/nodepool/+/78722016:30
openstackgerritClark Boylan proposed zuul/nodepool master: nodepool-functional : ignore errors copying logs  https://review.opendev.org/c/zuul/nodepool/+/78706516:31
clarkbenough jobs have passed that I'm confident that ^ should work now16:31
clarkbthat is the better commit message + commit reordering stack16:31
corvusclarkb: +3 both of those16:33
clarkbcool16:33
*** sam_wan has quit IRC16:34
*** ykarel|away has quit IRC16:39
*** hamalq has joined #zuul16:45
*** hamalq has quit IRC16:47
*** hamalq has joined #zuul16:48
clarkbheh py36 failed after like 3 runs16:49
clarkbI don't think it is related to this change16:50
openstackgerritJames E. Blair proposed zuul/zuul master: Pseudo-shard unique project names in keystore  https://review.opendev.org/c/zuul/zuul/+/78698316:53
corvustobiash, swest, fungi: i think the next steps for the keystore-in-zk work is to review https://review.opendev.org/786774 and https://review.opendev.org/786983  -- ignore the previous test failures (one docker related, one linter), they are ready for review.16:55
corvusclarkb: ^ and that last change is the optional "do some extra sharding" change we talked about yesterday16:56
*** rpittau is now known as rpittau|afk16:59
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: WIP: Roles to snapshot and cleanup image builds for digitalocean  https://review.opendev.org/c/zuul/zuul-jobs/+/78675717:02
avassthere's no way to get an artifact produced by the same change but from the pipeline that merged the change other than something like: https://review.opendev.org/c/zuul/zuul-jobs/+/786757/3/roles/promote-image-digitalocean/tasks/main.yaml right?17:03
avass(that queries the zuul api)17:05
corvusavass: see download-artifact role17:07
corvusavass: it uses the api; you might be able to reuse it17:08
*** jpena is now known as jpena|off17:08
avasscorvus: I don't want to download the image, just rename it17:08
corvusavass: good that's more efficient :)  you might just double check that role for details to see if it does anything differently17:09
corvusavass: but yes, i think that's a good approach; one other approach you might consider:17:09
corvusavass: is what we do with docker images which is rather than using artifacts, we use the remote system to store the mapping17:10
corvusavass: ie, we push images with complex tags derived from the change number, then we rename those tags17:10
corvusavass: if you could do something similar, or maybe add image metadata based on change number, etc, then you could avoid using zuul api17:10
corvus(but zuul api is fine and reliable; we use that for docs promotion)17:11
avasscorvus: yeah I think i've seen those. but that could cause problems where the image is created twice (somehow)17:11
avasssince docker images overwrite while machine images in digital ocean doesn't17:11
corvusack; makes sense17:11
avassso I sort of need the exact image id17:11
openstackgerritGonéri Le Bouder proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727117:19
clarkbthat must be the leaked conflict17:20
fungiaha!17:21
fungiyeah maybe the package removal doesn't clean up the named pipe?17:21
avasswow17:21
clarkbI left a review on https://review.opendev.org/c/zuul/zuul-jobs/+/787271 based on what I learned debugging the nodepool side of things17:33
*** goneri has joined #zuul17:40
corvusgoneri: clarkb was just saying you may have found the missing piece of the docker puzzle :)17:42
clarkbalso left acomment on the change. I think we need to be more careful about when we stop the unit17:42
corvusmight be interesting to do a no-op nodepool change that depends on 787271 before 787220 lands17:43
clarkbin particular if that unit doesn't already exist wouldn't stopping it be an error?17:43
corvus(i'm busy with other things right now; if anyone wants to try that be my guest)17:43
goneriHi! pabelanger told me to come here :-)17:53
clarkbgoneri: hello, I left a comment on https://review.opendev.org/c/zuul/zuul-jobs/+/787271 you may want to check17:59
goneriI actually just answered.18:00
clarkbgoneri: if it wasn't installed previously then what started that unit?18:01
clarkbsince it looks like in the ara that you skip the restart even18:02
goneriI think it's a post-install script from the rpm that starts the socket too early.18:02
clarkbhuh I thought rpms generally didn't start services. But I guess this is the upstream package so they can do what they want18:04
clarkbin the ubuntu case it seems to only be an issue if docker.io was there previously (though possibly also if docker-ce was there previously)18:05
clarkbbtu I guess if we're doing this late enough that the unit has started and we subsequently restart docker itself then this should be fine?18:05
gonerion Debian/Ubuntu, (sadly) the services are started during the installation.18:05
goneriyup, this should addresses both cases. https://github.com/kata-containers/tests/issues/3103 is with Ubuntu actually.18:06
gonerihttps://github.com/docker/docker-ce-packaging/blob/master/rpm/SPECS/docker-ce.spec#L11518:07
*** iurygregory has quit IRC18:15
openstackgerritMerged zuul/nodepool master: Stop installing docker via bindep  https://review.opendev.org/c/zuul/nodepool/+/78722018:35
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: WIP: Roles to snapshot and cleanup image builds for digitalocean  https://review.opendev.org/c/zuul/zuul-jobs/+/78675718:35
*** bhavikdbavishi has quit IRC18:46
openstackgerritMerged zuul/nodepool master: nodepool-functional : ignore errors copying logs  https://review.opendev.org/c/zuul/nodepool/+/78706518:47
*** ajitha has quit IRC18:56
*** vishalmanchanda has quit IRC19:09
openstackgerritGonéri Le Bouder proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727119:15
openstackgerritGonéri Le Bouder proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727119:50
openstackgerritGonéri Le Bouder proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727120:44
*** odyssey4me has quit IRC21:05
*** odyssey4me has joined #zuul21:06
goneriI don't think the last CI problems are coming from my PR: https://review.opendev.org/c/zuul/zuul-jobs/+/78727121:18
mordredgoneri: I agree, https://zuul.opendev.org/t/zuul/build/ade3b9c4127546f1873cecac8d3a434f doesn't seem particularly related21:24
ianwgoneri: i think we started running a periodic job for all zuul-jobs, we could/should look back and see where it started failing21:25
ianwi'd say here : http://lists.zuul-ci.org/pipermail/zuul-jobs-failures/2021-April/000028.html21:26
mordredianw: kube 1.21: Remove deprecated --generator, --replicas, --service-generator, --service-overrides, --schedule from kubectl run Deprecate --serviceaccount, --hostport, --requests, --limits in kubectl run21:30
mordredsorry, 1.20 actually21:30
mordredhttps://kubernetes.io/docs/setup/release/notes/21:30
mordredhttps://github.com/kubernetes/kubernetes/pull/9973221:31
mordredaccording to the PR, the generator flag  was a no-op21:32
mordredhttps://github.com/kubernetes/kubernetes/pull/99732/files#diff-bbbdda93ca43398e7c554a57f7934e126ed841d46078afe8d601edc2f695b4f9L177-L17821:32
ianwunfortunately i guess it means squishing changes into the docker fix21:33
mordredyeah21:35
mordredrun-pod/v1 <-- it seems that is the only generator that does anything - so it seems like, givne our use, removing the parameter is appropriate21:35
*** fsvsbs has joined #zuul21:36
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: Remove use of --generator=run-pod/v1  https://review.opendev.org/c/zuul/zuul-jobs/+/78729121:43
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: Remove use of --generator=run-pod/v1  https://review.opendev.org/c/zuul/zuul-jobs/+/78729121:43
mordredianw, goneri : ^^ I made that on top of the docker change - it should turn things green. if it does, we can squash21:44
ianwmordred: LGTM, thanks -- i can keep and eye and do the squash today21:44
mordredI'm curious if there is a way we could have caught this differently. I can't think of a good way21:45
ianwi guess modulo noticing deprecated arg warnings, just more scrutiny of the periodic job failure report21:47
mordredyeah. like - failing on deprecated args would have just shifted when the job failed - so not really much of a thing21:54
openstackgerritGonéri Le Bouder proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727122:00
ianwgoneri: ^ we don't actually want that depends-on the other change; we want to test mordred's change applied ontop of yours, and then to get through gate we'll need to squash them22:02
goneriok, sorry.22:03
ianwit's not always exceedingly clear in gerrit, but you can see "Relation chain" on the right hand side22:03
openstackgerritGonéri Le Bouder proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727122:04
mordredyeah - if you want to re-push yours up without that depends on (and while you;re at it, go ahead and take out the AWS line, since it's no longer true) - and I can rebase mine22:04
*** bodgix has quit IRC22:07
*** bodgix_ has joined #zuul22:07
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: Remove use of --generator=run-pod/v1  https://review.opendev.org/c/zuul/zuul-jobs/+/78729122:30
mordredianw, goneri: my patch went green on the k8s tests but broke on openshift - turns out I shouldn't have removed the arg from oc - just from kubectl *sigh*22:30
mordredtristanC: ^^ you know more things about openshift - kubectl has deprecated the generator argument but oc has not - should we assume oc will also deprecate it and should we do anything to future-proof ourselves?22:32
*** holser has quit IRC22:33
clarkbmordred: does your change need to be rebased under the other change?22:35
mordredclarkb: it's sitting on top of it22:39
mordredclarkb: we'll need to squash the two to land it22:39
*** tosky has quit IRC23:17
*** nils has quit IRC23:25
*** goneri has quit IRC23:27
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: Remove use of --generator=run-pod/v1 for oc  https://review.opendev.org/c/zuul/zuul-jobs/+/78730023:41
tristanCmordred: that sounds good, let's see with ^23:41
mordredI'm going to squash the two changes, the second is now green23:43
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: ensure-docker: ensure docker.socket is stopped  https://review.opendev.org/c/zuul/zuul-jobs/+/78727123:45
mordredtristanC, clarkb, corvus : ^^23:45
mordredthat's the squashed version of goneri's and my change. it should be green and good to go23:46
clarkbone thing I noticed in a recent ps is that we use a handler to start and stop the docker service23:46
clarkbare ansible handlers processed in order?23:46
clarkbotherwise we could start then stop23:46
openstackgerritJames E. Blair proposed zuul/zuul master: Lock node requests in fake nodepool  https://review.opendev.org/c/zuul/zuul/+/78730123:48
ianwcorvus: thanks!  i was just trying to figure that out  :)23:49
corvusianw: oh you saw that race too? :) that one hit me here: https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_332/786774/3/check/zuul-tox-py36/33292a2/testr_results.html23:50
corvuspresents as a failed nodepool request in the scheduler log, which should never happen with a fake nodepool (unless we intentionally fail)23:51
ianwnodepool.exceptions.ZKLockException: Did not get lock on /nodepool/nodes/0000000000/lock23:51
ianwhttps://zuul.opendev.org/t/zuul/build/8709fbeed01847949faeda8b7b4d8a5923:51
ianwis what i was looking at23:51
ianwwell i'm presuming it's the same thing as it involves the word "lock" :)23:52
corvusianw: could be related, but not 100% sure; my thing fixes a lock on the requests; that could cascade to another failure.  that error is at least a little different.23:53
corvusianw: oh that's nodepool; definitely not related23:54
ianwok then, back to the drawing board :)23:55
corvusclarkb: handlers are run in the order *defined* (!), we define docker.socket stop before docker.socket and docker restart23:55
corvusso i think that's gtg23:55
clarkbTIL23:56
corvusme too23:56
corvusclarkb: would you mind giving https://review.opendev.org/786983 at least a quick review on the idea?23:57
clarkbya the idea of using the "org" prefix makes sense23:58
corvuskk; hopefully we can merge that stack tomorrow23:59

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!