Monday, 2019-04-15

openstackgerritPaul Belanger proposed openstack-infra/zuul-jobs master: DNM: add ansible_network_os to vars  https://review.openstack.org/65242400:25
*** jamesmcarthur has joined #zuul02:19
*** jamesmcarthur has quit IRC02:26
*** jamesmcarthur has joined #zuul02:26
*** jamesmcarthur has quit IRC02:32
*** jamesmcarthur has joined #zuul02:46
*** jamesmcarthur has quit IRC02:49
*** jamesmcarthur has joined #zuul02:50
*** jamesmcarthur has quit IRC02:54
*** bhavikdbavishi has joined #zuul03:25
*** bhavikdbavishi1 has joined #zuul03:28
*** bhavikdbavishi has quit IRC03:30
*** bhavikdbavishi1 is now known as bhavikdbavishi03:30
*** jamesmcarthur has joined #zuul03:31
*** raukadah is now known as chandankumar03:53
*** bhavikdbavishi1 has joined #zuul04:48
*** bhavikdbavishi has quit IRC04:49
*** bhavikdbavishi1 is now known as bhavikdbavishi04:49
*** bhavikdbavishi1 has joined #zuul04:53
*** bhavikdbavishi has quit IRC04:54
*** bhavikdbavishi1 is now known as bhavikdbavishi04:54
*** chandankumar has left #zuul05:21
*** chandankumar has joined #zuul05:21
*** quiquell|off is now known as quiquell|rover05:46
*** bjackman has joined #zuul05:55
*** pcaruana has joined #zuul06:19
*** toabctl has joined #zuul06:32
*** gtema has joined #zuul07:02
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Add --check-config option to zuul scheduler  https://review.openstack.org/54216007:18
*** hashar has joined #zuul08:12
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Add support for smart reconfigurations  https://review.openstack.org/65211408:12
*** shanemcd has quit IRC08:22
*** shanemcd has joined #zuul08:23
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Add --check-config option to zuul scheduler  https://review.openstack.org/54216008:26
*** jangutter has joined #zuul08:39
*** electrofelix has joined #zuul09:05
*** gtema has quit IRC09:46
*** zbr has joined #zuul09:47
*** zbr__ has quit IRC09:50
*** panda has joined #zuul09:51
*** bhavikdbavishi has quit IRC09:59
bjackmanHow can I add known host keys to Zuul for a Git connection?10:10
bjackmanI can always just manually add them in the Zuul scheduler container but I'd rather do it in a way that persists..10:12
*** gtema has joined #zuul10:30
electrofelixbjackman: seems like you'd want to mount in from a volume?10:50
*** quiquell|rover is now known as quique|rover|eat11:06
*** bhavikdbavishi has joined #zuul11:10
*** mhu|off has quit IRC11:12
*** weshay_pto has quit IRC11:12
*** mhu has joined #zuul11:13
*** weshay_pto has joined #zuul11:13
*** panda is now known as panda|lunch11:23
*** gtema has quit IRC11:35
*** quique|rover|eat is now known as quiquell|rover11:49
*** rlandy has joined #zuul12:03
*** rlandy is now known as rlandy|ruck12:03
*** panda|lunch is now known as panda12:08
*** jamesmcarthur has joined #zuul12:24
*** rlandy|ruck is now known as rlandy|ruck|mtg12:33
*** gtema has joined #zuul12:34
*** jamesmcarthur has quit IRC12:35
bjackmanelectrofelix, yeah I suppose that's what I'll need to do12:37
bjackmanI was hoping there was some Zuul config I was missing that could do it fo rme12:37
bjackmanI was hoping there was some Zuul config I was missing that could do it for me12:37
*** jamesmcarthur has joined #zuul12:46
electrofelixbjackman: we have added some stuff to containers to retrieve the keys on start up, it's not really best practice though12:46
*** rfolco has joined #zuul12:47
bjackmanelectrofelix, right12:48
electrofelixby we I mean another operator of zuul ;-)12:48
*** quiquell|rover has quit IRC12:55
*** quiquell has joined #zuul12:56
*** bhavikdbavishi has quit IRC13:03
*** rlandy|ruck|mtg is now known as rlandy|ruck13:05
*** bjackman has quit IRC13:05
*** bjackman has joined #zuul13:06
*** jamesmcarthur has quit IRC13:12
bjackmanCaught my nodepool instance in its apparent deadlock again - this time I dumped the request-list and node-list from the nodepool webapp: https://paste.gnome.org/pyzdqk4gq13:13
bjackmanIt has a pending request and 4 "ready" nodes but for some reason isn't servicing that pending request13:13
bjackmanAny ideas why that might be?13:14
sshnaidmhow can I trigger a job in zuul if new container is uploaded to docker.io for example?13:32
*** jamesmcarthur has joined #zuul13:35
*** jamesmcarthur_ has joined #zuul13:46
*** jamesmcarthur has quit IRC13:49
Shrewsbjackman: the nodepool logs should say what's going on14:04
pabelangersshnaidm: you can add a job dependency after the upload to docker job runs14:05
*** rlandy|ruck is now known as rlandy|ruck|mtg14:05
pabelangeror are you asking about triggering another job in a different pipeline?14:05
*** gtema has quit IRC14:16
*** gtema has joined #zuul14:18
*** gtema has quit IRC14:19
*** hashar has quit IRC14:21
*** bjackman has quit IRC14:37
mordredpabelanger: I believe sshnaidm is looking for "I want a zuul trigger plugin that triggers on someone uploading a new image to dockerhub"14:40
sshnaidmmordred, yeah, exactly14:40
mordredpabelanger: so that one could, for instance, do a rebuild of zuul images if the underlying python base image got an update14:41
sshnaidmpabelanger, seems like I need two jobs, one will monitor and second will run checks14:41
mordredsshnaidm: at the moment there isn't a way to do this other than using a periodic pipeline - or something that watches dockerhub and submits patches to a git repo14:42
pabelangerwas just going to say that14:42
sshnaidmmordred, I see14:42
mordredit's a type of use case that's come up a few times before, but so far nobody has been able to fully describe what such a trigger would look like in the wider context of the system - so I wouldn't be surprised if we grew that ability, but I would be surprised if we grew that capability in the short-term, simply because there are some deeper design questions that come in to play14:43
*** quiquell is now known as quiquell|off14:46
*** bjackman has joined #zuul14:48
bjackmanShrews, was about to say that the nodepool logs seem to suggest the request was satsified - but when I went to grab something to paste I noticed that earlier on in the logs (I guess when the request first came in) is a kazoo.exceptions.ConnectionLoss14:54
bjackmanWill keep digging14:54
bjackmanhttps://paste.gnome.org/p2tpzkkmz14:55
*** rlandy|ruck|mtg is now known as rlandy|ruck14:56
Shrewsbjackman: that's ok, we expect the zk connection to become temporarily unavailable at times because "networks" and try to recover correctly, which it seemed it did. I see that the request was actually satisfied ("Fulfilled node request 200-0000002869") and the handler removed, so that request should NOT be pending anymore at that point. That's VERY confusing to me.15:02
Shrewsbjackman: if you give the --detail option to 'nodepool list', does it show that node 0000002811 is assigned to that request?15:03
Shrewsbjackman: the way it works is, a node is assigned to a request, but it's up to the zuul-scheduler to then change that node's status from READY to IN-USE15:04
Shrewsbjackman: so you may need to dig into zuul logs as well for that request15:04
bjackmanShrews, sorry I had to restart the launcher as this is all happening on prod15:05
bjackmanSo can't dig any further (zuul-scheduler logs will have wrapped by now, too)15:05
Shrewsoh, that's unfortunate. well, now you know what else to look for next time  :)15:05
bjackmanBut this happens pretty often so I will look in more detail next time15:05
bjackmanShrews, indeed, thanks15:05
bjackmanHas been happening for a while but each time it happens, I haven't had time to really dig into it15:06
bjackmanOnly found out about the nodepool web UI today15:06
Shrewsbjackman: could very well be bugs in the static driver, but hard to tell from just what you pasted. i thing tobiash uses that driver quite often. not sure he sees such problems15:06
bjackmanShrews, I don't really understand why there are deleted nodes, either15:07
bjackmanIs that expected?15:07
Shrewsbjackman: you're running the latest versions of nodepool/zuul?15:07
bjackmanNo zuul is 3.6.2.dev4 9abc640315:07
Shrewsbjackman: those should be transient. i wouldn't expect those to live very long15:07
bjackmanLet me see how to find Nodepool version..15:07
bjackmanAh - it's just happened again. Will gather everything I can...15:07
bjackmanAh the user uploaded a new patchset so the stalled change got dequeued before I could finish15:17
bjackmanShrews, anything I can do to look into the persistent "deleted" nodes in the meantime?>15:18
bjackmanShrews, anything I can do to look into the persistent "deleted" nodes in the meantime?15:18
Shrewsbjackman: pulling the log info for any of those deleted nodes would be very helpful, but looks like they've been around a long time, so you may not have that anymore15:21
Shrewsi think i can see a code path where losing the zookeeper connection in the middle of a delete could leave that state hanging around, but would be nice to have logs to confirm that15:22
Shrewsmay have to come up with a test for that...15:23
tobiashI think also a zk session timeout option could help here (we have that already in zuul)15:24
Shrewstobiash: i think getting an exception from zk.deleteRawNode() that we don't handle would leave a znode in the DELETED state, and that's not one of the standard cleanup states we look for in DeletedNodeWorker._cleanupNodes(). I think I can easily create a test to prove that.15:26
tobiashShrews: cool15:27
Shrewsyou'd have to be extremely unlucky with timing for that to happen, but looks like bjackman has seen that more than once  :(15:29
Shrews(if that is the problem)15:29
tobiashbjackman: what disks/san are you running zk on?15:30
tobiashbjackman: initially I ran it on ceph which was very unstable15:30
bjackmantobiash, everything is just running via docker-compose on the same physical machine in our office15:32
Shrewsooh15:32
Shrewsthat's, uh, not recommended15:32
Shrewsfor production15:32
bjackmanShrews, unfortunately in my industry, top brass tends to reject "cloud solutions"15:32
Shrewsyou're going to get a lot of zookeeper connection issues15:32
bjackmanShrews, why would that cause connection issues?15:33
tobiashbjackman: the problem is IO, if zookeeper is stalled for more than 10s by default, the clients loose their session15:34
Shrewszookeeper is very i/o intensive, and you'll see it stall a lot.15:34
tobiashand with the session the locks etc15:34
Shrewsyeah, that15:34
bjackmanAh i see15:34
bjackmanI don't think there's much scope for changing the situation there15:35
tobiashbjackman: you could think about putting it on a tmpfs and regularly copy the snapshots back to a disk by a different process15:35
bjackmanWe have zillions of huge servers but they all run Red Had Linux 199515:35
bjackmanFor HW dev stuff15:35
bjackmanOnly a couple of things modern enough for containers15:36
Shrewsgiven that architecture, i'm not surprised you're hitting this delete issue so much then. i can try and make that better for that case, but be aware you may hit more such odd cases we haven't quite ironed out yet15:38
Shrews(not that i'm aware of any, but they're probably there)15:38
bjackmanShrews, tobiash so do you think the issue would go away if zk had its own host?15:44
tobiashbjackman: possibly (if on ssd)15:44
pabelanger+1 for moving zk to own host15:44
bjackmanIt seems like this issue is eventually going to come back at some scale or another though no?15:46
bjackmanIn my case the max number of parallel jobs in the entire system is 815:46
bjackmanThe machine has 32 cores (granted, only one storage volume)15:48
pabelangerRunning mergers / executors also need a lot of disk IO15:49
pabelangerso, after splitting off zookeeper, I could see you also wanting to split them off15:49
bjackmanYeah I can see that that's going to pound the disks15:49
bjackmanBut there must be a way to make the impact of the IO-thrashing that the system is slow, instead of that it gets stuck15:51
pabelangergiven that the executor runs a merger, and you are all-in-one, you might be able to stop zuul-merger process and just rely on zuul-executor. that could save a little io15:51
bjackmanpabelanger, I have a feeling I already have that situation - my setup is based on the Getting Started Guide in the Zuul repo15:52
bjackmanI don't have a separate service for the merger15:52
Shrewsbjackman: Yeah, obviously getting stuck is not what we want. We want to fix such bugs causing that. But running everything in such a constrained environment, you're more likely to hit such bugs... which would be helpful for finding them, but not so convenient for you, I'm afraid.  :(16:11
Shrewsi'm curious as to which piece is getting "stuck" for you, though. feels like more zuul-side based on the logs you presented earlier16:12
Shrewsthe i/o contention could certainly make things "seem" stuck16:13
*** hashar has joined #zuul16:13
SpamapSDon't forget that Zookeeper is *extremely* sensitive to IO load.16:14
SpamapSYou won't see anything in top, but Zookeeper will feel almost like it's dead if it goes into IO wait for even 5% of the time.16:14
ShrewsSpamapS: yep, first thing we noted. he's certainly seeing that16:14
SpamapSk, I'm reading backwards16:16
SpamapSif it's just a toy, tmpfs is a real win there.16:16
SpamapSmordred:also regarding your "I want to trigger when something happens elsewhere"... a webhook trigger driver would be amazing.16:17
clarkbSpamapS: fwiw the disk on the registry filled up again after I freed ~700MB16:18
SpamapSclarkb: :(16:18
SpamapSclarkb:what's going on, too many transient images?16:18
clarkbso we'll have to figure that out today. Probably by nuking the registry contents and starting over. Then make plans for garbage collection and tag deletion periodically16:18
clarkbSpamapS: ya its filling the 40GB device with our zuul and nodepool and gitea builds16:19
clarkbwe aren't running garbage collection which won't do much anyway since you have to untag things for that to work, but running by hand hits an error16:19
clarkbI can't even start a container now to try and run it by hand though because that requires making changes on /16:19
SpamapSAh yeah, with our AWS ECR repos we set some policies to delete tags based on patterns. Works well. we keep 25 change_XXX's, 25 newrev_XXX's, all pipeline_XXX's, and never delete anything until it is at least 7 days old.16:20
clarkbSpamapS: ya I think we'll want similar. But also we need to get the garbage collector working which I think means destroying the registry (unless someone wants to go one by one identifying errosr and manually deleting the bad sha blobs)16:21
SpamapSclarkb:why not build a new registry?16:26
* SpamapS prefers to cloud hard16:26
clarkbits the same difference really. Do you rm -rf /var/lib/registry or openstack server delete16:26
SpamapSbut anyway, just tell me when we can recheck patches so they can land. ;)16:26
SpamapSclarkb:me personally, I always server delete. :)16:26
clarkbthe biggest downside to ^ is you don't learn as much about the failure modes16:26
SpamapSactually I server stop, server create.. get working.. then delete.16:27
SpamapSYou don't want to learn about failure modes, you want to automate yourself into a place where you avoid them. :)16:27
clarkbyou want both16:27
SpamapSmeh.. not realy.16:27
SpamapSyou might :)16:27
clarkbyou can't avoid the failure modes if you don't understand them16:28
SpamapSAnd I always archive root volumes for the RCA16:28
clarkbsure you can rebuild and start over but then you've got a 10 minute outage16:28
SpamapSDon't you have an outage now?16:28
clarkbyes, but if we can udnerstand the failuer and prevent it then no more outages16:28
clarkbis my point16:28
clarkbfor example it may be that the disk is super undersized16:28
SpamapSI think my point is that I prefer to decouple the learning from the recovering. :)16:28
clarkbthat is something that you need to feed back into the rebuild16:29
clarkboh sure16:29
SpamapSBut also I think if you build a new one you might see it fill up fast again.. seems like you have more than a GC problem.. you have a dangling reference problem.16:29
SpamapSclarkb:anyway, don't let me gum up your process with opinions. If I can help, let me know. And if I can recheck, also let me know. :)16:30
clarkbI took a quick look on the weekend my apologoes if I didn't fix it quickly enough for your preferences :P16:30
clarkbthe other issue with GC is you cannot run it whne the registry is online16:31
clarkbwhich severely limits our ability to GC I think16:31
SpamapSwow, that seems.. wow.16:32
SpamapSgood job docker16:32
*** bhavikdbavishi has joined #zuul16:32
SpamapSMy preference is for everything ot be 5 nines and in the minutes of downtime a year I prefer to be fed grapes.16:33
SpamapS;-)16:33
SpamapSsorry if I push.. just peeking in.16:33
*** bhavikdbavishi1 has joined #zuul16:35
clarkbah ok re GC happening when server is offline it doesn't always require that, only when you don't use the api to mark tags for deletion16:36
clarkbso we should use the api for that I guess16:36
clarkb(odd that people wouldn't use the api)16:36
*** bhavikdbavishi has quit IRC16:36
*** bhavikdbavishi1 is now known as bhavikdbavishi16:36
openstackgerritFabien Boucher proposed openstack-infra/zuul master: WIP - Pagure driver - https://pagure.io/pagure/  https://review.openstack.org/60440416:41
openstackgerritClark Boylan proposed openstack-infra/zuul master: Fix tox.ini cover target install command  https://review.openstack.org/65272716:45
corvusSpamapS: there's a WIP url trigger change16:50
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Fix loss of ZK conn during node delete  https://review.openstack.org/65272916:51
Shrewstobiash: bjackman: I think the test in that ^^ demonstrates the orphaned DELETED node problem16:51
SpamapScorvus:cool. I have a need to run some jobs whenever an EC2 autoscaling group changes membership.. and was just thinking "hm, how can I do this in Zuul?"16:52
tobiashzuul-maint: I'd love a second review on https://review.openstack.org/634597 and parent which should make job canceling more robust and handle failed paused jobs better17:06
*** gtema has joined #zuul17:08
*** gtema has quit IRC17:10
mordredtobiash: done17:22
tobiashmordred: \o/17:22
clarkbtobiash: mordred note that cannot merge right now due to the broken docker registry (I think)17:23
clarkbI will hopefully have it up and running shortly17:23
tobiashthat will be an interesting race, gate vs clarkb :)17:24
clarkbmy money is on the gate. I'm taking a snapshot really quickly and those never seem to be a predictable cost17:24
*** hashar has quit IRC17:37
electrofelixexperimenting with zuulv3 where's the best place to configure common env variables such as proxies that need to be set for jobs? As we might have some hardware behind a proxy and some in AWS is it better done on a per node basis rather than job basis?18:03
*** jamesmcarthur_ has quit IRC18:04
corvuselectrofelix: to set an env var, you'll have to have ansible do that -- so either your task, role, or play will need to set the env variable -- like this: https://docs.ansible.com/ansible/latest/user_guide/playbooks_environment.html18:16
corvuselectrofelix: to deal with region-specific information (like a specific proxy for this region)18:17
corvuselectrofelix: you can use information that nodepool provides to zuul to construct the contents of the variable18:17
corvuselectrofelix: let me dig up links for how we set our region-local mirrors in opendev's zuul18:18
corvuselectrofelix: this is the contents of our site-variables.yaml file: https://opendev.org/openstack-infra/project-config/src/branch/master/zuul/site-variables.yaml18:23
corvuselectrofelix: see how the zuul_site_mirror_fqdn variable is constructed18:23
corvuselectrofelix: the executors include that file in every job that's run (via https://zuul-ci.org/docs/zuul/admin/components.html#attr-executor.variables )18:24
corvuselectrofelix: so all those variables are always automatically set for ansible18:24
corvuselectrofelix: you could do something that, and then use that to set the http_proxy environment variables in ansible18:25
corvuselectrofelix: or you can use it in roles like this (which we use to configure mirrors/proxies/etc): https://zuul-ci.org/docs/zuul-jobs/general-roles.html?highlight=configure%20mirrors#role-configure-mirrors18:26
SpamapSThere's also site variables for shared values.18:29
*** bhavikdbavishi has quit IRC19:01
*** jamesmcarthur has joined #zuul19:01
clarkbSpamapS: registry should be working again now19:16
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Support fail-fast in project pipelines  https://review.openstack.org/65276419:32
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Support fail-fast in project pipelines  https://review.openstack.org/65276419:39
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Support fail-fast in project pipelines  https://review.openstack.org/65276419:53
clarkbno responses to either my bug or PR against docker for the ipv6 issue19:59
clarkb:/19:59
fungianybody else have trouble with react-scripts build? i can't figure out how to get it to provide more detail: http://paste.openstack.org/show/749324/20:00
fungiis the "EEXIST: file already exists" error coming from it, or something else?20:00
fungior are those fsevents compatibility checks relevant?20:01
fungii'm gonna take a break to find some food, but will continue poking at it when i return20:01
clarkbfungi: did you start with a clean repo (eg git clean -xfd)?20:02
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Fix for orphaned DELETED nodes  https://review.openstack.org/65272920:02
fungiyep, and this is master branch tip of zuul with no changes20:02
fungioh!20:03
fungi`git clean -dfx` is skippnig web/build20:03
fungiit says "Removing web/node_modules/"20:03
fungibut it leaves web/build intact20:03
fungiwhy?20:03
fungiaha, web/build is a symlink to ../zuul/web/static/20:04
fungimaybe git clean won't remove links?20:04
fungioh, nope, that web/build symlink is actually tracked in git, that's why it doesn't remove it20:05
fungiso i'm guessing the EEXIST in the output isn't relevant20:06
fungisince that path should exist20:06
corvusmordred: ^20:06
fungianyway, back shortly20:06
mordredso many words20:07
fungimordred: the paste is me trying to run `tox -e py37` after installing everything bindep wants20:08
* fungi is *really* heading out now, before christine gnaws my leg off20:08
mordred*weird* - looking20:08
clarkbis it possible nodeenv doesn't build node if you have node installed locally? and if so do you have node installed locally?20:10
mordredI mean - it shouldn't matter too much - that paste shows node v8 being installed20:11
mordredfungi: incidentally - the fs-events thing is sadly unavoidable but also is always there. it's a warning from a dependency20:12
SpamapSclarkb:thank you20:18
SpamapSmordred:it's not *always* there.. just on all production-ready platforms (Linux) ;)20:19
mordredSpamapS: like I said - it's always there :)20:19
SpamapSI'm always there too.20:19
* SpamapS creeps20:19
* mordred cannot hide from the watchful eyes of SpamapS20:19
clarkbfwiw I have py37 locally too and running tox -epy37 after a git clean -xfd worked. WHich makes me wonder about platform stuff20:21
clarkbI have nodejs v8.11.1 installed in the virtualenv too20:22
*** jamesmcarthur has quit IRC20:37
*** pcaruana has quit IRC20:38
mordredclarkb: I just ran the same stuff with py37 - but i have packages installed so I didn't get things instlaled into nodeenv - and it worked20:38
*** jamesmcarthur has joined #zuul20:39
clarkbis it possible fungi's link is broken?20:41
*** sshnaidm has quit IRC20:41
clarkbI know that I've run into weirdness with the .keep file locally20:41
clarkb(git keeps thinking I've deleted it for some reason20:41
openstackgerritPaul Belanger proposed openstack-infra/nodepool master: Gather host keys for connection-type network_cli  https://review.openstack.org/65277820:42
pabelangerwould love a review of ^20:43
pabelangerhelps gather SSH host keys when using conneciton-type network_cli20:43
pabelangerShrews: ^if you don't mind adding to your review queue20:43
pabelangerso far, testing with vyos and zuul is working20:49
*** jamesmcarthur has quit IRC20:56
*** sshnaidm has joined #zuul20:57
fungiokay, back and seeing what we've got21:00
fungiso... i wonder whether `react-scripts build` has a way to... like... actually say *why* it "failed to compile"21:02
fungii tried adding --verbose to its invocation in web/package.json, to no avail21:03
fungieven though it's mentioned in the output, https://yarnpkg.com/en/docs/cli/run is surprisingly unhelpful here21:03
clarkbfungi: adding --verbose to the command in the json file would've been what I tried. The docs don't seem to document flags21:08
clarkbmight be able to find the source and check /me looks21:08
clarkbhttps://github.com/facebook/create-react-app/blob/master/packages/react-scripts/bin/react-scripts.js21:09
fungiit's likely just my fault for trying this on debian21:09
clarkbhttps://github.com/facebook/create-react-app/blob/master/packages/react-scripts/scripts/build.js21:09
fungiwoo, yeah, not a lot of options there21:10
clarkbfungi: https://github.com/facebook/create-react-app/blob/master/packages/react-scripts/scripts/build.js#L57-L6021:10
clarkbit could be that it is failing there and that is where your mkdir error comes from?21:11
fungithat function *is* remarkably devoid of user feedback, so maybe21:12
pabelangertobiash: 2.8.0b1 of ansible was just cut, I'm going to poke into your 2.8 patch for zuul and see why logging is failing21:15
clarkbfungi https://github.com/facebook/create-react-app/blob/master/packages/react-scripts/scripts/build.js#L12621:15
clarkbfungi: I read that as it says "failed then the error youv'e got there about mkdir is in fact the error that caused it to fail"21:15
openstackgerritMerged openstack-infra/nodepool master: Implement max-servers for AWS driver  https://review.openstack.org/64947421:17
fungiokay, which leads me to wonder why it's concerned about that directory existing when i run it, but not for others, given the symlink and directory it targets are both tracked in the git repo21:20
*** jamesmcarthur has joined #zuul21:27
openstackgerritMerged openstack-infra/zuul master: encrypt: Fix SSL error when using file pubkey  https://review.openstack.org/65058921:29
clarkbalso mkdir should be fine creating a dir that already exists21:34
fungiwell, mkdir -p is21:38
fungimkdir will exit nonzero21:38
clarkbcould it be a permissions issue?21:38
clarkbwhere it doesn't see the dir so it tries to recreate it?21:38
fungiseems unlikely. everything in this repo is owner by my user, same one which is executing toc21:39
fungitox21:39
fungibut now i've tried deleting that symlink and tox hangs indefinitely on develop-inst21:39
fungilooks like the leafmost process is node running web/node_modules/react-scripts/scripts/build.js21:40
funginevermind, not hung, just slow. now it's building an ansible wheel21:42
fungior so it appears from the process list21:42
clarkbit should isntall all the ansible versions for testing21:42
clarkbso that is potentially good news21:42
fungilikely courtesy of today's ansible release21:42
fungireleaseS21:42
fungiwell lookie there, it's running tests now21:46
fungibut only after i delete that symlink21:46
fungileading me to wonder *why* nobody else is encountering this issue. it's a symlink shipped in the git repo, so it should be present for everyone21:46
fungigotta be something about my machine21:47
clarkbyes the symlink is present for me21:47
fungionce this tox run eventually completes, i'll repeat with and without the symlink just to be sure that's the deciding factor21:50
*** jamesmcarthur has quit IRC21:52
openstackgerritMerged openstack-infra/zuul master: Centralize job canceling  https://review.openstack.org/64060921:52
*** jamesmcarthur has joined #zuul21:55
*** jamesmcarthur has joined #zuul22:00
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: upgrade react and react-scripts to ^2.0.0  https://review.openstack.org/63190222:07
clarkbfungi: ^ heh I wonder if that is realted22:08
fungifamiliar words22:09
fungimaybe tristanC is on my wavelength again22:09
tristanCfungi: that *might* help, assuming react-script>2 has more fix/improvement than the curent (old) version zuul is using22:11
fungiyeah, still no clue why on debian/unstable react-scripts build seems to care that the web/build symlink exists22:12
openstackgerritMerged openstack-infra/zuul master: Reset dependent jobs when aborting paused job  https://review.openstack.org/63459722:22
*** jamesmcarthur has quit IRC22:25
openstackgerritClark Boylan proposed openstack-infra/zuul master: Fix dynamic loading of trusted layouts  https://review.openstack.org/65278722:26
openstackgerritClark Boylan proposed openstack-infra/zuul master: Config errors should not affect config-projects  https://review.openstack.org/65278822:26
clarkbpabelanger corvus mordred ^ fyi22:40
corvus+322:41
*** mattw4 has joined #zuul22:41
tristanCclarkb: shouldn't the storyboard story be public now?22:47
clarkbtristanC: probably22:47
clarkbsorry I'm juggling a release note change and an email right now. mabye corvus or fungi can make it public?22:48
corvuscan do22:48
corvusdone22:48
corvustristanC: ^22:48
openstackgerritClark Boylan proposed openstack-infra/zuul master: Add release note for broken trusted config loading fix  https://review.openstack.org/65279322:52
clarkbcorvus: fungi tristanC mordred pabelanger ^22:52
clarkbany idea if this will be a 3.8 or 3.7.2 release?22:54
corvusclarkb: i think it should be 3.8.0 due to the artifacts change22:57
corvushttps://zuul-ci.org/docs/zuul/releasenotes.html#in-development22:57
clarkbthanks22:57
corvusclarkb: so you can link to https://zuul-ci.org/docs/zuul/releasenotes.html#relnotes-3-8-022:58
pabelangercatching up22:58
clarkbhttps://etherpad.openstack.org/p/tBpqrLtzsP is my draft email;22:59
pabelanger+3 on 65279322:59
corvusclarkb: i think that covers it, other than striking "-ci" from "zuul-ci" :)23:00
clarkbdone23:00
pabelangerclarkb: lgtm23:02
clarkbk I think we are just waiting for things to merge then23:02
* clarkb finds some tea. it is quite cold today for some reason23:03
pabelangerHmm23:04
pabelangerwe might be having an outage in rax23:04
corvusgood timing23:04
pabelangermoving to #openstack-infra23:05
corvusShrews, tobiash: based on http://cacti.openstack.org/cacti/graph.php?action=zoom&local_graph_id=63831&rra_id=4&view_type=&graph_start=1522317252&graph_end=1555370436 it looks like we may have introduced a memory leak in nodepool-launcher late last year; we should dig into that soon23:21
corvusmy first thought is the zk tree caching stuff...23:22
*** bjackman has quit IRC23:38

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!