Monday, 2017-04-10

*** hashar has joined #zuul07:06
*** dmellado has joined #zuul07:08
*** bhavik1 has joined #zuul08:11
*** bhavik1 has quit IRC09:51
tobiashhi, just curious, was there a reasoning why zookeeper was chosen from the set zookeeper, etcd, consul, ...?10:11
*** jamielennox|away is now known as jamielennox11:25
mordredtobiash: yes - at the time we started looking, it was the only one of the three that implemented fair locking, which for us is important11:49
mordredtobiash: it's my understanding that etcd3 may now have support, as they increased the scope for what they wanted etcd to be able to do for v311:49
tobiashmordred: ah, good to know :)11:50
tobiashthx11:50
mordredsure thing!11:50
tobiashmordred: btw, I discovered a bug in my io sync optimization suggestion11:51
tobiashmordred: left a comment in https://review.openstack.org/#/c/448591/11:51
mordredooh - thanks11:55
openstackgerritJoshua Hesketh proposed openstack-infra/zuul feature/zuulv3: Put test id in zk chroot path  https://review.openstack.org/45492512:36
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Run unitests twice to see if this is possible  https://review.openstack.org/45493113:01
*** openstackgerrit has quit IRC13:33
clarkbmordred: you can also try eatmydata, but at least for tempest found it wasn't that much of a speedup (since its largely cpu bound)14:53
mordredclarkb: yah - I was mostly curious about impact on the clouds where we're our own noisy-neighbor - lke infra-cloud (and also there where the disks are slow too)15:02
mordredclarkb: I figure at places like osic we'll see no impact at all15:02
*** hashar is now known as hasharAway15:08
clarkbjlk: looks like http://logs.openstack.org/31/454931/2/check/gate-zuul-python27-ubuntu-xenial/12ce10f/console.html generally reproduces the issues we see locally. So good to know the test instances are doing something magical to get stuff to pass other than running the tests once then being deleted.16:32
mordredclarkb: WOOT16:40
clarkbmordred: I had initially expected to not merge ^ and use it as temporary test, however after seeing your +2 I wonder if it wouldn't be a bad idea to have it run twice until things are less flaky16:41
clarkb(of course that could make it much harder to make it less flaky)16:41
mordredclarkb: yah - the more I've thought about it the more I'm on the fence on the best way to do it ... I think double-running in the gate is a good idea,but maybe doing that directly in the py27 env in tox makes -epy27 locally a bit annoying16:43
mordredclarkb: maybe we should add an env that does it twice, or even just add a job that calls tox -epy27 twice in a row that we could start as non-voting while we get things stablized again or something16:43
jlkclarkb: I'm confused by your statement. YOu mean CI IS doing something magical, or is the magic just running in a "clean room" each time?16:49
clarkbsorry s/are/aren't/ basically the only magic is they run clean room each time16:50
clarkbthe job fails if you run tox -epy27 twice16:50
jlkawesome.16:52
jlkThat confirms my / our suspicions.16:52
jlkalthough doesn't jive with some things I'm seeing. MOre testing happening today for me.16:53
jlkbut first, coffee.16:53
clarkbthere are two big main effects of clean room each time. The first is no preexisting .testrepository which means test ordering is naive (alphabetical? something like that). The other is databases (mysql/zk) will potentially have data leftover and I have seen zk leak however I don't think those leaks should affect other tests as the random test chroot code seems to work properly16:55
*** jlk has quit IRC17:30
*** jlk has joined #zuul17:30
*** jlk has quit IRC17:31
*** jlk has joined #zuul17:31
jlkclarkb: what I was trying thus far was removing all of .testrepository, .tox, and all the contents of the zuul test root, re-running tools/test-setup.sh, and restarting zookeeper between each run. That may not be enough, might have to get more drastic.18:18
clarkbjlk: its possible that something in /tmp is leaking across jobs but I had good luck just removing .testrepository and the zk files18:19
jlkkk18:21
mordredjlk, clarkb: we're going to feel so good when we know what's wrong18:24
jlksuch knowledge usually leads to brown liquids.18:25
jlkwhich initially feel great.18:25
*** hasharAway is now known as hashar18:34
mordredmmm. liquid19:21
jlkclarkb: so even pruning all those files and restarting zookeeper, I can't get tip of feature/zuulv3 to pass locally. Rebooting to try again.19:48
clarkbjlk: are you clearing the zk data when you restart too?20:12
jlkso I did a stop of hte service, rm -rf of the tmpfs where zookeeper works in, and start of the service.20:12
jlkWhat size node does the gate use for these tests?20:47
clarkbjlk: 8 vcpu, 8GB ram, and ~80GB of disk (though in some cases much more disk)20:51
clarkbbut what 8vcpu means and how fast that disk is varies greatly across clouds20:51
jlkah20:52
jlkSo I'm testing with 4 vcpu, maybe I should do more20:52
jlkload average is between 4 and 820:52
clarkbjlk: testr will use the number of cpus available to determine how many processes to run20:54
clarkbso slightly above 4 is what I would expect for load average20:54
jlkresizing up my test VM21:17
jlkto a full 8 CPUs21:17
jlkMAGIC it passes.21:27
jlkwtf.21:27
*** hashar has quit IRC21:29
clarkbthat can affect test ordering so if it is an order issue that may be why it passed21:51
jlkAnybody around that understands in 2.5 how trigger requirements work? I'm trying to trace through the code how zuul decides that a trigger event should apply to a pipeline or not. Specifically thinking about trigger filters/requirements as opposed to pipeline requirements.21:53
*** openstackgerrit has joined #zuul22:12
openstackgerritMonty Taylor proposed openstack-infra/nodepool feature/zuulv3: Validate flavor specification in config  https://review.openstack.org/45187522:12
openstackgerritMonty Taylor proposed openstack-infra/nodepool feature/zuulv3: Add ability to select flavor by name or id  https://review.openstack.org/44978422:12
openstackgerritMonty Taylor proposed openstack-infra/nodepool feature/zuulv3: Cleanup from config syntax change  https://review.openstack.org/45186822:12
openstackgerritMonty Taylor proposed openstack-infra/nodepool feature/zuulv3: Add support for specifying key-name per label  https://review.openstack.org/45546422:12
*** yolanda has quit IRC22:18
*** yolanda has joined #zuul22:19
openstackgerritMonty Taylor proposed openstack-infra/nodepool master: Add support for specifying key-name per label  https://review.openstack.org/45546622:26
jlkI want to cry now.22:33
jlkthe tip of my patch branch completely passes tests on a fresh platform, when using 8 CPUs22:33
jlkIs there a way to tell testr to only use 4 cores instead of 8?22:33
jheskethMorning22:51
clarkbjlk: yes tox -e py27 -- --concurrency=423:05
mordredjlk: also - wow23:36

Generated by irclog2html.py 2.14.0 by Marius Gedminas - find it at mg.pov.lt!