Thursday, 2017-11-30

jeblairmordred: you may be able to provide input on https://review.openstack.org/505354 to help with that00:00
*** jasondot_ has quit IRC00:01
mordredjeblair: it's up in my browser for review ... I need to run - but will review first thing tomorrow00:03
openstackgerritMerged openstack-infra/zuul-jobs master: Honor constraints files in tox-siblings  https://review.openstack.org/52394900:11
openstackgerritMerged openstack-infra/zuul-jobs master: Combine tox-siblings and tox roles  https://review.openstack.org/52395000:11
*** jasondotstar has joined #zuul00:14
*** threestrands has quit IRC00:21
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Add inventory variables for checkouts  https://review.openstack.org/52197600:36
openstackgerritTristan Cacqueray proposed openstack-infra/zuul feature/zuulv3: executor: add log_stream_port and log_stream_file settings  https://review.openstack.org/52369700:42
*** neatherweb has joined #zuul01:45
*** threestrands has joined #zuul02:55
*** xinliang has quit IRC05:00
*** xinliang has joined #zuul05:01
*** baiyi has joined #zuul05:28
*** bhavik1 has joined #zuul05:29
baiyihello everyone,when I start zuul-executor in v3, the log show an error:  git.exc.GitCommandError: Cmd('git') failed due to: exit code(-13)05:31
tobiashbaiyi: do you have a full log?06:06
baiyi cmdline: git clone ssh://esadmin@192.168.210.32:29418/requirements /var/lib/zuul/executor-git/192.168.210.32/requirements06:08
baiyi  stdout: 'Cloning into '/var/lib/zuul/executor-git/192.168.210.32/requirements'...'06:08
baiyi2017-11-30 13:42:46,831 INFO zuul.Merger: Updating local repository gerrit/requirements06:08
baiyi2017-11-30 13:42:46,905 ERROR zuul.Merger: Unable to update gerrit/requirements06:08
baiyiTraceback (most recent call last):06:08
*** baiyi has quit IRC06:08
*** baiyi has joined #zuul06:08
baiyi cmdline: git clone ssh://esadmin@192.168.210.32:29418/requirements /var/lib/zuul/executor-git/192.168.210.32/requirements06:08
baiyi  stdout: 'Cloning into '/var/lib/zuul/executor-git/192.168.210.32/requirements'...'06:08
baiyi2017-11-30 13:42:46,831 INFO zuul.Merger: Updating local repository gerrit/requirements06:08
baiyi2017-11-30 13:42:46,905 ERROR zuul.Merger: Unable to update gerrit/requirements06:08
baiyiTraceback (most recent call last):06:08
baiyi  File "/usr/local/python3/lib/python3.6/site-packages/zuul/merger/merger.py", line 382, in updateRepo06:08
baiyi    repo.reset()06:08
baiyi  File "/usr/local/python3/lib/python3.6/site-packages/zuul/merger/merger.py", line 143, in reset06:08
baiyi    self.update()06:08
baiyi  File "/usr/local/python3/lib/python3.6/site-packages/zuul/merger/merger.py", line 284, in update06:08
baiyi    repo = self.createRepoObject()06:08
baiyi  File "/usr/local/python3/lib/python3.6/site-packages/zuul/merger/merger.py", line 136, in createRepoObject06:08
tobiashbaiyi: please use http://paste.openstack.org/ for posting logs06:08
tobiashbaiyi: does your executor have write permission in /var/lib/zuul/executor-git ?06:09
baiyihttp://paste.openstack.org/show/627791/06:10
baiyiyes06:11
tobiashbaiyi: try to execute the git command within the user context of the executor06:12
baiyiIt is normal to use zuul-executor -d06:12
tobiashmaybe it's a permission problem (no access to ssh key etc)06:13
baiyiBut I use "zuul-executor-d", and there's nothing wrong with that06:14
tobiashbaiyi: so you're starting it with your uid?06:15
tobiashdoes the git clone work if you do it by yourself?06:15
baiyiit works06:17
tobiashin the same path, using the same user?06:19
tobiashwithin the same environment?06:19
baiyiyes06:21
tobiashbaiyi: too bad it doesn't print stdout and stderr06:27
tobiashso currently I'm out of ideas06:28
tobiashyou could catch the GitCommanderror in File "/usr/local/python3/lib/python3.6/site-packages/zuul/merger/merger.py", line 128, in _git_clone and log stderr and stdout to further debug this06:29
tobiash(catch, log, reraise)06:29
openstackgerritTobias Henkel proposed openstack-infra/nodepool feature/zuulv3: Do pep8 housekeeping according to zuul rules  https://review.openstack.org/52294506:29
openstackgerritTobias Henkel proposed openstack-infra/nodepool feature/zuulv3: Use same flake8 config as in zuul  https://review.openstack.org/50971506:29
*** xinliang has quit IRC06:41
*** threestrands has quit IRC06:50
*** xinliang has joined #zuul06:54
*** baiyi has quit IRC06:57
*** baiyi has joined #zuul06:57
*** neatherweb has quit IRC07:29
openstackgerritRico Lin proposed openstack-infra/nodepool master: Fix getProviderManager in alien_list  https://review.openstack.org/52409208:08
openstackgerritKrzysztof Klimonda proposed openstack-infra/zuul feature/zuulv3: [WIP] Support autoholding nodes for specific changes/refs  https://review.openstack.org/51516908:21
*** baiyi1 has joined #zuul08:24
*** baiyi has quit IRC08:25
*** baiyi1 is now known as baiyi08:25
openstackgerritKrzysztof Klimonda proposed openstack-infra/zuul feature/zuulv3: [WIP] Support autoholding nodes for specific changes/refs  https://review.openstack.org/51516908:27
*** baiyi1 has joined #zuul08:32
*** baiyi has quit IRC08:33
*** baiyi1 is now known as baiyi08:33
*** bhavik1 has quit IRC08:57
*** flepied has quit IRC09:33
*** neatherweb has joined #zuul09:35
*** EmilienM has quit IRC09:39
*** EmilienM has joined #zuul09:40
*** EmilienM has quit IRC09:41
*** EmilienM has joined #zuul09:41
*** hashar has joined #zuul09:47
*** electrofelix has joined #zuul10:00
*** kklimonda has quit IRC10:09
*** kklimonda has joined #zuul10:15
*** jesusaur has quit IRC10:17
openstackgerritKrzysztof Klimonda proposed openstack-infra/zuul feature/zuulv3: Support autoholding nodes for specific changes/refs  https://review.openstack.org/51516910:20
openstackgerritKrzysztof Klimonda proposed openstack-infra/zuul feature/zuulv3: Support autoholding nodes for specific changes/refs  https://review.openstack.org/51516910:22
*** jesusaur has joined #zuul10:22
*** sshnaidm|off is now known as sshnaidm|rover10:48
*** jasondotstar has quit IRC11:02
*** jasondotstar has joined #zuul11:02
*** jasondotstar has quit IRC11:07
*** jasondotstar has joined #zuul11:08
*** jlk has quit IRC11:12
*** jlk has joined #zuul11:12
*** jlk has quit IRC11:13
*** jlk has joined #zuul11:13
*** jesusaur has quit IRC11:30
*** jesusaur has joined #zuul11:32
*** jkilpatr has quit IRC11:37
*** jkilpatr has joined #zuul11:56
*** jkilpatr has quit IRC12:14
*** jkilpatr has joined #zuul12:17
*** baiyi1 has joined #zuul12:26
*** baiyi has quit IRC12:27
*** baiyi1 is now known as baiyi12:28
openstackgerritKrzysztof Klimonda proposed openstack-infra/zuul feature/zuulv3: Support autoholding nodes for specific changes/refs  https://review.openstack.org/51516912:34
Shrewsmordred: re: 505354, i don't know how to reconcile that. the changes from your original fix (https://review.openstack.org/505626) do not seem to correspond to current code. i'm tempted to remove that squash from mine14:08
*** neatherweb has quit IRC14:15
openstackgerritMerged openstack-infra/zuul-jobs master: Move to dictionary list of projects zuul._projects (take 2)  https://review.openstack.org/51881514:17
*** dkranz has joined #zuul14:18
*** jasondotstar is now known as iamjasonc14:40
*** iamjasonc is now known as JasonC14:42
mordredShrews: yah - I say just remove that squash from yours14:43
mordredShrews: refactoring that whole file is coming up soon on my TDL14:43
*** JasonC is now known as JasonCL14:43
Shrewsk14:45
*** JasonCL has quit IRC14:48
*** JasonCL has joined #zuul14:48
*** JasonCL has quit IRC14:51
*** JasonCL has joined #zuul14:52
openstackgerritDavid Shrewsbury proposed openstack-infra/zuul feature/zuulv3: Changes for Ansible 2.4  https://review.openstack.org/50535414:56
*** JasonCL has quit IRC15:02
*** JasonCL has joined #zuul15:02
*** JasonCL has quit IRC15:03
*** JasonCL has joined #zuul15:03
pabelangermordred: friendly reminder https://review.openstack.org/521324/ :)15:26
kklimondahmm, can multiple nodepool launchers share same cloud?15:28
rcarrillocruzyeah, but caution with naming, you should name them differently15:29
openstackgerritDavid Moreau Simard proposed openstack-infra/zuul-jobs master: mirror-workspace-git-repos: Pass a dict, not a list containing a dict  https://review.openstack.org/52421115:30
pabelangerkklimonda: yes, I think we still have a setting to namespace them15:30
pabelangerlooks at docs15:30
openstackgerritMerged openstack-infra/zuul-jobs master: mirror-workspace-git-repos: Pass a dict, not a list containing a dict  https://review.openstack.org/52421115:31
pabelangerit was nodepool-id in v2, but I'm not sure nodepool-launcher needs that any more15:32
kklimondaI have two launchers registered with ZK, and one of them was printing "Quota exceeded" in the logs - I'm now thinking that's due to both of them competing for that last instance, and one of them receiving 403 from nova.15:35
rcarrillocruzyeah15:35
rcarrillocruzso that's the problem15:35
Shrewskklimonda: launchers should not share clouds15:35
kklimondaI've had a weird problem today, with nodepool leaving all nodes in state "Ready, Locked"15:35
rcarrillocruzi have two launchers sharing a cloud15:35
Shrewsrcarrillocruz: you should not15:35
rcarrillocruzbut had that thing you depicted15:35
Shrewssharing resources across launchers is a thing we have talked about for the future, but right now we can't support that15:36
kklimondawith a ton of errors like http://paste.openstack.org/show/627883/15:36
rcarrillocruzyeah, not saying you should in prod ( have in testing), but if you go down that route the only way to not have launchers fight each others resources is by having them with different name15:37
dmsimardBy the way, is there a reason why opening up a console that hasn't started yet will show ---- END OF STREAM ---- ? Could we either wait until it has started to make a link clickable or make the console wait for the stream to begin or something ?15:37
pabelangerYah, you need a single launcher for provider right now. You could share the cloud, but need to setup different providers in nodepool. I think that would work15:38
dmsimardRefreshing isn't a big deal but just wondering15:38
kklimondamhm, thanks - I'll shut down the other launcher, and try to actually track down and fix that ZK issue I've been having15:38
jeblairi'd like to draw your attention to this spec: https://review.openstack.org/524024  it's written generally for new projects that we might start hosting in openstack, but i'd also like to pilot it with zuul.  i'd particularly like to get these things in place before we relase zuul 3.015:38
jeblairdmsimard: if we had the executor drop a line in job-output.txt before sending the gearman status update, it should have the effect you want15:39
kklimondamy nodepool is getting "RuntimeError: ('xids do not match, expected %r received %r', 913440, 913439)" and then gets stuck in "ZooKeeper suspended. Waiting" loop15:39
jeblairdmsimard: just a log line like "starting" or something15:39
Shrewsrcarrillocruz: pabelanger: you can fudge it with different names, but that's really not how it's supposed to work and calculations of quota things may be weird15:39
dmsimardjeblair: Ah, so just something like "Console starting..." yeah15:39
dmsimardjeblair: I'll figure out where it is and send a patch15:39
jeblairdmsimard: thx++15:39
rcarrillocruzyup15:39
pabelangerShrews: Yah, that's what we did for OSIC under v2. Was single cloud, but different flavors. Which, isn't something we'd do with zuulv3 now15:40
kklimondajeblair: if you have time to review https://review.openstack.org/#/c/515169/ I'd appreciate it :)15:40
jeblairkklimonda: thanks -- i should get to it soon -- i'm working my way back through after returning from vacation/holiday :)15:41
kklimondahaha, I'll leave you to that then15:42
jeblairthere are a bunch of changes in gate that have been stuck as 'queued' for 14 hours15:45
jeblairit seems to be due to: http://paste.openstack.org/show/627887/15:45
pabelangerI've seen that in logs before15:47
jeblairthat node doesn't seem to be in zk (which makes sense from that message)15:48
Shrewsjeblair: well, if it was never locked (or lost its lock somehow), it could have been re-used or recycled15:48
Shrewswould need to check corresponding nodepool logs15:49
jeblairhttps://etherpad.openstack.org/p/o9Tv7MVq0x15:49
jeblairi'm accumulating info there15:49
pabelangerI do see some launch errors in grafana.o.o15:49
pabelangerlooking at logs now15:49
pabelangerwe also had a new shade release yesterday, something to also check15:50
jeblairpabelanger: this is more than likely related to the zuul restart15:50
pabelangerok15:50
Shrewsso it was set to USED by zuul, but lost its lock (due to restart?)15:52
Shrewshrm, set to in use twice?15:53
openstackgerritDavid Moreau Simard proposed openstack-infra/zuul feature/zuulv3: Print a message when we start the Zuul console  https://review.openstack.org/52422515:53
Shrews3 hours apart15:53
rcarrillocruzhey, so i got it working 3rd party CI with zuulv3 + github https://github.com/rcarrillocruz-org/ansible-fork/pull/13 (note the ricky2-zuul, that's the 3rd party CI doing a noop). However to make it work i had to set the untrusted project I want to run test against to exclude job and project, otherwise the reconfiguration of the untrusted project (loading the .zuul.yaml) was referencing jobs and projects not15:54
jeblairShrews: well, the last log block is being continuously repeated15:54
rcarrillocruzpresent on the 3rd party zuul15:54
rcarrillocruzhttps://github.com/rcarrillocruz-org/ansible-fork/pull/1315:54
rcarrillocruzerm15:54
rcarrillocruzhttp://paste.openstack.org/show/627889/15:54
rcarrillocruzis that ok approach, or I'm misunderstanding stuff and should do it some other way15:54
pabelangerrcarrillocruz: not following your question15:55
mordredrcarrillocruz: the .zuul.yaml in ansible-fork had jobs defined that referenced jobs you didn't have?15:55
rcarrillocruzfrom origin zuul15:56
jeblairShrews: i'm working on getting the restart times from the log now15:56
pabelangernm, just re-read it15:56
rcarrillocruzif i defined them on the third party zuul15:56
rcarrillocruzthen the funny thing is that it appears to try to run them15:56
rcarrillocruzbut what i just want to run the jobs defined in the 3rd party15:56
rcarrillocruzthat's the only way i could make it work15:56
mordredrcarrillocruz: yah - so - I think excluding project is important and definitley correct15:57
mordredrcarrillocruz: (when we add ansible/ansible to openstack zuul we'll definitely exclude project)15:57
mordredrcarrillocruz: since otherwise exactly whatyoumentoin happens15:57
rcarrillocruzk15:57
mordredrcarrillocruz: for jobs -it sorts of depends on whether or not you want to re-use any of hte jobs defined in the repo as part of your third party testing- but if you do, you might have to include additional repos to get the job depends15:58
mordredrcarrillocruz: but - yah - excluding job and project is a totally legit thing- andthe safest place to start15:58
rcarrillocruzk thx15:58
jeblairShrews, pabelanger: hrm, the restart was 3 hours before this node was created.15:59
*** JasonCL has quit IRC16:00
*** JasonCL has joined #zuul16:02
Shrewsjeblair: any zk disconnects between 03:16:38 and 03:21:44 ?16:03
Shrewsor session errors16:03
jeblairhrm, i don't see any in zuul's log16:05
Shrewsweird16:06
jeblairShrews: any idea why nodepool would unlock/lock/unlock from 03:13 -- 03:16?16:30
jeblairShrews: oh there are 2 requests16:31
jeblairpresumably something happened to request 3740 and then it got reattached to 377816:31
Shrewsjeblair: i haven't looked at the logs (except what you pasted). which launcher?16:31
Shrewsseems like you figured it out though16:31
jeblairShrews: nl01  (and i was just reading that from the ones in the etherpad)16:31
mordredjeblair: (I'm still operating under the assumption that it's unlikely this is related to the shade release so I'm hammering on otherthings, but if it starts to look like shade is involved, please ping)16:32
jeblairmordred: ++16:32
Shrewsah16:32
Shrewsjeblair: curious that it would be assigned to two requests. did zuul return it unused?16:33
jeblairtrying to dig into the first request now16:33
Shrewsi think nodepool would log if it cleared the first allocation16:34
* Shrews looks16:34
jeblairShrews: oh -- how do we tell if that first request was min-ready?16:35
jeblair(cause i don't see it in zuul logs, so i'm starting to suspect it as)16:35
Shrewsjeblair: it was16:36
Shrews2017-11-30 03:13:27,302 INFO nodepool.PoolWorker.tripleo-test-cloud-rh1-main: Assigning node request <NodeRequest {'reuse': False, 'node_types': ['ubuntu-xenial'], 'id': '100-0001293740', 'nodes': [], 'state': 'requested', 'state_time': 1512011599.1659088, 'requestor': 'NodePool:min-ready', 'declined_by': [], 'stat': ZnodeStat(czxid=742752127, mzxid=742752127, ctime=1512011599165, mtime=1512011599165,16:36
Shrewsversion=0, cversion=0, aversion=0, ephemeralOwner=98788114264752898, dataLength=172, numChildren=0, pzxid=742752127)}>16:36
jeblairShrews: ah there it is, thanks :)16:36
Shrewsrequestor16:36
Shrewsso that makes sense16:37
jeblairokay, so that's a red herring.  whatever interesting happened it was with the second request it was assigned to16:37
Shrewsnothing unusual about that 2nd request in the logs16:38
Shrewsnodepool logs, that is16:38
jeblairweird, there's a 2.5 hour gap between returinng the node and trying to use it again.16:40
pabelangeris that the timeout for job?16:42
jeblairpabelanger: oh interesting idea16:44
rcarrillocruzhmm16:45
rcarrillocruzinteresting16:45
rcarrillocruzso16:45
rcarrillocruzset up 3rd party CI on GitHub16:46
rcarrillocruzget the two zuul, the one managing the repo I push the PR and the 3rd party one, reporting16:46
rcarrillocruznow i uninstall the 3rd party CI Zuul App from the org16:46
rcarrillocruzi do a recheck16:46
rcarrillocruzand i get again the two zuuls on the status16:47
rcarrillocruzdespite obvciously the 3rd paarty Zuul not receiving events, as I uninstall it16:47
rcarrillocruzlooks as if the PR somehow caches it has two CIs16:47
rcarrillocruztristanC, jlk: any hint ^ ?16:47
jeblairrcarrillocruz: well, the report sets a status on the pr; i imagine that when you remove the app it doesn't unset the previously set status fields16:48
rcarrillocruzthe zuul bot does not report tho, BUT the status thingy does show teh two zuuls16:48
rcarrillocruzah, so it's a static thing16:48
rcarrillocruzk, that explains, thx16:48
rcarrillocruzprepping a demo for tomorrow to my peers16:48
*** kmalloc has quit IRC17:11
*** JasonCL has quit IRC17:17
*** JasonCL has joined #zuul17:18
jeblairokay, i think i'm narrowing this down to an interaction between pipeline window sizes and restarts17:28
jeblairer reconfigurations17:28
jeblairthere were 2 reconfigurations between 03:00 and 06:00.  i think after each reconfiguration, the window size is reset.  so after the first reconfig, the tripleo window size went from 32 to 20.  that put the change we're looking at out of the window during the second reconfiguration.  i think that caused build results not to be ported over, though nodesets were.  so once it entered the window again, it tried to restart the build with an old nodeset.17:30
jeblairi think that explains why there is a set of changes at the top of the tripleo queue all like this -- they were the ones that fit between positions #20 - 32 in the queue.17:31
jeblairie, the ones that got cut out of the window when it shrunk during the reconfig.17:31
clarkbhuh17:37
*** sshnaidm|rover is now known as sshnaidm|off18:02
*** hashar is now known as hasharDinner18:06
*** jkilpatr has quit IRC18:23
*** myoung|ruck has joined #zuul18:31
*** jkilpatr has joined #zuul18:39
*** openstackgerrit has quit IRC18:48
*** openstackgerrit has joined #zuul19:30
openstackgerritMerged openstack-infra/zuul-jobs master: Remove tox-siblings role stub  https://review.openstack.org/52399619:30
pabelangermordred: ^does that mean we can now push on removin tox_install.sh?19:33
pabelangerokay, I think updated-pti is the topic you are using19:34
*** electrofelix has quit IRC19:38
*** jkilpatr has quit IRC19:39
mordredpabelanger: yes - for the projects that are not using tox_install.sh to install additional software, it should be safe to make patches like https://review.openstack.org/#/c/508061/19:47
pabelangermordred: coolio! I'll start looking after some food19:48
*** kmalloc has joined #zuul20:07
jeblairyay i think i have a reproducing test case20:09
*** JasonCL has quit IRC20:12
mordredjeblair: \o/20:13
*** JasonCL has joined #zuul20:21
*** hashar has joined #zuul21:05
*** hasharDinner has quit IRC21:08
*** jkilpatr has joined #zuul21:09
*** threestrands has joined #zuul21:12
*** threestrands has quit IRC21:12
*** threestrands has joined #zuul21:12
*** dkranz has quit IRC21:30
*** JasonCL has quit IRC21:39
*** hashar has quit IRC21:39
*** JasonCL has joined #zuul22:22
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Remove nodesets from builds canceled during reconfiguration  https://review.openstack.org/52440922:26
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Don't shrink windows on reconfiguration  https://review.openstack.org/52441022:26
jeblairclarkb: if you could take a look at that, i'd appreciate it22:26
clarkblooking22:27
clarkbthat is a not small set of fixing22:28
clarkb(thankfully mostly tests \o/22:28
jeblairyeah, the actual changes are pretty small tweaks.  mostly tests.22:29
mordredjeblair: wow. those are fun22:34
clarkbobeservation, ExecutorClient.execute() is really long22:35
jeblairit's generally getting shorter :)22:36
clarkbindeed, one line at a time :)22:36
*** dkranz has joined #zuul22:39
clarkbjeblair: reading this can we just stop keeping track of the nodeset in the buildset entirely?22:40
clarkbseems more natural and less redundant that each build would track its own buildset?22:40
clarkb(thats a bigger change after digging through model.py so maybe something to do in the future)22:41
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Don't set job var override_checkout if null  https://review.openstack.org/52441422:42
jeblairthat ^ reduces it by one more line, thanks to comment trickery :)22:42
jeblairclarkb: the buildset holds the nodeset before the build is started, so it serves a pretty important role still22:43
clarkbjeblair: but couldn't you have a buildset(build(state: notstarted, nodeset: foo)) and track it that way? it just feels weird double accounting this22:44
clarkbbut yes not easy to change22:44
jeblairclarkb: that makes sense, but currently the build is not created until it's launched, so yeah, that's non-trivial22:44
jeblairanother option would be to remove the nodeset from the buildset as soon as the build is created, and have the routines that look for a nodeset check both places22:45
jeblairthat's slightly less invasive, but not quite as much of an increase in clarity as your suggestion22:45
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Remove nodesets from builds canceled during reconfiguration  https://review.openstack.org/52440922:49
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Don't shrink windows on reconfiguration  https://review.openstack.org/52441022:49
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Don't set job var override_checkout if null  https://review.openstack.org/52441422:49
jeblairi just added ordered=false to the asserts in the tests since the jobs can complete out of order (that was the test failure reported on the change)22:50
clarkbjeblair: https://review.openstack.org/#/c/524409/1/tests/unit/test_scheduler.py line 3886, what causes that reconfigure to cancel the other jobs? its just applying the old config again isn't it?22:50
clarkbs/old/previous/22:51
clarkboh do we remove non running builds on reconfiguration regardless of config?22:52
clarkbwell that can't be because then the first reconfigure would cancel them22:53
jeblairclarkb: ah i think i see it after the first reconfiguration, the item is still marked active even though it's outside the window, but as soon as the queue processor runs on it, it flips that bit to inactive.  so then the next reconfiguration, the item is considered inactive, and that cancels the job.22:58
jeblairiow, the key is that we shrink the window during a reconfiguration, perform a pass through the queue processor, then reconfigure a second time.22:58
clarkbwhat does the pass through the queue processor? does the second reconfiguration trigger that first?22:59
jeblairclarkb: after completing the first reconfiguration, we do pass through (that's what would cause new jobs to run)23:00
jeblair(i mean, we run the processor after every reconfiguration for that reason)23:00
clarkbgotcha23:01
clarkbstack lgtm23:04
clarkbthe keeping of old windows is slightly brain trippy, means you can't force a smaller window if things are going good, but also means we don't artificially inhibit queue growth if things are going good23:05
clarkb(and the intent is to let the current state of goodness determien values more so than human fiddling so that wins out)23:05
clarkblooks like tests are still unhappy though23:06
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Don't shrink windows on reconfiguration  https://review.openstack.org/52441023:32
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Don't set job var override_checkout if null  https://review.openstack.org/52441423:32
jeblairthat should fix the tests -- we were setting windows on queues which didn't have them.23:32
*** bramwelt has joined #zuul23:35

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!