Thursday, 2019-02-21

*** jamesmcarthur has quit IRC00:01
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: WIP: use-buildset-registry: support running before docker installed  https://review.openstack.org/63818000:47
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Split docker mirror config into its own role  https://review.openstack.org/63819500:47
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: WIP: use-buildset-registry: support running before docker installed  https://review.openstack.org/63818001:18
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Split docker mirror config into its own role  https://review.openstack.org/63819501:18
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: use-buildset-registry: configure as a pull-through proxy  https://review.openstack.org/63831201:18
*** logan- has quit IRC01:22
*** logan- has joined #zuul01:23
*** logan- has quit IRC01:31
*** logan- has joined #zuul01:31
*** ruffian_sheep has joined #zuul01:39
ruffian_sheepDoes anyone know how to deal it?  git.exc.GitCommandError: Cmd('git') failed due to: exit code(-13)01:43
ruffian_sheepI can do the cmd by myself.But I cannot be used by the service. http://paste.openstack.org/show/745241/01:43
*** sdake has quit IRC01:44
*** sdake has joined #zuul01:46
*** jamesmcarthur has joined #zuul02:10
SpamapSruffian_sheep: that looks like maybe zuul isn't using the right SSH key02:11
SpamapShm02:23
SpamapSzuul.change doesn't get set in post pipelines from what I can tell (at least, on my github)02:23
SpamapSwhich a) disagrees with the docs (zuul.change is in the generic section of zuul attributes), and b) makes using the promote pipeline more complicated.02:24
SpamapSI have a feeling we could make an attempt at setting change, since the github UI figures it out.02:35
*** jamesmcarthur has quit IRC02:37
*** jamesmcarthur has joined #zuul02:42
*** kratec has left #zuul02:45
*** sdake has quit IRC02:48
*** sdake has joined #zuul02:48
*** sdake has quit IRC02:57
*** sdake has joined #zuul02:58
*** jamesmcarthur has quit IRC03:08
*** pwhalen has quit IRC03:14
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: gerrit: add support for report only connection  https://review.openstack.org/56821603:18
*** pwhalen has joined #zuul03:19
*** saneax has joined #zuul03:25
*** jamesmcarthur has joined #zuul03:32
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add buildsets page  https://review.openstack.org/63004103:32
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add /{tenant}/buildset/{uuid} route  https://review.openstack.org/63007803:32
*** spsurya has joined #zuul03:34
*** jamesmcarthur has quit IRC04:00
*** jamesmcarthur has joined #zuul04:01
*** jamesmcarthur has quit IRC04:05
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: gerrit: add support for report only connection  https://review.openstack.org/56821604:08
*** saneax has quit IRC04:11
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add /{tenant}/buildset/{uuid} route  https://review.openstack.org/63007804:19
*** nilashishc has joined #zuul04:21
*** nilashishc has quit IRC04:25
*** ianychoi has quit IRC04:31
*** sdake has quit IRC04:50
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add /{tenant}/buildset/{uuid} route  https://review.openstack.org/63007804:54
*** sdake has joined #zuul04:54
*** sdake has quit IRC05:14
*** sdake has joined #zuul05:16
*** saneax has joined #zuul05:45
ruffian_sheepBut I can do the cmd to clone this project?Isn't mean I set the right ssh key? SpamapS05:46
*** fdegir has quit IRC05:48
*** sdake has quit IRC05:48
*** fdegir has joined #zuul05:48
*** sdake has joined #zuul05:50
*** kmrchdn is now known as chandankumar06:01
*** sdake has quit IRC06:03
*** [GNU] has quit IRC06:03
*** sdake has joined #zuul06:04
openstackgerritTristan Cacqueray proposed openstack-infra/nodepool master: Add python-path option to label  https://review.openstack.org/63733806:22
*** spsurya has quit IRC06:25
*** nilashishc has joined #zuul06:27
*** sdake has quit IRC06:37
*** quiquell|off is now known as quiquell06:50
*** [GNU] has joined #zuul07:02
*** spsurya has joined #zuul07:17
*** bhavikdbavishi has joined #zuul07:19
*** pcaruana has joined #zuul07:19
*** quiquell is now known as quiquiell|brb07:30
*** rlandy|bbl is now known as rlandy07:44
*** bhavikdbavishi has quit IRC07:56
*** quiquiell|brb is now known as quiquell07:59
*** themroc has joined #zuul08:09
*** jpena|off is now known as jpena08:29
*** saneax has quit IRC08:37
*** hashar has joined #zuul08:37
*** electrofelix has joined #zuul09:08
*** gtema has joined #zuul09:26
*** panda|off is now known as panda09:54
*** iurygregory has quit IRC10:01
*** spsurya has quit IRC10:22
*** bhavikdbavishi has joined #zuul11:10
*** ruffian_sheep has quit IRC11:21
*** nilashishc has quit IRC11:43
*** jpena is now known as jpena|lunch11:56
*** bhavikdbavishi has quit IRC11:59
*** bhavikdbavishi has joined #zuul11:59
*** sdake has joined #zuul12:01
*** quiquell is now known as quiquell|lunch12:05
*** nilashishc has joined #zuul12:11
*** bhavikdbavishi1 has joined #zuul12:20
*** bhavikdbavishi has quit IRC12:20
*** bhavikdbavishi1 is now known as bhavikdbavishi12:20
*** saneax has joined #zuul12:40
*** jesusaur has quit IRC12:50
*** jesusaur has joined #zuul12:56
openstackgerritSorin Sbarnea proposed openstack-infra/zuul-jobs master: Assure iptables is installed inside multi-node-firewall role  https://review.openstack.org/63841412:58
*** gtema has quit IRC12:59
openstackgerritSorin Sbarnea proposed openstack-infra/zuul-jobs master: Assure iptables is installed inside multi-node-firewall role  https://review.openstack.org/63841413:08
*** quiquell|lunch is now known as quiquell13:10
*** jpena|lunch is now known as jpena13:15
*** sdake has quit IRC13:23
*** sdake has joined #zuul13:25
openstackgerritMerged openstack-infra/zuul-jobs master: use-buildset-registry: configure as a pull-through proxy  https://review.openstack.org/63831213:28
*** bhavikdbavishi has quit IRC13:32
*** gtema has joined #zuul13:35
*** sdake has quit IRC13:36
*** sdake has joined #zuul13:37
*** rlandy has joined #zuul13:40
*** jamesmcarthur has joined #zuul13:43
*** nilashishc has quit IRC13:43
*** jamesmcarthur has quit IRC13:58
*** jamesmcarthur has joined #zuul13:58
*** jamesmcarthur has quit IRC14:03
*** sdake has quit IRC14:14
*** panda is now known as panda|ruck14:16
*** gtema has quit IRC14:18
*** jamesmcarthur has joined #zuul14:29
*** sdake has joined #zuul14:31
*** jamesmcarthur has quit IRC14:34
*** nilashishc has joined #zuul14:39
*** nilashishc has quit IRC14:42
*** sdake has quit IRC14:51
pabelangerShrews: https://storyboard.openstack.org/#!/story/2005064 is a fun bug in nodepool we found last week, I'm happy to work on it, but would need some design help14:58
pabelangerfor history, http://eavesdrop.openstack.org/irclogs/%23zuul/%23zuul.2019-02-12.log.html#t2019-02-12T15:23:56 is the discussion last week with clarkb14:58
*** sdake has joined #zuul15:01
Shrewspabelanger: interesting. i haven't read the irc log yet, but did you identify why the instance creation failed?15:04
Shrewsi don't see that in any of the links15:05
pabelangerShrews: cloud was in a bad state, along with lack of resources15:05
pabelangerquota for provider was superhigh, even though there was no compute capacity15:05
Shrewspabelanger: what do you mean by "bad state"?15:06
Shrewsor is the quota thing the reason15:06
Shrewsor a symptom15:06
pabelangerShrews: like the controllers for the cloud were in the process of dying, and needed to be rebooted15:07
pabelangerthe cloud in general isn't very stable15:07
corvusSpamapS: for 'promote' you need a change-based pipeline (not like 'post' which is sha-based).  my guess for github is you want to trigger on a 'pull_request' event, action 'closed', and require 'merged'.15:07
Shrewspabelanger: what confuses me is that an error is reported for the node creation, yet instances are still being created (or so it would seem). how would you go about solving that? the cloud seems to be lying to us15:08
*** saneax has quit IRC15:08
pabelangerShrews: however, once it recovered, there was no capacity for new nodes, because all of the ERROR'd instances nodepool created. And because of the reuse of nodepool_node_id, the clean up handler couldn't see those instnaces to try to delete15:08
pabelangerShrews: Yah, I am not too sure. The other key info, is retries on that provider is crazy high (99999999), as not to have nodepool report NODE_FAILURE results back to gerrit15:09
*** jamesmcarthur has joined #zuul15:10
pabelangerone idea clarkb and I discussed, was to update the nodepool_node_id on the ERROR'd node, when the exception is raise, but unsure if that is the right thing to do15:11
jkttristanC: I have an option for the runc driver to optionally use overlayfs for container's rootfs. Should I squash it to that change under review, or submit as a separate one?15:12
Shrewspabelanger: how can you update nodepool_node_id on a node that is reported to not exist when the exception is raised?15:12
Shrewspabelanger: https://softwarefactory-project.io/paste/show/1424/ shows that we don't see the node at that point15:13
Shrewspabelanger: if the root of the problem is that what the cloud reports to us does not match reality (i feel like that's what your bug report might be saying?), then i don't know how to solve that up front.15:13
Shrewspabelanger: let me read through the irc scrollback to see if i'm missing something15:15
pabelangerShrews: right, not sure why it doesn't exist, I suspect openstack side issue.  However, we do eventually see it in our cleanup handler, but fail to try to delete it because of: http://git.zuul-ci.org/cgit/nodepool/tree/nodepool/driver/openstack/provider.py#n51415:15
pabelangerif we didn't reuse the nodepool_node_id on failure, I don't think we'd have this issue (but unsure what that means for code base)15:15
pabelangersure15:15
*** jamesmcarthur has quit IRC15:17
*** swest has quit IRC15:17
Shrewspabelanger: can you explain how that code causes us to fail to delete it? is it because that zknode exists? if so, what is its STATE value?15:19
Shrewspabelanger: also, have you been able to create a test case that shows this?15:20
Shrewsi think we should start with that, if possible15:21
pabelangerShrews: yah, it still existed. The state was either READY or inuse, because a node eventually was build with the nodepool_node_id:  see like 86 at https://softwarefactory-project.io/paste/show/1428/15:22
pabelangerI don't seem to have the original zk output15:22
pabelangerI think what happens in the case of ^ paste, is once server 3cb0fc64-0f00-48b5-afeb-2fad08d941cf is finished running a job ( in this case usually 3 hours), the data will eventually be deleted in zk, then our clean up handler will start to see the leaked error nodes15:23
pabelangerbut, never confirmed that because they manually deleted the error'd nodes using openstack client15:24
Shrewspabelanger: ok, there are lots of question marks here... if you can create a test case that fails and demonstrates the issue, it would be much easier to dissect15:27
pabelangerack, I can work on that15:27
Shrewsbecause i'm still confused as to what state the zookeeper database and clouds were actually in15:28
pabelangerjust thinking, I'm not sure how to mock the openstack side, because we need to actually validate there will be 2 nodes created (1 error, 1 ready) but nodepool thinks there is only 1 zookeeper entry15:30
pabelangerdue to same nodepool_node_id15:30
Shrewspabelanger: do you think what's happening is that we 1) get an exception on launch  2) create a zknode for the errored instance  3) cleanup thread tries to delete errored instance but it does not yet exist  4) we delete the zknode  5) errored instance now appears in the cloud15:36
Shrewsi'm still trying to understand what's causing the errored instances to not get deleted15:38
pabelangerShrews: yes, that seems to be the flow.15:39
pabelangerI am unsure why that is15:39
Shrewsso why isn't http://git.zuul-ci.org/cgit/nodepool/tree/nodepool/driver/openstack/provider.py#n514 deleting it?15:39
pabelangerbut step 5) errored node appears, with original nodepool_node_id, which is likely been reused for the next node launch attempt15:40
Shrewsah, that's new info15:40
pabelangerShrews: sorry, that is what I was trying to show in: https://softwarefactory-project.io/paste/show/1430/ you can see 0001615360 exists twice for 2 different nodes15:41
*** sdake has quit IRC15:48
*** jamesmcarthur has joined #zuul15:48
Shrewsso nodepool_node_id is not matching the actual zk node sequence id, caused by the new code here: http://git.zuul-ci.org/cgit/nodepool/tree/nodepool/driver/openstack/handler.py#n26015:49
*** sdake has joined #zuul15:50
Shrewshowever, that *should* eventually resolve itself15:51
pabelangerShrews: right, the new nodepool_node_id is delete, step 3 above leaking original15:52
pabelangeryah, I agree it should resolve, however the time it takes to resolve could end up being a large amount of time15:52
Shrewsyeah, until the READY instance gets used and deleted, the errored instances will remain15:53
pabelangerwe retries set to 999999 we end up created a lot of ERROR'd nodes15:53
*** jamesmcarthur has quit IRC15:53
pabelangeryes15:53
Shrewsthat's an unusually high number15:53
pabelangeryup, single cloud provider, and don't want nodepool to report back NODE_FAILURES15:53
pabelangerso, basically setup to keep trying forever until a nodepool node comes online15:53
pabelangerkinda like v2 nodepool15:54
Shrewsok, wow. edge case i don't believe we've considered15:54
pabelangeryah, so not sure the best path there to have nodepool delete the error'd nodes, because it is resulting in people manually cleaning them up15:56
Shrewspabelanger: ok, now that i better grasp what you're seeing, i can now say with 100% certainty that i don't have an immediate answer.  :)15:56
Shrewswill require some thought15:56
pabelangerYay!15:56
pabelangercool, thanks15:56
*** jamesmcarthur has joined #zuul15:58
*** jamesmcarthur has quit IRC15:58
*** jamesmcarthur has joined #zuul15:58
Shrewsright now, nodepool assumes temporary excess usage for a short period. but your use case flips that on its head, where we need NO excess usage for a possibly long period of time16:01
Shrewsnot certain there is a quick fix for that, but... maybe?16:02
*** themroc has quit IRC16:03
*** rfolco is now known as rfolco|rover16:07
clarkbShrews: pabelanger my short summary of understading that bug is we don't delete it because the node request eventually succeeds. Once the successful node is used and makred for deletion the leaked nodes should be deletable as well16:09
clarkbso the issue is time to delete more than never gets deleted16:09
*** quiquell is now known as quiquell|off16:13
*** jamesmcarthur has quit IRC16:13
Shrewsclarkb: that's pretty much the clarifying comment i just left on the storyboard issue16:14
Shrewsclarkb: pabelanger: does my comment here seem to sum it up properly? https://storyboard.openstack.org/#!/story/200506416:15
clarkbShrews: yup that looks correct to me16:16
pabelangerShrews: yes, thanks16:16
clarkbI think the other thing that complicates this is we refer to the nova metadata to get the request id16:17
clarkbbut that ends up being stale. I wonder if we can replace that with a boot key that a node request updates for each node booted16:17
Shrewsyeah, if we could keep those values in sync, that would be the solution16:23
clarkbShrews: or maybe do node-request-id-iteration-number16:25
clarkbin nova metadata16:25
clarkbthen the node request can record the iteration that was successful. This has the nice effect of being human readable16:25
*** jamesmcarthur has joined #zuul16:27
*** hashar has quit IRC16:39
clarkbfollowing up on the github commit to PR cache in Zuul. It does seem to help the openstack deployment but I'm still seeing the occasional run up to a backlog of a couple hundred then back down to zero16:45
clarkbhttps://github.com/ansible/ansible/pull/52682 is a PR merging that caused it to run up todate.16:45
clarkbI haven't confirmed this but I think the merge_commit_sha is relative to the current branch head of the target branch16:45
clarkbthis means if say ansible were to merge two commits in succession the merge_commit_sha for the second would move and we wouldn't necessarily have that cached forcing us to do the full scan16:45
clarkbif this is what is happening I don't think there is much that zuul can do unfortunately. And really points to the need of better query api tooling in github (which they acknowledge so yay)16:46
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: Proposed spec: tenant-scoped admin web API  https://review.openstack.org/56232117:06
*** pcaruana has quit IRC17:07
*** nilashishc has joined #zuul17:20
*** jamesmcarthur_ has joined #zuul17:24
*** jamesmcarthur has quit IRC17:27
*** nilashishc has quit IRC17:31
*** nilashishc has joined #zuul17:31
*** nilashishc has quit IRC17:40
*** panda|ruck is now known as panda|ruck|off17:48
openstackgerritJan Kundrát proposed openstack-infra/nodepool master: runc: Allow overlayfs mounts for container's rootfs  https://review.openstack.org/63846917:57
*** sdake has quit IRC18:05
*** sdake has joined #zuul18:05
*** jpena is now known as jpena|off18:29
jlkhave y'all brought up TravisCI news in here yet?18:30
pabelangerhaven't heard anything recently18:32
jlkhttps://www.reddit.com/r/devops/comments/at3oyq/it_looks_like_ibera_is_gutting_travis_ci_just_a/18:32
pabelangerI know they were acquired a few weeks ago18:32
jlkA bunch of sudden layoffs, in some cases without the person's manager knowing about it18:32
Shrewswow18:34
*** electrofelix has quit IRC18:37
pabelangereep18:37
mordredyuck18:55
jlkThere's going to be some good people on the job market who understaned CI systems.18:56
mordredjlk: not to be morbid, but it's a good reminder that the issue is never the intention of the original well-intentioned startup ... it's the potential behavior of whoever buys them that's the issue18:56
mordredjlk: ++18:56
jlkmordred: that's some uh... relevant musings given uh... recent acquisitions.18:57
mordredjlk: not, of course, that you have any experience being part of a startup that was bought by people who then proceeded to have no clue or anything18:57
jlkyeah...18:57
jlkI am currently rather pleased with my giant corp overlord.18:57
* Shrews gets a cold chill for some reason18:57
mordredjlk: your new overlord seems to have become one of the most clued in corp overlords, which really just makes me doubt reality18:58
jlkthis timeline is so weird18:58
*** hashar has joined #zuul18:59
mordred:)18:59
*** manjeets has quit IRC19:14
*** manjeets has joined #zuul19:14
*** hashar has quit IRC19:14
fungithat timeline where the empire commits to a lengthy effort to rebuild coruscant, and then paints the death star a cheerful shade of yellow and starts selling flowers out of the docking bay19:56
*** jamesmcarthur_ has quit IRC19:56
*** jamesmcarthur has joined #zuul19:57
fungino, wait, i meant rebuild alderaan (just gotta find all the pieces first)19:58
*** jamesmcarthur has quit IRC20:01
*** jamesmcarthur has joined #zuul20:18
*** jamesmcarthur has quit IRC20:40
*** jamesmcarthur has joined #zuul20:40
*** jamesmcarthur has quit IRC20:45
*** jamesmcarthur has joined #zuul20:53
daniel2So nodepool has a bunch of stale builds: http://paste.openstack.org/show/745652/ I can't seem to clean them out, any suggestions?20:54
clarkbdaniel2: this is still older nodepool not v3 right?20:55
daniel2clarkb: yes, 0.5.020:55
daniel2We were running 0.3.0 originally but I was able to make a case to at least upgrade to 0.5.020:55
clarkbdaniel2: I think that happens if the builder and main server which talks to the mysql db get out of sync (this was a big reason for movign to zookeeper for this data). In that case I think you just have to drop the database entries20:55
daniel2the mysql database tables are empty.  I forgot about zookeeper20:56
daniel2It must be storing in that.20:56
daniel2clarkb: thanks, cleared the node in zookeeper which fixed that.21:00
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: Proposed spec: tenant-scoped admin web API  https://review.openstack.org/56232121:26
tristanCjkt: it sounds like overlayfs could use a separate review, but maybe not, i don't know the implementation detail21:27
*** hashar has joined #zuul21:32
*** badboy has quit IRC21:41
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: run-buildset-registry: run a dual registry  https://review.openstack.org/63851421:54
*** jamesmcarthur has quit IRC21:57
*** jamesmcarthur has joined #zuul22:05
*** jamesmcarthur has quit IRC22:07
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: use-buildset-registry: support running before docker installed  https://review.openstack.org/63818022:12
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Split docker mirror config into its own role  https://review.openstack.org/63819522:12
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Use buildset registry push endpoint  https://review.openstack.org/63852022:12
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Use buildset registry push endpoint  https://review.openstack.org/63852022:15
clarkblooking at the github event processing we definitely seem to take a minute an a half to process consecutive ansible merged PR status events22:22
clarkbI would expect the first to populate the cache then the second and third and so on to be looked up that way22:22
clarkbbeginning to suspect either the shas are not up to date as expected (like when does github update merge_commit_sha ?) or we've got a bug22:23
clarkbjlk: ^ do you know how eventually consistent the merge_commit_sha data is?22:23
jlkI do not. and I'm just walking out teh door now22:23
clarkbno worries then22:24
clarkbwe've had this problem for a lot longer than a day:)22:24
jlkyou're using merge_commit_sha as the key, not the HEAD of the PR ?22:26
clarkbjlk: our cache maps shas to PRs. We are putting the regular head sha and the merge_commit_sha both in the cache22:27
*** sdake has quit IRC22:27
clarkbjlk: the idea being once the PR merges the status event will be on the merge_commit_sha22:27
jlkI think that merge_commit_sha is a background task, it may update whenever a human hits the PR page, or some other background timer? I'm not totally sure. I don't get the feeling that it's calculated at every API get.22:28
clarkbbut it wouldn't surprise me if the merge_commit_shas are updated in batches as evaluating what things merge to every time tehre is another merge is likely not cheap22:28
clarkbya so that is a best effort caching and may not be completely accurate22:28
jlkyeah, and I think that assumes a particular type of merge, which if it's a human it could be different22:28
clarkb(sometimes we'll get lucky)22:28
*** hashar has quit IRC22:28
jlksquash vs rebase vs whatever22:28
clarkboh good point22:29
jkttristanC: I'm almost going to bed, but https://review.openstack.org/638469 . Also, I'm quite tempted to run a full init (systemd in my case) in the container just so that I could use ansible roles for e.g. starting httpd22:31
jkttristanC: do you see a huge blocker in that?22:31
tristanCjkt: well there is security concern if you enable privileged access.22:41
tristanCjkt: if you want r/w access to /, then we might want to look at better runtime like lxd22:44
jkttristanC: it's on overlayfs22:46
tristanCjkt: i meant, you'll need to become root, and then you are exposed to issue like CVE-2019-573622:49
tristanCjkt: the intend of the runc driver is to be as safe as possible by locking down the environment to only enable unprivileged task like tox22:50
tristanCjkt: it can surely be extended to authorize privileged access, but instead of managing overlayfs manually, we may want to look at another runtime for that22:51
tristanCjkt: for example i think "podman --systemd" does a better job at running systemd in a container22:52
mordredtristanC, jkt: I'm halfway through writing an email on that very topic actually - hopefully I'll be finished by tomorrow23:07
tristanCmordred: i'm looking forward to it :)23:11
mordred\o/23:12
*** sdake has joined #zuul23:13
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: gerrit: add support for report only connection  https://review.openstack.org/56821623:20
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add /{tenant}/buildset/{uuid} route  https://review.openstack.org/63007823:25
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: zuul-runner: add quick-start integration test  https://review.openstack.org/63570123:28
tristanCcorvus: can you please reconsider your -2 on https://review.openstack.org/555153, last PS should have addressed your concerns23:32
corvustristanC: yes, sorry i have a backlog of reviews to do, i've been unusually busy with other work23:33
*** sdake has quit IRC23:33
corvustristanC: i did just add a -2 to 635701 though.  of course feel free to continue using that change to debug problems with testing -- i just didn't want you to spend too much time on writing a quick-start based test.23:34
*** sdake has joined #zuul23:38
openstackgerritMerged openstack-infra/zuul-jobs master: run-buildset-registry: run a dual registry  https://review.openstack.org/63851423:39
openstackgerritMerged openstack-infra/zuul-jobs master: use-buildset-registry: support running before docker installed  https://review.openstack.org/63818023:39
corvustristanC: it looks like the web trigger change is injecting job variables?  that also violates the model.  the model is that events enqueue refs.  once the ref is enqueued, we forget the information about the triggering event.  any trigger driver needs to be able to enqueue a ref and cause the same jobs to run.23:40
tristanCcorvus: arg... how would the web trigger generate a ref then?23:42
openstackgerritMerged openstack-infra/zuul-jobs master: Split docker mirror config into its own role  https://review.openstack.org/63819523:43
openstackgerritMerged openstack-infra/zuul-jobs master: Use buildset registry push endpoint  https://review.openstack.org/63852023:43
corvustristanC: it doesn't have to create a ref, it has to tell zuul to enqueue a ref.  for example, the timer trigger enqueues the branch-tip ref.23:43
tristanCright, but how to set custom variables then?23:43
corvuswe can't -- triggers can't set variables.23:44
tristanCwell, i would argue that a gerrit triggers event can set variables...23:44
corvusthe whole point of the model is that all of the workflow actions are encoded in configuration stored in git.  there's no ad-hoc running jobs manually with special arguments.23:44
corvustristanC: the gerrit and github triggers can not set job variables23:45
tristanCi could create a review with new variable, and the resulting gerrit triggered event would apply those variables to the jobs23:45
corvustristanC: right, but you've created a git commit to do that.  that's part of the git-driven workflow.23:46
*** sdake has quit IRC23:46
corvusthere's a record of it, other people can see it, copy it, etc.  there's no reliance on humans setting parameters manually.23:47
*** sdake has joined #zuul23:47
tristanCcorvus: well you could delete the review in recent gerrit23:47
corvusthat doesn't sound like a *good* git-driven workflow, in fact, it sounds like a very bad one, but it does still sound like a git-driven workflow.23:50
*** sdake has quit IRC23:51
*** sdake_ has joined #zuul23:51
corvusthe fundamental issue is that we must not create a system where the only way to do something in zuul is manually.23:51
tristanCcorvus: i understand the point of git-driven workflow, but asking our user to create fake review to start a job with a custom parameter is a no-go unfortunately23:54
corvustristanC: so we're going to have to continue to dive into the underlying use case23:54
tristanCcorvus: perhaps the webtrigger could take care of creating and closing the review using the gerrit connection?23:55
tristanCcorvus: iiuc it's really a common use case for jenkins user, that is to start a job with custom parameter23:55
corvustristanC: right, that's not a use case23:55
corvusthat's "jenkins does something this way"23:56
corvusi want to know what thing you're trying to accomplish23:56
corvusif the answer is "a user wants to run a job with custom parameters" then the next question is "why do they want to do that?  what are they trying to cause to happen?"23:57
tristanCcorvus: iiuc, Q[AE] folks wants to run job manually, for example to reproduce a customer case, they would start the same ci job but with the right set of option to investigate23:58
tristanCcorvus: which is something they can do with jenkins through the web interface23:58
corvustristanC: well, won't that give them the same answer?23:59
corvusit's a gating system, presumable it has already run with those parameters and passed.23:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!