Friday, 2017-09-08

pabelangerSpamapS: ya, I've just been using zookeeper from fedora for testing, so centos-8 will have something. But, we can likely push zookeeper-lite from tristanC / dmsimard to COPR or something00:03
SpamapSmeh00:03
SpamapStarball seems to be working-ish00:03
pabelangerI have an ansible-role for zookeeper I plan on adding tarball support for also00:03
pabelangercool00:03
SpamapSpabelanger: -> BonnyCI/hoist00:03
pabelangerSpamapS: are you centos now?00:04
SpamapSpabelanger: aye00:05
mordredjeblair: http://logs.openstack.org/02/500202/24/check/devstack/a0ebcd2/job-output.txt.gz#_2017-09-07_22_29_05_05631600:06
mordredjeblair: I think you got further00:06
pabelangerSpamapS: cool00:07
SpamapSsorta ;)00:07
pabelanger:)00:08
ianwSpamapS: if you want "shoved into a rpm" then i've got -> https://copr.fedorainfracloud.org/coprs/iwienand/zookeeper-el7/packages/00:08
SpamapSianw: you're about 20 minutes late00:09
ianwSpamapS: it's ok, nothing happens for years ,then everything happens in 20 minutes ;)00:10
SpamapSyep00:11
SpamapSrealistically, I'm just pushing this bonnyci work to completion, and then I'll probably just install Software Factory and be happy00:11
openstackgerritMerged openstack-infra/zuul-jobs master: Add support for debian in configure-mirrors  https://review.openstack.org/50153700:11
openstackgerritMerged openstack-infra/zuul-jobs master: Add support for fedora in configure-mirrors  https://review.openstack.org/50153800:18
openstackgerritMerged openstack-infra/zuul-jobs master: Add support for opensuse in configure-mirrors  https://review.openstack.org/50153900:19
openstackgerritMerged openstack-infra/zuul-jobs master: Remove project pipeline definition from zuul-jobs  https://review.openstack.org/50184900:20
ianwhave i done something wrong or is https://review.openstack.org/#/c/501904 not getting picked up for testing?00:28
ianwhmm, neither is a recheck of an old job00:29
ianwit seems to be in a pretty hard loop around http://paste.openstack.org/show/620679/00:36
ianwjeblair / mordred: ^ ...00:39
mordredianw: hrm00:42
mordredianw: I agree with you - that definitely looks like a hard loop00:43
ianwvariations of00:43
ianw2017-09-08 00:43:14,011 INFO zuul.DependentPipelineManager: Reported change <Change 0x7fecd74a41d0 501849,2> status: all-succeeded: True, merged: True00:43
ianw2017-09-08 00:43:14,115 INFO zuul.DependentPipelineManager: Reported change <Change 0x7fecd7010c88 501538,5> status: all-succeeded: True, merged: True00:43
ianwover and over00:43
mordredianw: 501849 removed a pipeline definition from zuul-jobs that was in the zuul-jobs repo00:45
mordredianw: a few bugs have flushed out related to pipeline removals - so maybe something went south?00:45
mordredianw: in any case, at that rate we're going to run outof disk space due to logging00:46
mordredianw: so how about we restart the scheduler (if there's not enough logging now there never will be)00:46
ianwsgtm00:46
mordredI have stopped it - restarting00:47
mordredit's now reading is config00:48
dmsimardwhat did I break00:49
mordredianw: I rechecked your change and it seems to be enqueued00:49
dmsimardmordred: I made sure to put depends-on on my forklift patches to ensure things did not merge out of order00:50
dmsimardmordred: https://review.openstack.org/#/q/topic:zuulv3-forklift00:50
mordreddmsimard: I don't think it was that - my hunch is that we tickeled a weird edge-case somewhere00:50
dmsimardoh wow my patch series finally landed \o/00:50
dmsimardneed +3 on two patches to bring back zuul-jobs jobs into place https://review.openstack.org/#/q/topic:zuulv3-forklift00:51
SpamapSgah.. so many apt: calls to fix :-P00:52
mordredianw: although we're now just sitting in queued state00:53
mordredianw: nevermind. I'm just impatient00:53
dmsimardI'm not sure I understand the deal with that apt issue00:54
dmsimardansible is unable to use python-apt until you do apt-cache update or something like that ?00:55
ianwdmsimard: just been looking at them, as i was doing 50190400:55
ianwmordred / dmsimard: are we at a point we can merge https://review.openstack.org/#/q/topic:zuulv3-forklift ?00:58
dmsimardianw: yes, it's tech debt from merging my configure-mirror tree00:58
ianwit lgtm, if zuul's up to it :)00:58
ianweverything seems to be moving, so ok00:58
dmsimardianw: for the ci creation script, go ahead00:59
dmsimardianw: if it's part of what would be the base job, add the role in the base.yaml test playbook00:59
dmsimardthat's where validate-host will end up as well00:59
dmsimardalthough I still need to figure out what to do with that one, it has tasks that can only run on the executor..01:00
ianwdmsimard: yep ... so for testing purposes, just assert: that: tests that things look right?01:00
ianwfile is there, perms are right, etc?01:00
ianwor is there a better way?01:00
dmsimardianw: checking that it runs without horribly failing is already a good start (the job will fail if it fails horribly)01:00
dmsimardianw: asserts are good too.01:00
dmsimardianw: there's some examples here if you want: https://github.com/ansible/ansible/tree/devel/test/integration01:00
dmsimardianw: some examples of what I've written https://github.com/ansible/ansible/blob/devel/test/integration/targets/sensu_client/tasks/main.yml01:01
dmsimardhttps://github.com/ansible/ansible/tree/devel/test/integration/targets/include_vars01:01
ianwexcellent thanks; looks like roughly what i had in my head so a good sign ;)01:02
*** harlowja has quit IRC01:11
tristanCianw: fwiw, there is a zookeeper-lite rpm package built in software-factory repository01:33
ianwtristanC: cool, is that more or less the tarball in rpm foramt?01:40
tristanCianw: it's built from source for el7, the trick was to remove all the client and netty, hence the "-lite" suffix01:42
tristanCpackage is https://softwarefactory-project.io/kojifiles/repos/sf-2.6-el7-release/Mash/zookeeper-lite-3.4.10-1.el7.x86_64.rpm, .spec is https://softwarefactory-project.io/r/gitweb?p=software-factory/zookeeper-lite-distgit.git;a=tree01:42
ianwtristanC: excellent, updated https://etherpad.openstack.org/p/zookeeper-epel7 in case any googlenauts find it01:54
dmsimardwanted to be productive tonight but spent the evening recovering from rdo infrastructure outage01:54
dmsimardianw: if you want to take a stab at devstack centos/fedora/suse/debian go for it, I'll follow up tomorrow01:55
ianwok, see how i got this afternoon.  need some lunch first! :)01:56
dmsimardianw: if you try out fedora, I can probably just pattern off of that for01:56
dmsimardcentos* and the others01:56
tristanCianw: thanks!01:58
*** olaph1 has quit IRC02:08
*** xinliang has quit IRC02:12
*** xinliang has joined #zuul02:24
*** xinliang has joined #zuul02:24
dmsimardall the cloud outage things are fixed02:53
dmsimardgoing to bed now gnight02:53
dmsimardo/02:53
*** jkilpatr has quit IRC02:59
tristanCSpamapS: there is an upcoming blog post about using software-factory as a third-party-ci: https://softwarefactory-project.io/r/#/c/9473/3/Openstack_3rd_Party_CI_with_SF_26/2017-08-28-openstack-3rd-party-ci-with-software-factory.html.md03:07
tristanCSpamapS: and the general documention is https://softwarefactory-project.io/docs/operator/deployment.html03:08
SpamapStristanC: awesome. :) I really do intend to give it a shot. I just figured I invested half a day in centos-ifying BonnyCI/hoist, I might as well finish :)03:08
tristanCAs you wish, I think the main difference is that software-factory uses rpm packages for everything and the sfconfig script automatically generate all the secrets03:09
SpamapSthere are likely tons of differences :)03:09
tristanCmay i ask how are you getting python3 on centos?03:10
SpamapSbut ultimately, if I can get a zuulv3 up that talks to our internal Github and starts running jobs on our cloud... I can show people the magic. :)03:10
SpamapStristanC: haven't crossed that bridge yet. :)03:10
SpamapSI'm still in the stage of converting all the apt's to yums.03:11
SpamapSand went through the really silly-feeling process of making mariadb wori03:11
SpamapSwork03:11
tristanCalternatively, you could try using software-fatory repository so that you could yum install zookeeper, zuul and nodepool03:12
tristanCwell fwiw, all the distgit are available over softwarefactory-project.io gerrit, and package update goes through full integration test so contribution are welcome too :-)03:14
SpamapSneat03:14
ianwjeblair / mordred : zuul has hung again, and the logfile is up to 16gb ...03:23
ianwhttp://paste.openstack.org/show/620684/03:24
ianwi'm going to restart it, because this doesn't end well03:25
SpamapShrm so why isn't there python3.5 in EPEL? :-P03:35
SpamapSor does python3.4 work ok for zuul/nodepool these days? :-P03:35
* SpamapS can't remember what the min version was03:35
tristanCSpamapS: you could get python35 using softwarecollections03:38
SpamapStristanC: ah03:38
tristanCSpamapS: but then, zuul-executor needs libpython35 in /usr/lib because bwrap will drop a custom ld_library_path03:38
SpamapSround and round we go03:39
SpamapStristanC: you're scoring points :)03:41
* SpamapS is having to update bonnyci's code to be very explicit about pip/python executables03:41
tristanCSpamapS: sort of, imo the real point is using zuul, what ever distribution/confmgmt ;)03:43
openstackgerritIan Wienand proposed openstack-infra/zuul-jobs master: Expand PATH for SUSE systems to include /usr/sbin  https://review.openstack.org/50173703:45
clarkbpy35 is required bcause of the type annotations03:50
clarkbI dont follow why bubblewrap factors into that03:51
tristanCclarkb: when using software-collections, python35 is installed in /opt and the .so isn't declared in the main ld.so.conf03:52
clarkbI had zuul running on py34 for gerrit testing but that was before the type annotations03:52
clarkbtristanC: oh software collection specific thing03:52
tristanCclarkb: yes, regarding python35 on centos7 with scl03:53
tristanCthe issue is that bwrap is setuid and it will drop the custom LD_LIBRARY_PATH from zuul-executor, so ansible-playbook will failed with missing libpyton3503:55
clarkbgot it03:56
tristanCthough it's easy to fix, just symlink the python lib from /opt to /lib03:56
*** xinliang has quit IRC03:59
ianwdmsimard: ahh, i think we've had a bit of a split-brain thing happening between depends-on, zuulv2 and zuulv3 merging with your jobs moving tests.  just untangling it now ...04:02
openstackgerritMerged openstack-infra/zuul-jobs master: Expand PATH for SUSE systems to include /usr/sbin  https://review.openstack.org/50173704:02
*** xinliang has joined #zuul04:06
ianwoh, you know what ... it's because there's no jenkins reporting on the depends-on04:10
SpamapSweird... pip: is using 'pip2' instead of 'pip' which is causing all kinds of weirdness for me :-P04:27
openstackgerritIan Wienand proposed openstack-infra/zuul-jobs master: Add integration tests for validate-host  https://review.openstack.org/50154304:34
tristanCSpamapS: btw, you'll also need git-2 because zuul is using GIT_SSH_COMMAND, which is also avail in scl with rh-git2905:05
tristanCand well, if you don't want to bother with those details, you could get the service installed now using "yum install -y https://softwarefactory-project.io/repos/sf-release-master.rpm && yum install -y rh-python35-zuul-* rh-python35-nodepool-*"  ... just saying05:11
openstackgerritMerged openstack-infra/zuul-jobs master: Add a role to emit an informative header for logs  https://review.openstack.org/50149505:33
*** harlowja has joined #zuul06:19
*** harlowja has quit IRC06:40
*** fbo_ has joined #zuul06:43
*** fbo_ is now known as fbo06:48
openstackgerritTobias Henkel proposed openstack-infra/zuul feature/zuulv3: Use connection type supplied from nodepool  https://review.openstack.org/50197606:58
*** yolanda has joined #zuul07:07
*** electrofelix has joined #zuul08:38
*** hashar has joined #zuul09:01
openstackgerritTobias Henkel proposed openstack-infra/nodepool feature/zuulv3: Support username also for unmanaged cloud images  https://review.openstack.org/50080809:04
openstackgerritTobias Henkel proposed openstack-infra/nodepool feature/zuulv3: Add password to build and upload information  https://review.openstack.org/50201109:04
tristanCI started the process of writting a SELinux policy for zuul, once it passes some integration test i'd like to submit upstream for review if you are ok with this09:55
tristanCfwiw, the type enforcement file looks like this: https://softwarefactory-project.io/r/#/c/9593/2/zuul/sf-zuul.te09:56
*** jkilpatr has joined #zuul10:57
*** olaph has joined #zuul12:03
dmsimardjeblair, mordred: so I was thinking last night.. with ara 1.0, there is the opportunity to make a *very* bare bones callback with (almost) just 'requests' as a dependency for use with the HTTP REST API. The python API leverages the flask app so it still depends on some things (less than the entire webapp, like no xstatic, etc.).. as things progress along, I'll see what are the opportunity for a more "bare12:08
dmsimardbones" python API callback12:08
*** odyssey4me has joined #zuul12:43
*** hashar has quit IRC13:28
*** hashar has joined #zuul13:28
*** dkranz has joined #zuul13:56
rcarrillocruzhey folks13:59
rcarrillocruzhow's bubblewrap installed13:59
rcarrillocruzvia distro package?13:59
rcarrillocruzsetting up a zuul, getting failures due to missing bwrap binary13:59
odyssey4mercarrillocruz are there instructions around for zuul v3 - or is this zuul v2.5 ?14:03
rcarrillocruzsetting up v314:03
rcarrillocruzhmm14:05
rcarrillocruzhttps://github.com/openstack-infra/puppet-zuul/blob/master/manifests/executor.pp14:05
mordredrcarrillocruz: there'sa  PPA we use for ubuntu-xenial14:06
rcarrillocruzyah, just enabled https://launchpad.net/~openstack-ci-core/+archive/ubuntu/bubblewrap/+index14:07
rcarrillocruzthx14:07
mordredrcarrillocruz: it's already in fedora - I'm not sure about the story for centos14:07
mordred++14:07
rcarrillocruzoh14:07
rcarrillocruzfedora 26?14:07
rcarrillocruzi mean, right now i'm in AIO, scheduler/merger/executor all in a xenial14:08
rcarrillocruzbut that's interesting14:08
rcarrillocruzin other news14:09
rcarrillocruzhttps://github.com/rcarrillocruz-org/zuul-tested-repo/pull/514:09
rcarrillocruzi have GH reporter working14:09
rcarrillocruz\o/14:09
rcarrillocruzodyssey4me: oh sorry, didn't follow, you are willing to spin up a v3?14:11
rcarrillocruzi'm doing an install using ansible-role-zuul14:11
rcarrillocruzgetting notes of things i'm encountering that need manual fix14:11
rcarrillocruzfor later patches14:11
rcarrillocruzbut overall, the role sets up a v3 AIO nicely14:11
mordredrcarrillocruz: wot!14:12
odyssey4mercarrillocruz yes, I'm particularly interested in nodepool at this stage - not sure how tightly coupled it is to zuul v314:12
mordredodyssey4me: it can totally be run on its own14:12
rcarrillocruzodyssey4me: well, i'm using nodepool standalone for a very rough CI in Ansible networking14:13
mordredodyssey4me: nodepool in the feature/zuulv3 branch is the one you want if you're setting up a nodepool14:13
odyssey4meah, excellent - not that I wouldn't prefer our CI to use zuul... but for now we have jenkins :(14:13
rcarrillocruzused ansible-role-nodepool14:13
rcarrillocruzhappy to help14:13
mordredodyssey4me: so - there's a thing that I think would be GREAT if someone wrote14:13
mordredodyssey4me: but I'm not going to because I'm busy14:13
odyssey4memordred aha, that'd be what I'm looking for then14:13
mordredodyssey4me: which is a nodepool plugin for jenkins14:13
odyssey4mercarrillocruz so ansible-role-nodepool works with the feature branch version?14:14
rcarrillocruzyep14:14
rcarrillocruzby default it pulls feature/zuulv314:14
odyssey4memordred hmm, yes - that might actually just end up happening14:14
mordredodyssey4me: nodepool v3 uses zookeeper for zuul to request nodes from it - it should be VERY easy for someone with the javas to write a plugin for jenkins that could request nodes from nodepool using the same api14:14
odyssey4mercarrillocruz sweet, that helps a bunch - thanks14:14
mordredodyssey4me: I'd personally like it because I think it makes a really nice migration path for folks - or for situatoins where you want to run zuul and jenkins side by side14:15
rcarrillocruzthings you need to do before ansible-role-nodepool invocation:14:15
rcarrillocruz1.bootstrap python14:15
rcarrillocruz2. bootstrap pip14:15
rcarrillocruz3. generate ssh key14:15
mordredodyssey4me: both a zuul and a jenkins witha zk plugin could totally consume nodes from the same nodepool with no problems14:15
rcarrillocruz4. copy over clouds.yaml to nodepool home folder14:15
rcarrillocruz5. invoke ansible-role-zookeeper14:16
odyssey4memordred yeah, I saw the intended lock feature - so that makes sense14:16
rcarrillocruz6. invoke ansible-role-nodepool, passing as param the nodepool_file_nodepool_yaml_src var (it creates nodepool.yaml)14:16
rcarrillocruzi'm in the process of creating a network-infra, it's private repo, once i split off secrets from it i wanna make it public14:17
rcarrillocruzwill ping you when done14:17
odyssey4mercarrillocruz that'd be great14:17
rcarrillocruzbut the nodepool playbook is pretty much the steps i depicted above14:17
odyssey4meI wonder if we should etherpad that little procedure and share war stories as we go14:18
odyssey4meI suppose a review to the repo would probably make better sense.14:18
odyssey4meI'll take these as notes and prep a patch to the README14:18
odyssey4methanks rcarrillocruz once again14:18
rcarrillocruz++14:18
rcarrillocruzi need to app myself a doc patch describing what perms are bare minimum to create a zuul gh app14:20
rcarrillocruzs/app/write14:20
*** olaph has quit IRC14:34
*** olaph has joined #zuul14:36
dmsimardrcarrillocruz: what are you deploying on ? centos ?14:51
rcarrillocruzxenial14:51
dmsimardrcarrillocruz: okay -- if you deployed on centos I might have been able to help :)14:52
dmsimardrcarrillocruz: https://www.rdoproject.org/blog/2017/03/standalone-nodepool/14:53
dmsimardout of software factory14:53
dmsimardrcarrillocruz: oh, that might be for odyssey4me instead though14:53
rcarrillocruzotoh, i believe sf is not on zuulv3 yet14:53
rcarrillocruz?14:53
dmsimardrcarrillocruz: it's there: https://softwarefactory-project.io/zuul3/14:54
dmsimardrcarrillocruz: tristanC might have more details but I believe all the necessary bits to use it are there but the next version of SF should contain the "real" v3 release14:54
rcarrillocruzorly? yanis told me SF v3 was still to be deterimned14:54
odyssey4methanks dmsimard - that may prove useful :)14:54
dmsimardmordred: which brings the question, do you guys plan on merging feature/v3 back into master and start tagging releases at some point ? :)14:55
rcarrillocruzhmm14:55
rcarrillocruzsudo yum install -y --nogpgcheck https://softwarefactory-project.io/repos/sf-release-2.5.rpm14:55
rcarrillocruzi wonder if that v3 was hand rolled14:55
rcarrillocruzand is what yanis referred to, that there no v3 packages14:56
dmsimardrcarrillocruz: v3 is not in 2.514:56
dmsimardrcarrillocruz: it landed in 2.614:56
rcarrillocruzyanis == spredzy on IRC for those who don't know14:56
dmsimardrcarrillocruz: install 2.6 instead14:56
rcarrillocruzah, so SF releases are not related to zuul versioning14:56
mordreddmsimard: yup!14:56
rcarrillocruzgood14:56
mordreddmsimard: well - we plan on merging v3 back into master for sure14:57
mordreddmsimard: and I believe we plan on tagging a 3.0 release once we think it's 'ready' for other folks14:57
dmsimardrcarrillocruz, odyssey4me: docs here https://softwarefactory-project.io/docs/operator/deployment.html14:57
dmsimardrcarrillocruz: correct, sf releases are not tagged according to zuul release :)14:58
mordreddmsimard: ongoing releases are an interesting question - since we run CD from master (or will be once it's re-merged) - releases wind up being arbitrary snapshots -so we'll need to think about the semantics of that14:58
dmsimardrcarrillocruz: 2.6 ships with jenkins, zuul-launcher (so zuul v2.5 without jenkins) *and* zuul v3, it's a release to help with transition14:58
rcarrillocruzhmm, k14:59
rcarrillocruzpabelanger: ^14:59
tristanCrcarrillocruz: the current roadmap is sf-2.x support both zuulv2 and zuulv3, and a future sf-3 will be zuulv3 only14:59
rcarrillocruzwe talked about SF and v3 not that long15:00
rcarrillocruz++15:00
tristanCrcarrillocruz: sf-2.6 does have a working "tech-preview" zuulv3, and we are waiting for upstream release of zuulv3 to release it part of sf-2.715:00
tristanCthe master repository (that will be sf-2.7) does have all the nodepoolv3/zuulv3 bits in place15:02
tristanC+ the static and opencontainer driver so that you can test the full stack without a cloud15:04
rcarrillocruzOH15:05
rcarrillocruzthat's neat!15:05
rcarrillocruzi thought static driver was in review15:05
rcarrillocruzhas that been merged?15:05
tristanCrcarrillocruz: not yet but I've added it to the nodepool3.rpm so that it's easy to test15:06
jeblairmordred: post-hoc -1 review of https://review.openstack.org/50188615:06
tristanCrcarrillocruz: and we just added an integration test that does create oci slave to verify zuulv3 can merge patch15:06
rcarrillocruzhmm, what's ansible_user value on the executor15:07
rcarrillocruzis it root by default?15:07
rcarrillocruzthe user it ssh to on nodepool nodes15:07
tristanCrcarrillocruz: yes, nodepool needs root access to create slave15:08
tristanCrcarrillocruz: you can have a look at this integration test: https://github.com/softwarefactory-project/sf-ci/blob/master/health-check/nodepool3.yaml15:08
tristanCrcarrillocruz: which use this nodepool.yaml: https://github.com/softwarefactory-project/sf-ci/blob/master/health-check/templates/nodepoolV3.yaml.j215:08
tristanCsimilarly, this zuulv3 verify a change can be merged: https://github.com/softwarefactory-project/sf-ci/blob/master/health-check/zuul3.yaml15:09
jeblairpabelanger, clarkb: fyi see comments on https://review.openstack.org/50188615:09
tristanCthose tests are run on every software-factory change :)15:09
pabelangermorning15:23
pabelangerjust catching up on backscroll15:23
pabelangerjeblair: clarkb: thanks, keep forgetting about that15:24
mordredjeblair: GAH. thank you15:30
tristanCleaving for denver soon, see you there folks!15:35
*** hashar is now known as hasharAway15:45
rcarrillocruzare we good to go https://review.openstack.org/#/c/500808/315:50
rcarrillocruzi need this to consume user from nodepool settings, to log into network appliances images15:51
rcarrillocruzmordred: ^, you +2'd previously15:58
rcarrillocruzhmm, on the other hand, where in the code of zuul we consume 'username' from nodepool? i see executor/server.py it creates ansible_user off executor.default_username, but don't see setting the user on hostvars off nodepool node15:59
rcarrillocruzi guess by passing the full provider, as it contains labels, then username16:02
mordredtristanC: see you in denver!16:07
mordredrcarrillocruz: tobiash has patches for zuul to consume the username field once it's there16:08
mordredrcarrillocruz: you might also want to review https://review.openstack.org/#/c/500800/2 and https://review.openstack.org/#/c/453968/316:08
rcarrillocruzthx for the +116:08
rcarrillocruzi'm a bit torn exposing passwords on nodepool.yaml16:08
rcarrillocruzmaybe it should go in something like secure.yaml or the likes16:08
rcarrillocruzlike16:08
rcarrillocruznodepool.yaml have it in core review public repo16:08
rcarrillocruzsecrets in private repo or somewhere else16:09
rcarrillocruzi refer to https://review.openstack.org/#/c/502011/16:09
mordredrcarrillocruz: I just left that same -116:10
mordredrcarrillocruz: we have an optional secure.yaml already - I think the docs in this case should talk about using it for that field16:10
mordredrcarrillocruz: also - I tihnk we need doc mentions on securing zookeeper16:11
rcarrillocruzmy mem may fail on me, secure.yaml was for clouds.yaml passwords?16:12
pabelangerrcarrillocruz: we used secure.yaml for database connection string16:14
pabelangerright now, it is not used16:15
mordredrcarrillocruz, pabelanger: both of you are right ...16:15
pabelangerbut, I think we want to use it for zookeeper auth at some point16:15
mordredos-client-config supports a clouds.yaml and a secure.yaml16:15
mordredand nodepool support a nodepool.yaml and a secure.yaml16:15
rcarrillocruzwhat i'm getting at: will secure.yaml become a general thing to store nodepool creds16:15
pabelangerOh, TIL about os-client-config16:15
mordred(os-client-config supports secure.yaml because nodepool did and it seemed like a nice thing to copy)16:15
rcarrillocruzimages creds16:15
rcarrillocruzconnection strings16:15
rcarrillocruzetc16:15
pabelangerand secure.yaml16:16
mordredrcarrillocruz: secure.yaml already is a general thing in which any setting can go16:16
rcarrillocruzif so, it sounds like we want to have a mini schema here16:16
rcarrillocruzack16:16
mordredrcarrillocruz: it's basically just two files so you can, as an admin, put some in one file with stricter perms as you see fit16:16
mordredor not, if your env is just you and you don't care :)16:16
rcarrillocruzheh16:17
mordredpabelanger: yah - we could split our nodepool clouds.yaml files if we wanted, put secure.yaml files out there with our passwords managed in system-config and put the other bits into project-config16:18
pabelanger++16:18
jeblairtobiash: http://paste.openstack.org/show/620735/ errors related to connection cache16:21
rcarrillocruzbtw, timer trigger doesn't work on github repos16:23
rcarrillocruzhttp://paste.openstack.org/show/620737/16:23
rcarrillocruzit's supposing events will have commits16:23
jeblairthat error means that we're not completing the reconfiguringation process16:23
rcarrillocruzthat logic seems to need a move16:23
rcarrillocruzwill prep a patch16:23
jeblairrcarrillocruz: thx16:23
jeblairmordred, ianw: it is possible the maintainCache bug may be the cause of the stuck changes in queue.  when we reconfigure (because we land a config change), we re-enqueue all the changes in new pipelines, then maintain the connection cache (which fails now and aborts the process).  *after* that we re-establish the shared change queues in the pipeline postConfig method.  it seems plausible that since that's not happening, that a change that was in ...16:29
jeblair... a pipeline across a reconfiguration may not be removed from the correct queue.16:29
jeblairi will prepare a revert16:29
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Revert "Enable maintainConnectionCache"  https://review.openstack.org/50212116:38
jeblairmordred: ^16:40
tobiashjeblair: looks like maintainCache is untested yet in github16:48
jlkhuh, yeah I don't think we've tested that16:48
tobiash+2 for the revert16:49
jeblairsince it was probably copied, it's worth double checking that the bug in the github version of that function isn't also in the gerrit version.  it's possible we ran that on github before we would have run it on gerrit.  in other words, that particular code path may not be tested in gerrit as well.16:51
tobiashjeblair: yes, I assumed it worked in v2 and wasn't modified much16:53
mordredjeblair: +A16:53
mordredjeblair: and ah - your explanation of how that could cause the things we saw seems plausible16:54
pabelangermordred: jeblair: is something wrong with zuulv3.o.o? I am seeing what I think is a loop in the debug.log17:00
*** harlowja has joined #zuul17:00
jeblairi'll restart it17:02
jeblairpabelanger: see scrollback and https://review.openstack.org/50212117:02
pabelangerjeblair: ah, thank you17:05
tobiashmordred, rcarrillocruz: for production use of storing passwords in zookeeper we for sure need to secure zookeeper with auth and encryption17:11
tobiashmordred, rcarrillocruz: both should be supported in zookeeper, but we will have to add tls support to kazoo17:12
clarkbthis is login credentials?17:12
clarkbanother option may be to negotiate those out of band like is done with ssh keys?17:13
tobiashclarkb: yes, that's login credentials to the nodes17:14
tobiashclarkb: then we would need secrets on the executor also for user supplied images and means on the executor to configure different credentials for different labels17:15
clarkbtobiash: yes, which is how ssh works isn't it?17:17
clarkb(the private key being the secret)17:17
tobiashyes17:18
tobiashcurrently there is just a single private key on the executor for all nodes17:18
tobiashand accessing windows nodes unfortunately doesn't work with ssh17:19
clarkbtobiash: but you can use ssl client cert auth with windows17:19
clarkbwhich is similar ish to ssh keys17:19
tobiashclarkb: hm, didn't think about that possibility yet17:22
tobiashsounds cool17:22
tobiashclarkb: so you suggest to add a winrm_client_cert setting to the executor to avoid password passing?17:23
clarkbtobiash: ya, I mean I've never had to actually use it but seems like a good match up to how things work with ssh keys17:23
clarkband that might simplify bootstrapping17:24
mordredit's worthwhile checking about how windows guests work on openstack, but also on aws - do you get an administrator password returned in the nova server or ec2 instance data from the API?17:24
mordredmostly because if that's the mechanism available for those, then figuring out password passing will still need to happen at some point17:25
tobiashmordred: in our v2 env the windows guests work quite well17:25
tobiashthey just take longer to boot17:25
tobiasharound 2min instead of 50s17:25
tobiashbut we preconfigured the login stuff in the image17:26
mordredtobiash: how does auth work with that? also - someone was asking about image building too17:26
jeblair2017-09-08 17:26:19.542490 | ubuntu-xenial | 2017-09-08 17:26:19.542 | stack.sh completed in 944 seconds.17:27
jeblairclarkb, mordred: ^17:27
jeblairhttp://logs.openstack.org/02/500202/25/check/devstack/dbd6e11/17:27
tobiashmordred: we built that image by hand and added cygwin ssh service with public key authentication such that nodepool and jenkins are happy17:27
tobiashbut windows was a side use case before and now we need to professionalize this with automated image building and so on17:28
clarkbjeblair: nice17:29
tobiashok, client cert auth should be possible (although undocumented) with ansible: https://github.com/ansible/ansible/issues/1624317:34
mordredjeblair: WOOT17:38
mordredtobiash: cool17:39
pabelanger502121 looks stuck in gate :(17:39
clarkbdoes make me wonder why windows doesn't do ssh key auth for powershell. Its all bsd licensed too isn't it?17:39
jeblairmordred, Shrews: do you know if anyone working on the websocket streaming test failures?17:40
jeblairpabelanger: i'll force-merge it and restart i guess17:40
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Revert "Enable maintainConnectionCache"  https://review.openstack.org/50212117:41
pabelangerjeblair: ty17:41
tobiashclarkb: https://blogs.msdn.microsoft.com/powershell/2015/10/19/openssh-for-windows-update/17:43
tobiashBut still ansible would have to support win modules via ssh17:44
jeblairclarkb: do you want to take a look at 496959, 496301, 500202 ?17:45
jeblairmordred: and you the last 2 of those? ^17:45
jeblairi think that's a good checkpoint; we can land those and iterate from there (i think we still need to add the disk partitioning role to that, and then multinode, and tempest... :)17:46
clarkbmordred: jeblair for 959 homedir is moving from /opt/stack/new to /opt/stack ?17:49
*** electrofelix has quit IRC17:50
jeblairclarkb: yep; since we're starting from scratch here, i'm trying to be as close to devstack default as possible  (i'd like to actually have devstack do that -- it is capable of doing so, but there's some chicken/egg stuff with git repos we'd need to work out first)17:50
clarkband tempest homedir is default path which isn't a change I don't think17:51
clarkb?17:51
jeblairclarkb: i believe that's correct17:52
mordredjeblair: +2 from me all around17:53
clarkbya confirmed tempest isn't moving just continues to use default17:54
mordredjeblair, Shrews: I am not working on websocket streaming test failures17:54
clarkbjeblair: and the localrc sort method is not alnum or similar it is instead writing out referenced vars first before they are referenced?17:57
clarkbya ok thats what vargraph is17:59
mordredjeblair: I just rechecked https://review.openstack.org/#/c/500365 to see how it goes18:05
pabelangermordred: jeblair: puppet has installed 502121 on zuulv3.o.o, safe to restart scheduler?18:05
mordredpabelanger: fine by me - I just rechecked a job, so you may want to do graceful save/restore - or else I can just recheck once you're done18:06
pabelangerI am not sure, I think zuul is stuck in loop? cc jeblair18:06
mordredoh- then absolutely safe to restart18:06
pabelangerk18:06
pabelangerokay, restarted18:07
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Add migration tool for v2 to v3 conversion  https://review.openstack.org/49180518:09
mordredjeblair, clarkb, pabelanger: ^^ that's ready for people to take stabs at18:09
mordredI added a script "tools/run-migration.sh" that you can use to run it locally if project-config is adjacent to zuul in your local git structure18:10
mordredit expects the project-config repo to have the zuul/mapping.yaml file in it though18:10
mordredI also added a check job with that patch that runs the script on project-config and collects the output18:10
mordredand I put in some notes at the top of the file on things that need to be implemented - I think we can divy those up18:12
mordredalso, while there are some classes in this tool, it's also vintage mordred code, which means it's not always doing things the most sanest way18:12
clarkbjeblair: I am +2 on those changes too. Not approving because have plumber here again and distracted18:15
mordredjeblair: errors like this when running a job:18:18
pabelangeroh noes18:18
mordred"shade-ansible-devel-functional-devstack shade-ansible-devel-functional-devstack : ERROR Unable to find playbook /var/lib/zuul/builds/f244873046ed46a8abcdbb7a036008ea/work/src/git.openstack.org/openstack-infra/shade/playbooks/devstack/pre-run"18:18
pabelanger2017-09-08 18:17:00,357 DEBUG zuul.AnsibleJob: [build: 924886064b4f45d28785c6851c60bc27] Ansible output: b'ERROR! A worker was found in a dead state'18:18
pabelangerthat was on ze0118:18
mordredpabelanger: ugh18:19
pabelangerwe seem to still have our PPA version installed18:20
mordredjeblair: how hard would it be to catch those in the job parser / validator? I mean, I guess it would involve the parser making additional cat job calls to get the playbook content18:20
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Add migration tool for v2 to v3 conversion  https://review.openstack.org/49180518:20
mordredpabelanger: also - could you review https://review.openstack.org/#/c/500320/ for me please?18:21
mordredclarkb: or you - you +2'd it before, but then I squashed it with the previous patch18:22
pabelangerlooking18:22
mordredpabelanger: and https://review.openstack.org/#/c/501001/ to go along with it18:23
jeblairmordred: yeah, the problem with catching that at validation is that we don't know where to look.  we'd have to either ask the merger for a list of *all* files, or have two round trips to the merger.18:24
mordredjeblair: nod18:25
Shrewsjeblair: mordred: i was not aware of websocket test failures. example?18:26
mordredShrews: it's a sporadic failure18:26
mordredShrews: one sec- lemme find one18:27
jeblairmordred, pabelanger, clarkb, jlk, tobiash, dmsimard, Shrews, anyone else: how about we take a look at mordred's migration code, and maybe convene around 20:00 utc and work on a plan for next steps / ptg?  does that time work for folks?  alternate suggestions?18:27
Shrewsmordred: didn't you submit a ipv6 patch for something yesterday?18:27
mordredShrews: http://logs.openstack.org/92/500592/6/gate/tox-py35-on-zuul/679b5c4/testr_results.html.gz18:27
mordredShrews: I did, and it has landed - and it still doesn't work :(18:27
mordredjeblair: ++ I'll be there18:27
mordredShrews: test_websocket_streaming is the only failure in that log that's relevant18:28
*** hasharAway has quit IRC18:28
*** hashar has joined #zuul18:29
Shrewsmordred: any idea when this started happening?18:29
mordredShrews: yah - when we added IPv6 enabled test nodes into the v3 nodepool18:30
mordredShrews: best I can tell it only happens when we run those unittests on one of them18:30
pabelangerjeblair: wfm18:30
dmsimardjeblair: where is the migration code ?18:30
clarkbjeblair: yes that should work for me, though I doubt I will be able to look at the code much.18:31
clarkbgood news is the saga of the plumbing and washing machine is almost over \o/18:31
Shrewsmordred: hrm, seems more zuul_stream related (possibly). no logging output whatsoever18:31
Shrewsfun18:31
jlkjeblair: sounds right18:33
Shrewsor finger server... hrm18:33
jeblairdmsimard: mordred just pushed it up -- 49180518:35
Shrewsi'm betting socketserver may not be ipv6 friendly18:36
Shrewsaddress_family = socket.AF_INET18:38
mordredShrews: gah. I thought I got all of those18:38
Shrewslikely culprit... seeing how to override18:38
Shrewsmordred: this is in socketserver itself18:38
mordredShrews: oh - that's in socketserver itself?18:38
mordrednod18:38
Shrewsmordred: specifically, https://github.com/python/cpython/blob/3.4/Lib/socketserver.py#L41518:40
tobiashjeblair: I'll be around18:40
Shrewsmordred: so, we can override that value... but AF_INETV6 would not support ipv4, right? not sure how to tell which we should use18:41
mordredShrews: AF_INET6 supports both18:42
mordredShrews: you can grep in zuul source for a few places we use it18:43
openstackgerritDavid Shrewsbury proposed openstack-infra/zuul feature/zuulv3: Also support IPv6 in the finger log streamer  https://review.openstack.org/50213718:46
Shrewslet's see what that gets us18:46
openstackgerritDavid Shrewsbury proposed openstack-infra/zuul feature/zuulv3: Also support IPv6 in the finger log streamer  https://review.openstack.org/50213718:57
dmsimardjeblair, mordred: added two small comments to mordred's patch, it's hard to me to review without seeing the end result -- I'll test with the provided bash script and report back if I see anything weird18:59
dmsimarddidn't see anything shocking18:59
mordreddmsimard: sweet. I've got an update/followp coming in just a sec - so I'll go look at your comments before I push it up19:00
Shrewsmordred: where is your zuul/mapping.yaml file referenced in 491805?19:02
dmsimardShrews: I think that's where the mapping ends up being generated ? /me looks again19:03
dmsimardah, nope, the file is expected to be there indeed19:06
dmsimardShrews: https://review.openstack.org/#/c/491804/5/zuul/mapping.yaml19:06
tobiashmordred: I didn't find the generated playbooks in the build result19:07
dmsimardtobiash: I see the new layout in zuul.d/99converted.yaml but I don't see the actual jobs19:10
tobiashdmsimard: yeah, also just noticed that19:10
dmsimardwould jobs be "compiled" at runtime and this is just a migration of the layout ?19:10
tobiashhow is it supposed to be distributed over the repos at the first try?19:11
dmsimardtobiash: it's not afaik19:11
tobiashso all will be in project-config at first?19:11
dmsimardtobiash: it will live in project-config until projects start migrating over ?19:11
mordredtobiash: we are not generating the playbooks yet19:16
mordredthat's one of the next things needed19:16
mordred(we have a bunch of the code for that already in 2.5 that we can copy-pasta in)19:16
dmsimardmordred: I wonder if we should prefix or suffix jobs that have been automatically migrated19:16
mordreddmsimard: aha - but we do! :)19:17
mordreddmsimard: legacy-{name}-ansible-func-centos-7 ... for instance19:17
dmsimardmordred: yeah I saw legacy but it's not everywhere so I thought it was something else19:17
mordreddmsimard: we have a bunch of places where we have already defined new jobs for tolks19:18
mordredso people who were using gate-nova-python27 before just get "tox-py27"19:18
dmsimardmordred: so you ended up defining nodesets after the node names just like we discussed ?19:18
mordredbut if we don't have a nice new job to migrate you to (that's what the mapping file is for)19:18
mordreddmsimard: oh - yah - need to write that patch, but yes, that idea makes this quite nice19:19
tobiashmordred: looks like the jobs used in 99converted are also sill missing?19:19
mordredtobiash: yes indeed19:19
dmsimardmordred: yeah I don't see the nodesets but I see the intent of using that19:19
mordredtobiash: the job+playbook content will come next19:19
dmsimardI think jeblair was against the idea though19:19
tobiashah, ok19:19
dmsimardor maybe it was about mixing label and name without requiring the other19:19
dmsimardI forget19:19
mordreddmsimard: that's a different thing- but the thing you just mentoined, defining some nodesets that match each of our labels - and also the old v2 multi-node labels19:20
mordredmakes for a nice transition19:20
mordredI'll get that patch up in just a sec19:20
dmsimardwfm19:20
mordredWOOT19:24
mordred"19:24
mordredAccepted python3.5 into zesty-proposed. The package will build now and19:24
mordredbe available at19:24
mordredhttps://launchpad.net/ubuntu/+source/python3.5/3.5.3-1ubuntu0~17.04.0 in19:24
mordreda few hours, and then in the -proposed repository.19:24
mordredclarkb, jeblair, pabelanger, Shrews, SpamapS: ^^19:24
jeblairyay!19:28
jeblairdmsimard, mordred: i am all for doing the convenience nodeset definitions (they should be in project-config)19:29
jeblairmordred: what things require ordereddict?  (i'm concerned about the stuff going into zuul/lib/yamlutil)19:37
mordredjeblair: if you don't use ordereddict the resulting file is illegible19:38
mordredjeblair: happy to just copy-pasta all of that into migrate.py though19:38
jeblairmordred: well, i was mostly looking at ordered_load, which would all be on the input side.  what makes it all the way through the system ordered?19:39
mordredjeblair: by illegible, I mean the source data has been maintain in alphabetical order, so scanning through and comparing source data to produced data with produced data being in an arbitrary order is hard to process19:39
pabelangermordred: cool19:40
jeblairmordred: oh, you mean: "projects: [nova]"19:40
mordredyah. - and also we have a generally accepte practice of having pipeline definitions look like project: name: foo template: - a - b check: ... etc19:41
jeblairmordred: we could maybe drop that and re-alphabetize it (for project_name in sorted(self.layout['projects'].keys()))19:41
openstackgerritDavid Shrewsbury proposed openstack-infra/zuul feature/zuulv3: Support IPv6 in the finger log streamer  https://review.openstack.org/50213719:42
pabelangermordred: jeblair: Shrews: SpamapS: so, before we dive into any we got another dead worker on ze01, lets upgraded to -proposed version of python, and see if we can reproduce19:42
jeblairmordred: okay.  i'm going to leave some comments about moving it into migrate and why.19:42
mordredjeblair: kk19:42
mordredjeblair: I'll do that in my next iteation19:42
jeblairmordred: okay, left comment; that's all i have after a quick look, will play with it now19:45
mordredjeblair: cool. also, please ignore the fact that the the output is missing a 'jobs' ... fixing that right now :)19:46
jeblairpabelanger: probably a good plan19:46
tobiashShrews: have thoughts on 50213719:51
Shrewstobiash: b/c 'localhost' did not work for me19:52
tobiashShrews: ok, that's an argument19:52
Shrewsdammit. now some sort of race....   "Streamed: Build ID c8ace3ccb68d44198522a428ac1440b3 not found"19:57
dmsimardWhy is there nova patches in the v3 check queue ?20:02
dmsimardAlso, it seems like there's a glitch in the UI where you can't expand the box to see queued jobs if none of the jobs have started yet ?20:03
clarkbo/ here for 2000UTC convening20:05
jlko/ same20:05
jeblairpabelanger, Shrews, tobiash, mordred, dmsimard: ping ^20:05
dmsimardpong20:05
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Add migration tool for v2 to v3 conversion  https://review.openstack.org/49180520:06
mordredjeblair: heya20:07
jeblairdmsimard: yeah, that's the thing that mordred wanted to fix with a ui patch (but it ended up making it harder for us to diagnose issues).  we should still think about ways to improve it, but probably later at this point.  in short, every change that *might* be enqueued *is* enqueued at least long enough for zuul to determine whether it should be.  that's when it shows up there, and it may stay there with no jobs until the merger comes back with ...20:07
jeblair... information on whach jobs it should run.20:07
jeblairs/whach/which/20:07
tobiasho/20:07
dmsimardjeblair: ah somehow I thought it was a merger queue thing so I wasn't entirely wrong20:07
dmsimardok, let's look at it later :)20:08
openstackgerritMerged openstack-infra/zuul-jobs master: Add UPPER_CONSTRAINTS_FILE file if it exists  https://review.openstack.org/50032020:08
jeblairhow about we start real quick by going over the etherpad from earlier: https://etherpad.openstack.org/p/zuulv3-pre-ptg20:08
jeblairwe're really close on devstack jobs.  i think there's a little more work to go into the v3 native job, and then we'll be ready to point people at it to start building new simple devstack jobs.20:09
jeblairit's probably close enough at this point not to consider it a migration blocker.20:10
jeblairthat sound right to folks?20:10
mordredagree20:10
clarkbis the devstack job running tempest?20:10
mordredwe can emit jobs for devstack-legacy for the most of them20:10
clarkb(I want to say the log I saaw did run tempest)20:11
jeblairclarkb: not yet20:11
jeblairclarkb: not the v3 native one (devstack-legacy, yes)20:11
clarkbah ok20:11
dmsimardjeblair: devstack doesn't worry me too much, it's the non-devstack things that do20:12
tobiashwas large scale dynamic layout reconfiguration tested already?20:12
dmsimardjeblair: especially deployment projects like puppet-openstack, openstack-ansible, kolla, tripleo..20:12
jeblairtobiash: no, only static configuration loading20:12
jeblairtobiash: and that with a null configuration20:14
jeblairokay, next on the list -- jobs with special slaves:20:14
jeblairhttps://etherpad.openstack.org/p/zuulv3-special-jobs20:14
jeblairmordred, pabelanger: what's left from that list?20:15
mordredjeblair: there's 4 release jobs20:15
mordredjeblair: npm-upload, jenkinsci-upload, mavencentral-upload and forge-upload20:16
jlkoh crap, I need to finish up the translation-update20:16
jlkupstream-translation-update20:16
jlkand while I'm there, I might be able to do propose-translation-update{suffix}20:16
mordred++ those are the others20:17
jeblairwhat's forge-upload?20:17
mordredjeblair: the stack in the middle just need to be handled by the migration script, I'll add those to the mapping file in the next patch20:17
jeblairpuppetforge20:18
jeblairthat could probably be more clear in the future :)20:18
jeblair(hardly the first thing called forge)20:19
mordredright?20:19
jeblairjenknsci and mavencentral are probably not critical to cutover.  npm, forge, and translations probably should be considered blockers?20:19
clarkbeven puppetforge is probably not critical20:19
pabelangerjeblair: I'm just retesting publishing to afs, and tagging a release, then good to mark them as done20:19
clarkbI don't think we consume any of our puppet modulse via puppet forge20:19
clarkbso question would be if puppet-openstack does20:20
jlkhttps://review.openstack.org/#/c/499845/ needs to be rebased by me, but could use reviews after.20:20
pabelangerjeblair: so, ready to work on next task20:20
jlkit's for propose-project-config-update20:20
jeblairclarkb: yeah, i was thinking of p-o20:20
dmsimardjeblair, clarkb: puppet-openstack used to publish to the forge20:20
mordredforge-upload is used by two projets20:20
dmsimardbut that was back in the day of hodgepoge PTL20:20
mordredpuppet-httpd-forge-upload and puppet-storyboard-forge-upload20:20
dmsimardI don't think we publish anymore20:20
clarkbmordred: if it is just those two then I don't think it is criticial to treat that job as a blocker20:21
jeblairoh, then i agree we can lower priority of that20:21
jeblairwhat about npm-upload?20:21
mordredit's used by openstack/tripleo-ui20:21
mordredand that's it20:21
jeblairlet's call that a blocker?20:22
mordredyah - it shouldn't be too hard to translate20:22
jeblairokay, so i put npm and translation-upload on the blocker list, and the rest on a new list of things to do right after cutover20:23
mordredit runs a script from slave_scripts, so the pattern used in the other proposal jobs can be copied easily20:23
jeblairokay, moving down the list (but saving migration script for last): migration docs20:24
jeblairi'm happy with where they are at the moment, but i'm sure we will want to update them with info about the migration script once we have it20:24
jeblairso let's call this done20:24
jeblairthe wget breakage is fixed20:25
jeblair"    docs jobs incorrecly publishing"20:25
jeblairpabelanger: i think you just fixed that?20:25
mordredyah20:25
mordredjeblair: for migration docs - perhaps we should keep an etherpad going at the PTG to jot down notes as folks ask us questions20:26
jeblairzuul-cloner shim -- Shrews that's in the base job now, right?20:26
Shrewsyep20:26
jeblairmordred: ++20:26
jeblair    configure-mirror parity20:26
jeblairthere are reviews there; sounds like we're almost done20:26
jeblairdmsimard: when those 3 are landed, are we good?20:27
jlkmordred: good idea.20:27
jeblair(note, i think as part of this that populating /etc/ci is being separated from configure mirror role, which makes a lot of sense)20:27
dmsimardjeblair: that should provide backwards compat for the /etc/ci/mirror_info.sh -- would need to compare or better, use, in a job that consumes it in order to tell if it's good20:28
clarkbdmsimard: I believe the dib cross build jobs make extensive use of it20:29
jeblairdmsimard: ack.  devstack-legacy uses one line from it20:29
dmsimardat first glance the reviews seem okay but I haven't had the chance to review them thoroughly yet20:29
dmsimardjeblair: I believe tripleo uses that file too, would need to double check20:29
jeblairok.  so still a blocker but likely we can wrap it up today20:30
jeblairnew servers20:30
jeblairpabelanger: ^ what's the status there?20:30
dmsimardare we keeping the same set of zuul mergers ? is there anything we are taking the opportunity to reinstall/install in xenial ?20:30
clarkbdmsimard: I think we need to update to xenial because of python3.520:31
pabelangerjeblair: I'd like to stand up nb01/nb02 shortly20:31
clarkbthe type annotations won't work as is in < 3.520:31
dmsimardso we have 8 new mergers to stand up ?20:31
pabelangerjeblair: still need to work on zuul-executor / zuul-mergers20:31
jeblairokay; we don't strictly need to do nb01/nb0220:32
clarkbfor the mergers I expect that we can rollback to v2 on xenial if we have to and just replace the existing set with xenial nodes20:32
jeblairshould we consider deferring that to save time+quota?20:32
jeblairclarkb: that sounds reasonable20:32
*** hashar has quit IRC20:32
pabelangerclarkb: oh, so just upgrade not new servers20:33
pabelangerI do think we need to do some puppet work for puppet-zuul on mergers20:33
clarkbthey are largely stateless (if you wave your hands around v2 needing to fetch from them) so easy enough to make rollback be redeploy on trusty as well20:34
jlkI assume once in production we'll have a better idea of how to balance executors vs pure mergers20:35
jeblairjlk: ++20:36
jeblairclarkb: what would you recommend we do?20:36
clarkbjeblair: re mergers? maybe split the current 8 in two and have 4 trusty for easy rollback and 4 xenial for v320:37
jeblairthat works for me.  we can also disable the merge-check pipeline temporarily to reduce load on them.20:37
jeblairpabelanger: sound good ^?20:37
jeblairif not, you can work it out later :)20:39
jeblairlast thing i just added to the list and can't believe i forgot -- logstash emitter20:39
jeblairi started some prep for this a couple weeks ago20:39
jeblairmy plan is to have a trusted post-playbook in base that submits a background job to the logstash gearman queue20:40
dmsimardemitter reminds me of the stuff like firehose/openstack-health, we don't need to do anything special for those ?20:40
jeblairwhat it emits will be compatible with what we currently emit20:40
jeblairbasically, it'll be a copy-pasta of the jenkins-log-client code into a post-playbook, skipping the zmq step20:41
mordred++20:41
jeblairdmsimard: largely to assist with this, zuul emits no firehose yet20:41
pabelangerjeblair: clarkb: sure, wfm20:41
dmsimardok.20:42
jeblairregarding openstack-health, i think the subunit processor is hooked up to the logstash system20:42
clarkbya its the same sort of job submission with different parameters20:42
clarkbshould be straightforward to solve both of those together20:42
jeblairok, so probably our post playbook will submit 2 jobs20:43
mordredjeblair: and the log-gearman-client.py, once zmq is done, only depends on gear, which is installed on the executor so should be piece of cake20:43
jeblairya20:43
jeblairi think this will not take long to write; i am unsure whether i can also get it safely merged into the base job today.20:44
jeblairi'll try to get as far as i can, but of the things so far, this seems most likely not to happen by EOD today20:45
jeblairokay, last thing -- migration script20:45
jeblairit runs!20:45
jeblairand outputs 19.5 kilolines of configuration!20:45
jlkegads20:45
jeblairwhich doesn't include the job definitions yet :)20:45
clarkbwhat is it if not job definitions?20:46
jeblair(that's actually 600 more than the current layout.yaml, interestingly enough)20:46
clarkboh right layout20:46
jeblairclarkb: project definitions (ie, what's in current layout.yaml)20:46
jlkcurrent makes more use of templates, no?20:46
jeblairclarkb: but the "jobs" section with all the crazy regex stuff is gone, folded into the project-pipeline definitions20:47
jeblairso it is slightly longer and *much* more readable20:47
jeblair      - legacy-grenade-dsvm-redis-zaqar:20:47
jeblair          voting: false20:47
jeblairfor instance ... can you guess whether that job is voting?  :)20:47
jeblairin v2 the answer is, no, you can not guess.20:47
mordred\o/20:48
jeblairanyway20:48
jeblairmordred: do you have incomplete code to do job configuration yet, or are we at the start of that?20:48
mordredjeblair: the existing 2.5 code20:49
mordredjeblair: so - no, I don'thave it pulled out or stiched in to the migration script yet - but that was going to be the starting place20:50
jeblairmordred: right, so we have that to do, as well as formulating the 'job:' configuration stanza20:50
jeblairand somewhere in there we need the role to do ZUUL_ variable compatability20:50
mordredyup20:51
jeblairthe change listed some other todos20:51
mordredthere isa todo at the top of the  migration script in the comments, fwiw20:51
mordredyah20:51
jeblairshared job queues20:51
jeblairalso, filters from the jobs section20:51
mordredyah - and continuing to add various mapping information to mapping.yaml20:51
jeblairmordred: oh, so i guess the current voting: settings just came from '-nv' ?20:51
mordredjeblair: yes. that is right20:52
*** olaph has quit IRC20:53
jeblairmordred: this is great progress and i can see it coming together, but i feel like we probably have a couple of days work yet until we get to what we'd consider final output stage.  would you concur?20:53
*** olaph has joined #zuul20:53
mordredjeblair: maybe? the migration/mapping stuff actually goes fairly quickly - but it's also 4pm on a friday here, so it's tough to say20:54
pabelangermordred: Hmm, release jobs are broken20:55
jeblairbased on the other stuff i still need to do, i probably personally can't pick up a migration script task until the plane flight on sunday20:55
pabelangermordred: I think mirror-workspace-git-repos is the issue20:55
pabelangeras it doesn't checkout tag?20:55
jeblairpabelanger: can we hold that for a minute?20:56
pabelangerjeblair: sure20:57
jeblairso here's a strawman -- personally, i don't think we're going to be in a position to cut over monday morning; i think it's more likely that we'll finish up the migration script and continue to flesh things out early in the week and we can perhaps attempt cutovers mid or late next week.20:58
mordredjeblair: yah- that's kind of what it feels like to me too - we're able to make extreme progress when we're all heads-down on it, so knocking out the final things while we're in the room will likely work quite well20:59
dmsimardFWIW I've put name on some risks at the bottom of the pad21:00
jlkthat's how it sounds to me as well21:00
dmsimardAnd I also believe that doing this *tomorrow evening* is risky, everyone will be travelling21:00
jlkI'd rather see us work hard on getting things right, rather than doing the cut over badly and firefighting frantically for the few days after21:00
mordred++21:00
jeblairdmsimard: ya thanks that helps a lot21:00
dmsimardjeblair: I'm trying to de-risk things that are uncharted territory right now21:01
dmsimardbut my time is short, won't have time to hack on things tonight or tomorrow as I prep the family for the time I won't be around21:01
clarkbya and juggling knowing who is around when is hard when its all travel time21:02
mordredat the ptg we'll also be in a decent position to do some targetted trial runs of some of the unknowns with folks - dmsimard mentioned non-devstack integratoin tests as an example21:02
pabelangerI'll be working late today21:02
pabelangerbut travel in the morning to PTG, but around all after Saturday / Sunday21:02
dmsimardwhat mordred said, I would very much like to de-risk the migration by opting in higher risk projects first gradually, somehow21:02
jeblairdmsimard: we have to do the cutover all at once, but we *can* start running check jobs on any project we want21:03
mordredwe're pretty sure that our base jobs and ZUUL_ vars role should put things in the right state for those, but won't know until we try one21:03
dmsimardjeblair: yes, sure, I'd basically like to try and run real "migrated" v3 jobs on projects ahead of the cutover21:03
mordredluckily we can generate some, put a few into a .zuul.yaml and run a check job on them21:04
mordredbut for many of these projects we don't know how to assess if they are behaving properly, or if they aren't, what isn't happening that should21:04
clarkbmaybe instead of attempting fisrt cutover on saturday we use daytime sunday to prep and get check jobs running in more places?21:05
dmsimardmordred: I have experience and/or contacts in all the deployment projects to help21:05
clarkbI think a lot of us arrive on saturday evening ish time so potentially sunday is good sit down and crank things out time21:05
mordredclarkb: agree. fungi and I will be in tc/board meeting, but I usually laptop-hack in those anyway :)21:06
dmsimardYeah I'll be there sunday21:06
dmsimardtristanC too21:06
jeblairclarkb: i'm not expecting us to be in a position to even do trial cutovers saturday21:06
mordredI get in saturday evening21:06
dmsimardmordred: from China ?21:06
mordreddmsimard: no - woundup not going to china21:06
dmsimardOh, ok, no jet lag then21:06
mordredjeblair: yah- I think clarkb was suggesting we not try to do trial cutovers on saturday21:07
jeblairso i think we just keep plugging away at this and whenever the migration script gives us output we can use, we start throwing some of it at zuulv3.  but i don't think we should plan on any 'scheduled' events this weekend.21:07
dmsimardAgreed21:07
mordredagree21:08
dmsimardLet's make as much progress to de-risk the migration as much as we can and see if it's doable thursday ?21:08
dmsimardCause, you know, read only friday21:08
jeblairthe last question i have is about notification/communication.  the last thing we said about this to the dev community was 1 month ago, and we said we'd (probably) cutover monday.  normally, we would have sent some reminder announcements, but the situation hadn't really gotten any clearer.  do we want to send something out now?  if so, what?21:09
jeblairor do we want to cutover and say "surprise!"21:09
mordredI definitely don't think we should attempt to throw the switch on friday :)21:09
mordredjeblair: heh21:09
mordredI think sending out an update is a good idea21:09
dmsimardAn update to say it's not happening either saturday or monday is the minimum21:10
mordredsomething along the lines of "we're close, but we're not satisfied with the migration/cutover and are going to work on it for a few more days while we're in denver. blah blah safe than sorry see you soon kthxbai"21:10
clarkbmordred: ++21:10
jeblairmaybe also include "zuul might start leaving info on patches before then.  we'll let you know when it happens.  read this page: https://docs.openstack.org/infra/manual/zuulv3.html  attend the ptg session on monday."21:11
clarkbespecially since I know a lot of people are wondering when they will get first crack at zuulv321:11
jlkWe should probably send something that outlines what we've discussed today, and set expectations that there is high chance of a cutover next week.21:11
mordredjeblair: there is also a chunk of time monday scheduled in a room for some value of "us" to talk to some value of "people" about v3 and what it means21:11
mordredjlk, clarkb: ++21:11
jlkor what mordred said21:11
jeblairmordred: you want to draft that up?21:11
mordredjeblair: yah - I can take a swing at it real quick21:11
dmsimardjeblair: yes, something about potentially getting non-voting reviews from zuul as we dry run things could be good21:11
mordred"if you start seeing comments from zuul, don't freak out - unless you want to"21:12
jeblairokay, i think we have a Plan(tm)21:12
jeblairanything else?21:12
Shrewsand possibly note that we are still holding "office hours" to discuss v3 things?21:13
jlkWe've got budget for the cut-over celebration right?21:13
jeblairShrews: ++21:13
mordredjlk: there's always budget for that21:13
jlkhopefully it's before Thursday evening.21:13
jeblairjlk, mordred: free booze! *now* i'm motivated!21:13
Shrewsfree booze AND free steak is better motivation. just sayin'...21:14
jlkI'm happy to donate my entire per diem for that day to the cause21:14
mordredwe may also want to ponder, before we show up, that there are a bunch of future/post-ptg zuulv3 related things that people are going to want to discuss while we're in denver21:14
jeblairmordred: yes, and unfortunately, we still have a self-generated backlog of weeks/months21:15
mordredand think about how to allocate some amount of time to that so people don't explode, but not so much that we stall on the last mile of rollout21:15
pabelangerSo, is there an updated list of things we need to focus on? Or is https://etherpad.openstack.org/p/zuulv3-pre-ptg been updated21:15
mordredit's also possible the time for that will just be at the bar in the evening :)21:15
jeblairmordred: yeah.  if we can collect use cases, etc, that will be good.  if we can also convey it'll probably take a bit to get around to them on account of we still need to do some basic things that would be great21:16
jeblairpabelanger: that's current21:17
jlkprobably need to make clear that cutover != v 3.0 release, right?21:17
jeblairjlk: indeed21:17
jlkthat we still have some things to finish before calling it 3.021:17
pabelangerokay, I've added my name to some tasks21:18
pabelangerbut release jobs appear broken, so making note21:18
Shrews3.0a   alpha release21:19
jeblairShrews: i don't think we're even there yet21:19
jlk3.lol21:19
jeblairmore like that yeah21:19
Shrewsi will call it 3p0 to myself, b/c i find humor in that21:20
* Shrews must do dinner things now21:20
pabelangerI'm relocating as ansiblefest is over, going to get food then get back online21:20
jeblairit's sort of like the opposite of tex version numbers.  we started at 3.lololololololol, and we're at 3.lolol, almost down to 3.lol.  :)21:21
jlksoon enough, it'll be 3.yolo21:21
jeblairfriday21:21
jlk(that's when the drinking begins)21:21
mordredhah21:21
* jeblair puts away flask21:21
jeblairokay, i'm going to give my chair a break, then back to logstash things21:22
mordredwoot!21:22
mordredjeblair: before you do logstash things ...21:22
jeblairthanks everyone!21:22
jlkcheers. I'm out for a bit too, dog needs a walk21:22
mordredjeblair: I may have another zuul scheduler issue for you21:22
jeblairtoo late already stood up21:22
mordredjeblair: https://review.openstack.org/#/c/500365/ is not running any jobs21:23
mordredjeblair: nor is its ancestor21:23
mordredjeblair: the job content itself isn't a big deal - but zuul ignoring the uploads and the rechecks is potentially worrisome - I looked briefly and didn't see anyting immediately21:23
jeblairmordred: ack; i'll check logs in a bit21:24
mordredjeblair, clarkb, jlk, pabelanger, Shrews, dmsimard: https://etherpad.openstack.org/p/6NTJxufjNU how's that look?21:37
pabelangerlooking21:38
mordredjlk: if you get bored and run out of other things to do after the translation proposal things, https://review.openstack.org/502185 adds the secret needed for the npm upload job, so anybody can take the ball and run with that one too21:39
pabelangermordred: minor change, looks good21:41
mordredpabelanger: ++21:41
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Update roadmap in the README  https://review.openstack.org/50218821:52
mordredpabelanger: so what's up with the release jobs?21:57
pabelangermordred: let me find log, but I don't think we were on right tag, we ran against master21:58
mordrednod21:58
pabelangermordred: http://logs.openstack.org/da/0a0162e080f09d5593effd4fcb837c3554d014da/release/release-openstack-python/44c7a5b/job-output.txt.gz#_2017-09-08_21_04_03_78062021:59
mordredpabelanger: also - I was just scanning through the project-template section of zuul/layout.yaml and I think there are a few more templates we could define based on content we have now - and a couple that we might need to add (there is an openstack-server-release-jobs that bulds and uploads tarball like normal but does not publish to pypi)22:00
mordredpabelanger: cool, looking22:00
pabelangermordred: ya, happy to look and get the templates working22:01
mordredpabelanger: if you happen to look at the project-template list, adding stuff as we go to that mapping file will, I think, help us not forget things (I know I've already lost track of a few of the things we have myself :) )22:03
mordredpabelanger: that looks like something EXTRA weird with Determine Local Head22:04
pabelangermordred: let me look at your migration script first22:04
pabelangerso to get up to speed22:04
jeblairmordred: msg looks good!22:06
jeblairpabelanger: if you can make that release job run with keep enabled, that would be great.  we can look at the local repo state on the executor.22:08
mordredyah ... I can't find anything that would cause that output22:09
mordredalthough it does make me want to change something in the base job ...22:09
mordredoh. hah. I already wrote a change for this somewhere22:11
dmsimardWould it be worthwhile to plug jeblair's Barcelona Zuul v3 talk ?22:17
dmsimardI re-watched it recently to refresh my memory on things :D22:17
dmsimardOtherwise message lgtm22:18
jeblairdmsimard: heh, i feel like it's a bit dated and lacks practical information; but i'll let you/others decide if it's useful.22:18
dmsimardI think it's easy for us to take a lot of things as granted or "common sense", the talk goes into a bit of the background for people unfamiliar with what even zuul v2 would be.. but I don't have a strong opinion, just thinking out loud22:20
*** olaph1 has joined #zuul22:21
mordredjeblair, pabelanger: https://review.openstack.org/#/c/501242/ would be nice to land - also would let us see the inventory we produced for the borked release job22:22
*** olaph has quit IRC22:23
mordredjeblair, dmsimard I added a PS to the end with a link to that talk and other things  - I can't decide if I like including that section or not22:28
mordredclarkb, jlk, pabelanger: ^^ thoughts?22:29
pabelangerjeblair: mordred: sure, I can do --keep now22:29
dmsimardmordred: worst that can happen is that people don't watch/read/care, best case we get more people on the same wavelength22:31
dmsimardDoesn't sound like a bad tradeoff22:31
jeblairmordred: we don't have a great way to share a module between two (roles, playbooks, etc) do we?22:34
dmsimardjeblair: openshift-ansible has a solution for that22:35
jeblairmaybe symlinks?  otherwise, we'd have to have a role for it, then include_role from 2 other roles?22:36
dmsimardOne sec22:36
jeblairmaybe we should handle /library like we handle /roles?22:36
jeblairdmsimard: thx22:36
dmsimardjeblair: https://github.com/openshift/openshift-ansible/tree/master/roles/lib_openshift22:36
dmsimardjeblair: https://github.com/openshift/openshift-ansible/tree/master/roles/lib_utils22:36
dmsimardThey're basically roles with more or less just a library folder with modules/things in them22:37
mordreddmsimard: dont you have to use a role before the module in it is accesible?22:38
dmsimardmordred: meta dependencies22:38
dmsimardmeta dependencies will *run* the role.. but if there's nothing to run.. :)22:38
mordreddmsimard: neither of those roles declar depends22:38
mordredah. yah22:38
dmsimardmordred: https://github.com/openshift/openshift-ansible/blob/3409e6db205b6b24914e16c62972de50071f4051/roles/docker/meta/main.yml#L1322:38
mordredjeblair: so yah - make a role that just has library in it, then put dependencies: - library-role into meta/main.yaml of the roles that need it22:39
dmsimardopenshift-ansible does a lot of really cool things, I've sent a few patches their way. awesome stuff.22:39
mordredjeblair: but - that's also basically what you were saying with include_role - same thing22:39
jeblaircool (the library role will be the "submit a gearman job" and the depending roles will be "submit a logstash job" and "submit a subunit job")22:39
pabelangerjeblair: mordred 76fa1c4a60314afe843e0c5cf8c96803 on ze01.o.o was the release job22:40
jeblairmordred: yeah, but then i can use the module directly in a task?22:40
dmsimardmordred, jeblair: this one is particularly cool, it's more or less the equivalent of our validate-host but on steroids https://github.com/openshift/openshift-ansible/tree/3409e6db205b6b24914e16c62972de50071f4051/roles/openshift_health_checker22:40
mordredjeblair: root@ze01:/var/lib/zuul/builds/76fa1c4a60314afe843e0c5cf8c96803/work/src/git.openstack.org/openstack-dev/sandbox# git status22:40
mordredNot currently on any branch.22:40
mordredjeblair: yes22:40
jeblairk.  i'll do a local mockup first :)22:41
dmsimardAnd then they even have a callback that prints what the actual failures were: https://github.com/openshift/openshift-ansible/blob/3409e6db205b6b24914e16c62972de50071f4051/roles/openshift_health_checker/callback_plugins/zz_failure_summary.py22:41
jeblairmordred: huh, i thought i checked tags22:41
mordredjeblair: well- status there is not showing what it does for me locally22:42
jeblairthat'll do it22:42
jeblairgit describe22:42
jeblair0.0.2322:42
dmsimardjeblair: happy to review your "zuul-lib" role when you got something going, add me as reviewer22:42
dmsimard(btw we might want to move the module from validate-host there)22:43
mordredjeblair: what's the state difference - locally if I do "git checkout 0.0.23"22:43
mordredgit branch shows me:22:43
mordred* (HEAD detached at 0.0.22)22:43
jeblairyeah me too22:44
jeblair.git/HEAD is the same22:44
jeblairhuh, git status consults .git/logs/HEAD22:46
mordredjeblair: so is this becaus it's cloned directly from /var/lib/zuul/executor-git/git.openstack.org/openstack-dev/sandbox to that ref rather than cloned and then checked out?22:47
jeblairmordred: it should be cloned and checked out; i think executor is checking it out differently than git cli22:48
mordred\o/22:49
mordredjeblair: so - fwiw, for tag jobs we do have zuul.tag for the main project22:50
dmsimardoh, hey, we got a devstack-legacy pass on centos7 but a failure on suse so it's not too horrible22:51
dmsimardis devstack on debian a thing 622:52
jeblairmordred: i think the only viable option is for the executor to make things correct in its work root, and to synchronize that exactly to the remote nodes.  i don't want anything bypassing that process.22:52
dmsimard?22:52
mordredjeblair: and the normal "find local branch" logic works fine for branches that aren't the one that's tagged22:52
mordredjeblair: nod22:52
pabelangerdmsimard: no22:53
pabelangerdmsimard: centos should pass, I am not sure about opensuse. might want to ask dirk in openstack-infra22:53
dmsimardpabelanger: it should, we have it in devstack-gate22:53
dmsimardpabelanger: I'll try and see if I can find anything obvious22:54
dmsimardin the meantime re-checking a cleaner patch with f26 thrown in22:54
clarkbmordred: +2 on 50124222:54
clarkbdidn't approve because I am very distracted by home things before having to get on a plane22:54
pabelangerdmsimard: systemd failed to start peakmem_tracker22:54
pabelangerno idea why22:55
dmsimardpabelanger: what, on the suse job ?22:55
pabelangerdmsimard: yup22:55
pabelangerhttp://logs.openstack.org/47/502147/2/check/devstack-legacy-opensuse-423-tempest-dsvm-neutron-full/1d87625/logs/devstacklog.txt.gz#_2017-09-08_20_39_49_06722:55
jeblairthis is what the executor does: https://etherpad.openstack.org/p/BPndl6rF4722:56
jeblairyou can use that to reproduce the weird state22:56
dmsimardpabelanger: "['pmap', '-XX', '1']' returned non-zero exit status 1"22:57
openstackgerritMerged openstack-infra/zuul-jobs master: Make validate-host read from site-variables  https://review.openstack.org/50059222:57
mordredclarkb: thanks!22:57
pabelangerdmsimard: this could be devstack bug, I think they did screen removal recently22:58
pabelangerdmsimard: I'd check if zuulv2.5 jobs are working22:58
dmsimardpabelanger: yeah doesn't look like -X is even an arg for pmap http://www.unix.com/man-page/suse/1/pmap/22:58
jeblairmordred, pabelanger: maybe we want the executor to do: r.git.checkout('0.0.22')22:59
mordredjeblair: maybe because of reset --hard -- to do that after setting the head reference?22:59
mordredjeblair: yah - I agree - but it seems we only really need to do that for tags?22:59
pabelangerdmsimard: http://logs.openstack.org/47/502147/2/check/gate-tempest-dsvm-neutron-full-opensuse-423-nv/a30d9c4/logs/devstacklog.txt.gz#_2017-09-08_21_00_40_012 failing on zuulv2.5.22:59
mordredjeblair: like, we certainly don't want to try to checkout speculative refs :)22:59
jeblairmordred: we're already using a different method for branches23:00
dmsimardpabelanger: yeah I found https://review.openstack.org/#/c/496301/ which ran today23:00
mordredjeblair: cool23:00
jeblairmordred: speculative refs are on branches now, so we just checkout the branch in that case23:00
jeblair(that's one of the super subtle awesome things about v3 :)23:00
dmsimardpabelanger: I'll ping some suse folks on -infra23:00
pabelangerjeblair: mordred: will defer to your expertise23:00
pabelangerdmsimard: +123:00
mordredjeblair: so yah - in that case I think r.git.checkout is gonna be more better23:01
jeblairoh neat, we can even 'git checkout refs/tags/0.0.22'23:01
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Use 'git checkout' when checking out a tag  https://review.openstack.org/50219523:07
jeblairmordred, pabelanger, clarkb: ^  we'll need to merge that and restart executors then try the release job again23:07
pabelanger+223:10
mordredjeblair: I like that idea that we can get branches and tags working the same23:12
jeblairmordred: yeah, if it were any other day, i would have gone ahead and done that :)23:13
mordredjeblair: ++23:15
pabelangerze01 still has --keep, I've disabled it on ze02 already23:29
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Use 'git checkout' when checking out a tag  https://review.openstack.org/50219523:29
mordredjeblair, pabelanger: the last couple of times I've restarted the scheduler it has seemed like I needed to restart the executors too23:32
mordredjeblair, pabelanger: is that a thing? like, do we need to restart the executors after a full scheduler restart?23:32
mordredor am I just too impatient23:32
jlkmordred: might be impatient. I haven't seen that in my testing locally23:41
jlkmordred: what I _have_ seen from time to time is that the way the VMs work that existing nodepool VMs wouldn't necessary come back, because they check in at boot time, and since it's not boot time, they don't check in23:41
mordredjlk: cool. I'll go with impatient :)23:43
jlkmordred: the etherpad looks good, for a mail to go out.23:43
mordredjeblair: figured out why the shade changes weren't being tested23:43
jlkwhat time on Monday is the zuul session?23:43
mordredjeblair: there was a dependency loop23:43
mordredjlk: /me looks scared and hides23:43
mordredjlk: it's sometime in the afternoon I think ... not really sure23:46
jeblairmordred, jlk  2pm vail23:46
jeblairhttps://ethercalc.openstack.org/Queens-PTG-Discussion-Rooms23:46
jlkah I'll probably miss it. I land at like 1pm pacific time.23:46
jlkso I land right when that session is starting23:47
jeblairmordred: updated your email etherpad to include that23:48
jeblairmordred: what made you think you should restart the executors?23:48
mordredjeblair: jobs weren't executing23:48
mordredjeblair: but - I think last time it was early in the morning and I wasn't coffeed enough - so I think it should stay in the anecdote bucket23:49
pabelangermordred: I've had success with just scheduler restarts23:50
mordredok. good to know23:50
pabelangermordred: say when to try tagging again23:51
jlkI do it a ton in docker so that I can alter code and just restart scheduler23:51
mordredpabelanger: I have not restarted anything, but the change has landed23:51
jeblairpabelanger: i don't see the fix commit on ze01 yet.  we can manually update though if you're ready23:51
pabelangersure, I am ready if you want to manually update23:52
jlkhrm, I'm looking at upstream-translation-update, trying to find where it sets the node to proposal23:53
jeblairthere were 3 zuul-executors running on ze01 again23:53
jlkI see it runs a proposal-slave-cleanup though23:53
jlkoh I see it, yaml buried the lede there.23:54
pabelangerjeblair: what would cause that?23:55
jeblairdmsimard: i may have aborted some devstack jobs with a restart just now23:55
jeblairpabelanger: someone restarting it incorrectly?23:55
pabelangerok23:55
jeblairsystemd not doing the one thing it's supposed to be good at?23:55
dmsimardjeblair: ok let me know when I can do a recheck23:55
jeblairdmsimard: you're good23:55
pabelangerjeblair: does that drop back to 1 process when zuulv3 mergers come online?23:55
jeblairpabelanger: you can tag now23:56
pabelangerk, tagging23:56
jeblairpabelanger: well, i mean, we're never supposed to have 3 running.23:56
jeblairis not related to mergers23:56
pabelangerk23:56
pabelangersandbox 0.0.24 tagged23:57
jeblairmordred: it looks like you were the last to restart ze01, can you tell me exactly what you did?23:57
mordredjeblair: yah - I did a service zuul-executor stop - then I _believe_ I checked the process list for any python processes (which I do because I have absolutely no trust in our init scripts currently)23:58
pabelanger:(23:58
pabelanger2017-09-08 23:57:55.899000 | ubuntu-xenial -> localhost | ERROR: AnsibleUndefinedVariable: 'zuul_traceroute_host' is undefined23:58
mordredjeblair: I tend to not issue any starts until I've seen that there are no processes remaining23:58
mordredpabelanger: oh good!23:58
pabelangerlet me see why23:59
mordredI just landed the 'use site_variables' change23:59
jeblairmordred: that sounds like a good procedure.  maybe a straggler just escaped your attention.23:59
dmsimardbtw thanks #zuul for being patient with me while I learned zuul v3 and devstack/devstack-gate :D23:59
mordredjeblair: yah. I really wish it didn't have to be a procedue23:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!