Monday, 2017-11-20

tristanCSpamapS: that would be definitelly nice to have, do you think nodepool could/should do that (manage speculative image)?00:31
tristanCSpamapS: i'm affraid the ansible kubectl connection plugin only seems to work with the raw/shell modules...00:35
tristanCSpamapS: perhaps another workflow would be to let zuul-executor build and push images to k8s, and then write the tests as part of a Job: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/00:37
tristanCbut then again, we'll lose the capability to write test definition in Ansible00:39
tristanCSpamapS: and what about lxd, why would it be a better choice?00:41
clarkbI think you'd build and push the image as a pre step00:56
clarkbsince it will have access to the git repos there00:56
tristanCclarkb: maybe zuul could pass the speculative git repos as part of the requestNodes() call :)00:57
clarkbtristanC: one of the redesign goals of v3 was to avoid needing to serve git repos though00:58
clarkbI think ideally anything requiring speculatively merges code runs on the executor aide of things00:59
tristanCwell the test instances do require the speculatively merged code... so instead of building the environment once with nodepool or a "containerpool" service, we end up pushing the merged code to each environment01:02
tristanCthinking about a kernel ci, it would be nice to have glance image with the speculative kernel so that it could run test in parallel01:04
clarkbfor that I think you just install it and reboot as part of job?01:05
clarkbpotentially with an earlier job building it01:05
clarkbespecially since the kernel may not boot at all01:06
clarkbso youd want that job side to record it properly01:06
tristanCclarkb: yes, that's exactly what jeblair suggested :)01:06
clarkbhaving built kerbels that nuked grub you definitely want a away to account for that01:07
tristanCwe could even get a post-nova-console.yml play01:07
tristanCclarkb: i am mostly interrested in the kernsec/bwrap/oci stack test, for which a aki/ari/ami would work fine01:10
tristanCi mean, as a thought experiment, it's an interresting workflow to address01:12
clarkbya though I dont think it needs to be complicated, just install kernel and reboot01:13
clarkbif it doesnt come back treat it as a failure. if it comes back test it01:13
tristanCyes, thanks for that suggestion, that would easily work indeed01:14
SpamapStristanC: nodepool would build the image that the pre job FROM's01:40
openstackgerritRui Chen proposed openstack-infra/nodepool feature/zuulv3: Fix nodepool cmd TypeError when no arguemnts  https://review.openstack.org/51958201:44
SpamapStristanC: and lxd is meant to run containers that act like VMs.02:02
tristanCSpamapS: iiuc, lxd doesn't provide an api, so you would still have to manage host instances (which is what the oci driver does)02:10
tristanCoh my bad, it does have a rest api02:17
tristanCSpamapS: anyway i'm not convinced lxd is a better choice over oci/k8s, at least not for tests like tox, go test, rpmbuild... Those seems to work fine with the containerized sshd trick02:25
tristanCit seems like more comparable to openstack instance, e.g. when you need the whole system stack02:28
tristanCwell i never used lxd so i may be missing something. anyway it looks like a good nodepool driver candidate02:43
*** threestrands has quit IRC06:03
*** bhavik has joined #zuul06:47
openstackgerritIan Wienand proposed openstack-infra/zuul-jobs master: Ignore missing .tox/env/logs directorys for copy  https://review.openstack.org/52143606:52
openstackgerritIan Wienand proposed openstack-infra/zuul-jobs master: Ignore missing .tox/env/logs directories for copy  https://review.openstack.org/52143606:55
*** xinliang has quit IRC07:06
*** xinliang has joined #zuul07:18
*** hashar has joined #zuul08:33
*** bhavik has quit IRC09:53
*** electrofelix has joined #zuul09:55
*** bhavik has joined #zuul10:03
*** bhavik1 has joined #zuul10:37
*** bhavik has quit IRC10:39
*** bhavik1 is now known as bhavik10:39
*** bhavik has quit IRC10:55
openstackgerritMerged openstack-infra/zuul-jobs master: Ignore missing .tox/env/logs directories for copy  https://review.openstack.org/52143610:56
*** kmalloc has joined #zuul11:01
*** openstackgerrit has quit IRC11:32
*** jkilpatr has quit IRC11:52
*** jkilpatr has joined #zuul12:24
rcarrillocruzhttp://38.145.34.35/logs/ansible-networking/check/github.com/rcarrillocruz-org/ansible-fork/2/1e759d67fe084e1297cab4fba9440cd1/run-openvswitch-integration-tests/logs/job-output.json12:32
rcarrillocruz\o/12:32
rcarrillocruzfirst job run of openvswitch ansible modules on my testing zuulv3 server12:33
rcarrillocruzhooked with GH12:33
SpamapSrcarrillocruz: congrats!13:52
SpamapStristanC: also sorry for getting all engineer-crit on your thing. I think its really cool to have a k8s option for zuul. :)13:53
tristanCSpamapS: heh, no offense taken :) i'm very new to docker/k8s and your suggestions are much appreciated14:01
tristanCsorry if i sounded defensive14:02
SpamapSnot at all14:18
SpamapSjust woke up and realized I had only been critical14:18
SpamapSAnd really I just want to boil it down to the thinnest thing possible.14:19
SpamapSsshd is certainly thinner than a whole VM :)14:19
*** openstackgerrit has joined #zuul14:22
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114214:22
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114214:23
leifmadsenmordred: jeblair: fyi pabelanger helped me last Thursday/Friday, and I was able to get far enough to run a "hello world"!14:36
leifmadsenso I'll be circling back around in the next week or two, and building out a cleaned up set of docs/notes for a "Zuul From Scratch" (I'm thinking about avoiding use of "quickstart" in the docs, because it's not really very quick or light :))14:36
leifmadsenonce I get that far, I'll propose some changes to the existing feature/zuulv3 branch14:37
mordredleifmadsen: \o/15:00
leifmadsenindeed heh15:02
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Make build-python-release job  https://review.openstack.org/51392515:10
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Remove old python-sdist job  https://review.openstack.org/51392615:10
rcarrillocruzfolks, not finding much docs about how to stream console jobs on zuul. What is it needed? web-console up and zuul_console spawned on each job on the executor, what else15:15
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Half-Revert "Revert "Add ensure-reno and ensure-babel roles""  https://review.openstack.org/52155815:24
tristanCrcarrillocruz: what is web-console? you meant zuul-web right?15:24
*** bhavik1 has joined #zuul15:24
rcarrillocruzerm, yeah15:25
rcarrillocruzzuul-web sorry15:25
tristanCnot sure what's wrong with your setup, but that should be enough. here is an apache conf to sit in front of zuul-web: https://review.openstack.org/#/c/505452/9/zuul/web/static/README15:26
tristanCwell, without that patch cherry-picked, change it to ProxyPassMatch /console-stream ws://localhost:9000/console-stream nocanon retry=015:27
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Half-Revert "Revert "Add ensure-reno and ensure-babel roles""  https://review.openstack.org/52155815:27
*** bhavik1 has quit IRC15:39
rcarrillocruzhmm, k, was mixing app webapp *and* zuul-web15:41
*** jkilpatr has quit IRC16:13
*** jkilpatr has joined #zuul16:29
*** bhavik1 has joined #zuul16:39
*** hashar is now known as hasharAway16:45
*** bhavik1 has quit IRC16:50
*** bhavik1 has joined #zuul16:50
tobiashianw: I left an answer on https://review.openstack.org/#/c/52085516:53
openstackgerritMerged openstack-infra/nodepool feature/zuulv3: Fix nodepool cmd TypeError when no arguemnts  https://review.openstack.org/51958216:56
*** bhavik1 has quit IRC16:59
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114217:01
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Update fetch sphinx output to use sphinx vars  https://review.openstack.org/52159017:01
mordredjeblair: if you didn't see in the scrollback, electrofelix was asking questions in #openstack-infra about the ability to push merges from zuul and how adding that might or might not relate to merger/executor interactions ... I thought that was a good jeblair conversation :)17:08
*** jkilpatr has quit IRC17:09
*** jkilpatr has joined #zuul17:10
mordredjeblair: also - important to note - there is what I think is an issue with override-checkout not working as expected17:15
mordredjeblair: we disabled tox-py35-on-zuul from zuul-jobs as it was bombing out due to zuul master being checked out: https://review.openstack.org/#/c/521096/ and I failed at figuring out what was happening17:16
jeblairmordred: does that affect override-branch too?  (or have we removed it?)17:17
electrofelixmordred: thanks, I keep forgetting this is the best room for zuul discussions, I'll break my habbit at some point17:17
mordredjeblair: well, first stab at fixing was "swich from override-branch to override-checkout"17:17
mordredjeblair: http://logs.openstack.org/42/521142/1/check/tox-py35-on-zuul/81bff30/ is an log from a failure17:18
mordredjeblair: but switching from one to the other did not have any impact17:18
jeblairelectrofelix: i caught up on the earlier infra discussion -- but what's the issue you're trying to solve?17:18
jeblair(i think i need just a little more context)17:18
electrofelixWe build artifacts during the gate that we don't want to rebuild afterwards, because what we built is what actually passed tests17:19
jeblairmordred: ok, so it seems to affect both.  did this use to work and something changed?17:19
electrofelixsome teams like to attach the metadata of the git repos to those artifacts17:19
jeblairelectrofelix: okay, so this may actually touch on two related issues:17:19
electrofelixbut the commit SHA1 won't line necessarily up subsequently, because it corresponds to the proposed merge by zuul is just a test rather than the actual merge17:20
electrofelixjeblair: need to switch from office to home, so maybe hold off discussion until tomorrow and I'll ping you again about it17:21
jeblairelectrofelix: 1) using the result of the initial zuul merger calculation in the executors (rather than repeating the action in the executors), as well as 2) pushing the result of the initial zuul merger calculation to the target branch rather than having gerrit/github perform the merge.17:21
jeblairelectrofelix: sounds good17:21
*** electrofelix has quit IRC17:21
mordredjeblair: yes, I believe it worked at some point in the past as that was our initial thing to make sure zuul-jobs changes didn't break unittest jobs17:25
mordredjeblair: THAT SAID - it's possible we never actually validated that it was running zuulv3's unittests and not zuulv2s'17:25
mordredjeblair: I'm also not sure, even if it WAS running zuul v2's unittests - that that would break due to the patch in question17:25
jeblairmordred: yeah, i guess that sort of thing could slip by (were we running py3 tests on v2?)17:26
mordredjeblair: but there were too many variables in the air and I was more focused on helping get releasenotes jobs fixed, so I didn't diagnose deeply17:26
jeblairmordred: thanks, i'll try to dig in shortly17:27
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114217:41
jeblairmordred: there's a good chance we forgot to restart the executors after adding support for override-checkout.  that may explain why override-checkout isn't working, but it would not explain why override-branch wasn't.17:46
jeblair(override-checkout change merged nov 1, a random executor start time is oct 31)17:47
jeblairShrews: in shutting down openstack's executors, i see that 2 of them are stuck at the following:18:05
jeblairsendto(7, "59.676215 | wheel-mirror-ubuntu-"..., 4096, 0, NULL, 018:05
jeblairzuul-exec 9708 zuul    7u  IPv6 3547817819       0t0        TCP ze04.openstack.org:finger->zuulv3.openstack.org:34988 (ESTABLISHED)18:06
jeblairShrews: it looks like it's stuck transmitting data to the multiplexer18:06
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add support for warning-is-error to sphinx role  https://review.openstack.org/52161818:07
jeblairShrews: these were last restarted on nov 10 and nov 15, so they should be running the old process code18:07
jeblair(though i suspect this would be the same in either case -- process or threading)18:07
mordredtristanC: what version of angular is required? (I just realized that we didn't add an entry into etc/status/fetch-dependencies.sh18:13
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: DNM: test tox-py35-on-zuul  https://review.openstack.org/52162318:14
mordreddmsimard: do you perhaps know the answer ^^ ?18:15
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Add angular to fetch-dependencies.sh  https://review.openstack.org/52162518:23
mordredtristanC, dmsimard: ^^ based on looking at xstatic packages and also storyboard-webclient depends I'm guessing that's the version you were using18:25
dmsimardThere's no angular in ARA18:30
dmsimardMissing context, let me read back..18:30
dmsimardlooking at https://softwarefactory-project.io/r/gitweb?p=scl/zuul-distgit.git;a=blob;f=zuul.spec;h=bfaaa32b7bb1b61ed84fa5695f665f05bd9af0b8;hb=HEAD there's no obvious version for it.. seems like it's being pulled from either in-tree or elsewhere18:35
dmsimardThis is the angular.js file from a test deployment http://paste.openstack.org/raw/626839/ .. there's no version header in it. Something about 1.3.7 but looks about error handling instead.18:41
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add support for warning-is-error to sphinx role  https://review.openstack.org/52161818:53
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114219:09
dmsimardmordred, SpamapS: fyi along the lines of our discussion last week, I'm starting a thread on openstack-dev around leveraging zuul v3 jobs/roles/playbooks outside OpenStack -- it's focused around migrating some "downstream" TripleO jobs to Zuul v3 but I feel like the discussion will be worthwhile to see what we can do, what we can't, what works and what doesn't19:20
dmsimardI feel the need to mention it since you might filter out [TripleO] threads :)19:20
mordreddmsimard: sweet19:22
jeblairdmsimard: sounds cool :)19:23
dmsimardjeblair: oh, we started a pad to document issues with running zuul-jobs outside the gate: https://etherpad.openstack.org/p/downstream-zuul-jobs-issues19:24
dmsimardthere's not much there yet, we only scratched the surface19:24
dmsimardWe were also discussing how we can safely include zuul configuration from zuul-jobs/openstack-zuul-jobs/etc without incurring things as potential syntax failures due to clashing names or whatever else.19:25
jeblairdmsimard: cool, when things settle down, we should probably establish a storyboard tag or something for issues like that19:25
dmsimardSo far we're using parameters like shadow and selective includes19:25
jeblaircool, it'll be good to know whether those work as intended or need further work19:26
dmsimardI was surprised to even see that those were available in the first place, it means someone thought about the use case19:28
dmsimardso whoever added that in, +++19:28
tobiashdmsimard: I added a point to your etherpad19:47
dmsimardtobiash: yeah, I haven't quite figured out that one yet19:49
jlktristanC: SpamapS haven't read full backlog, but I also was thinking in teh direction of a kubectl driver for ansible, so that you can do kubectl exec type things and not need ssh inside the images.19:49
tobiashI looked a bit into that some weeks ago but didn't have time for really working on that yet19:49
jlkI see "ssh" as a byproduct of how openstack VMs work19:50
jlkand if they're not necessary to execute things inside the contained (vm or otherwise) environment, then all the better.19:50
pabelangerjlk: the comment that interested me from tristanC was how would do synchronize task for logs from container?19:51
jlksynchronize stil works with the docker module19:52
jlkthe docker connection module for Ansible, which just uses docker exec. Presumably the same sort of thing can work on k8s19:53
jlkgranted, I haven't looked close enough at _how_ it works for Docker, but I've used it before.19:53
pabelangerkk19:53
*** hasharAway is now known as hashar20:00
openstackgerritAndreas Jaeger proposed openstack-infra/zuul-jobs master: Half-Revert "Revert "Add ensure-reno and ensure-babel roles""  https://review.openstack.org/52155820:07
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114220:15
mordredjlk: ++ to using native k8s/docker connection for the things - I believe flaper87 wrotea kubectl module for ansible - but I think in this case we'd need a kubectl connection plugin, yeah?20:19
clarkbin fairness ssh is a byproduct of how ansible works :P20:21
dmsimardwhat's a native k8s connection ? I know openshift has "oc rsh" but I haven't done much pure k8s20:21
clarkbif we used some other rpc system we'd use whatever protocol that speaks over20:22
dmsimardon openshift if you do "oc rsh <pod>" it opens a shell in the pod, not sure what it does behind the scenes (or how it would pick one of the containers in the pod)20:23
dmsimardopenshift-ansible probably have a module/plugin for that actually20:23
clarkb(it also happens to be the case that jenkins used ssh as well, so that wasn't a change for us)20:23
clarkb(but nothing about it is openstack vm specific20:23
dmsimardyeah doesn't look like openshift-ansible folks have something like that yet but it'd be "insanely cool"20:26
dmsimardthe different upstream connection plugins are here https://github.com/ansible/ansible/tree/devel/lib/ansible/plugins/connection20:27
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114220:49
mordredclarkb, dmsimard: yah - I think it's actually "ssh is a byproduct of the fact that we're currently using VMs that look like computers for our test nodes" - due to the connection plugin support, if we produce test nodes that are not intended to be connected to over ssh, that should not be a blocker20:59
mordredwe're already aware of at least win_rm for the windows nodes21:00
*** jkilpatr_ has joined #zuul21:00
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Normalize daemon process handling  https://review.openstack.org/51738121:00
*** jkilpatr has quit IRC21:01
mordredso yah - being able to support docker connection plugin for docker container build nodes or a theoretical kubectl connection plugin for k8s pods seems like a good way to handle those as we grow support for them21:01
*** jkilpatr_ has quit IRC21:01
jeblairreminder, meeting is in 1 hour (this may have changed in reference to your local time for folks in the usa)21:01
mordredjeblair: nice daemon process handling cleanup patch21:03
clarkbmordred: ya I just think its important to decouple that from openstack, nothing in openstack says you have to do it that way and in fact you could run windows on openstack if you wanted and do whatever it is you do with windows for example.21:05
jeblairyeah, in working with leifmadsen i was able to see what worked well and didn't in zuul/nodepool.  nodepool's handling of pidfiles wins because of the way they use paths, and zuul's default logging config wins -- that's something we should work on porting to nodepool.21:05
clarkbmordred: the determining factor for us is ansible (and prior to that was jenkins)21:06
mordredclarkb: I agree - it's not caused by openstack ...21:06
mordredclarkb: but I think it's just as important to point out that it's not actually driven by ansible either, as ansible has plenty of non-ssh connection plugins21:07
mordredit is the combination of the fact that we are running ssh capable VMs and that is the default ansible connection mechanism21:07
clarkbmordred: right thats fine we could use whatever ansible supports21:07
mordredyah21:07
clarkbit does ssh by default so we've ended up there by default21:08
mordredyup21:08
clarkbjenkins too fwiw, there were non ssh methods21:08
mordredand I thnik it's 100% the right choice of how to connect to linux-based things that look and behave like multi-user/multi-process computers21:08
mordred(whether those come from bare metal, vms or containers)21:09
clarkbya21:09
jeblairokay, i'm going to go out on a limb here... is there anything new happening in this discussion, or are we just repeating the discussion we have every few weeks?21:09
jeblairi'm asking because i'd like for us to, as a group, avoid spinning our wheels21:10
jeblairthese discussions take a *lot* of time and attention21:10
jeblairand this is an important topic21:10
clarkbif nothing else, I think it shows that we may have confusion that ansible is what drives this21:10
jeblairand we're not going to be able to address it fully right now.  we do have an item (maybe more than one item) on the roadmap to deal with non-vm test hosts, after the release21:10
clarkbzuul doesn't really care, nodepool doesn't really care and openstack definitely doesn't care. Ansible is what determines this so maybe we need to write that down "if you want to connect to something not sshd then you need to make ansible talk to it"21:10
jeblairis there a way we can structure our work so that we can say "yeah, we totally know this is a thing, let's work on it later?"21:11
jeblairor do we need to further delay the zuulv3 release so that we can work on this now?21:11
clarkbI don't think we need to work on it, its not something that zuul has to be directly aware of I don't think (but I could be wrong about that)21:12
jeblairthis is what i was trying to accomplish with a roamap.  have a place where we say we know what we're going to do in the future, but it's in the future.  now we have boring^Wexciting things like "actually make system usable and release" to do :)21:12
clarkbit would just be documentation for zuul users that says "make ansible do it"21:12
clarkbregardless of what the system is (windows, k8s, docker, etc)21:13
jeblairclarkb: well, i think this needs a significant amount of thought and discussion but i'm not prepared to have it now, and i'd like other folks to have the space to focus on near-term goals.21:14
jeblairso how can we do that?  or am i wrong?  do we just need to accept that we need to solve containers right now?L21:14
jeblairbasically, in my mind, the roadmap is flexible -- we can change it.  but if we agree about it, let's stick to it.21:15
clarkbI don't personally think we need to solve containers right now (we've not needed them in ~5 yaers). I do think it is something people are interested in tackling though and if there isn't any direction for them that maybe isn't a great thing21:15
jeblairi'm happy to work on that, but i'm afraid we don't have enough people signed up to do the things we need to get to the v3.0 release.21:16
leifmadsenisn't this the type of discussion we have every 6 months? :)21:17
leifmadseni.e. isn't that was devcons are for?21:17
mordredjeblair: the main thing I was trying to communicate by responding to this particular incarnation of the topic was to remind folks that support for non-ssh connection plugins is going to be important no matter what the thing on the other end of the connection winds up being - and I think from a zuul pov much of that doesn't have a ton to do with containers vs. non-containers21:17
jeblairleifmadsen: yes, i'd like to get us synchronized with that.  hopefully in the next cycle we can take advantage of it more.21:17
jeblairleifmadsen: we did indeed sketch out our roadmap at the last one21:17
jeblairleifmadsen: i'm wondering if it has meaning.  :)21:18
leifmadsendepends where it is, and if you're working from it21:18
leifmadsenand if everyone knows what it is, and that it's being worked from :)21:18
jeblairleifmadsen: indeed21:19
leifmadsenalso, it's good to document these things as you go for sure, because you'll forget all this stuff when you're building an agenda for the next devcon :)21:19
leifmadsenso a place to document these discussions so they can be useful and had, then documented, then added to roadmap, is a good place to be21:20
jeblairleifmadsen: that's another reason to put the roadmap in storyboard i suppose21:20
leifmadsenis it just in an etherpad right now? :)21:20
jeblairleifmadsen: it was an etherpad, became an email, and i believe at the last meeting pre-summit, clarkb and i signed up to make it be in storyboard21:21
leifmadsenyea, so basically, it doesn't exist :)21:21
jeblairleifmadsen: not sure why not -- we're all on the email list21:21
leifmadsenI don't consider etherpads and email documentation :D21:21
leifmadsenemail lists are incredibly hard to go and lookup information. The ideal place, in my head, is a link to a sane location from the README21:22
jeblairit's not documentation, it's meant to be a discussion, and something to agree on21:22
leifmadsenit's documentation21:22
leifmadsenyou're documenting a plan21:22
leifmadsenotherwise, you're just pontificating21:22
jeblairleifmadsen: i think there's a miscommunication here21:22
jeblairleifmadsen: i think it's really important for us to discuss something like this, which is why i started it in person, and online, at the summit in an etherpad21:22
leifmadsenI'm not sure there is :)21:22
leifmadsenyes, I agree21:23
jeblairleifmadsen: then followed up on the mailing list to make sure even more people were included21:23
leifmadsenwhat I'm saying, is your roadmap is not yet documented21:23
leifmadsenif it's only on an email list and an etherpad21:23
jeblairleifmadsen: it will eventually end up somewhere.  at our last meeting before the summit, we discussed whether it should be in a readme in repo or in storyboard21:23
jeblairleifmadsen: and we decided to put it in storyborad21:23
jeblairleifmadsen: i think that should make you happy.  but i suggest your criticism is unwarranted.21:23
leifmadsensure, that works, and then you can link to it from a README so that people getting to the project know where to look21:24
leifmadsenunwarranted?21:24
leifmadsenummm ok21:24
leifmadsenI'm not intending my tone to be harsh21:24
jeblairhere's the email: http://lists.openstack.org/pipermail/openstack-infra/2017-November/005657.html21:24
pabelangerre:containers, it does seem to be something people want before adopting zuulv3. But at the same token, we are seeing requests to tag zuulv3 now. So feel like catch-22 in that aspect. But agree, we need to stablizing things more before new features21:25
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114221:26
leifmadseneveryone wants all the things all the time21:26
leifmadsenjust gotta have your must haves, then move along with your life; development is never ending21:26
mordredjeblair: having just now gone to look at the roadmap again real quick, I notice that "nodepool backends" is in "long term / design", which I think we may want to at least partially reconsider, as I know getting windows nodes is important to tobiash21:26
jeblairmordred: that's a backend?21:27
mordredjeblair: the part that of that that I think may be worth considering on the pre-3.0 roadmap is ensuring that we can support backends that don't use ssh ... as I could imagine that there might be some sort of weird breaking change associated with that21:27
jeblairmordred: i don't see why that can't be a forward-compatible thing in v3.1?21:28
pabelangerI could sign up for 'demonstrate openstack-infra reporting on github' (gtest-org project) and 'add command socket to scheduler and merger for consistent start/stop' items myself21:29
jeblairleifmadsen: yep, we have to make compromises21:29
mordredit might be able to be, for sure- but I know that talking through the win_rm auth needs with tobiash from a nodepool->zuul perspective exposed a few design assumptions21:29
jeblairpabelanger: thank you so much!21:29
jeblairmordred: could maybe someone write that up as an email or something?21:29
mordredjeblair: sure21:30
mordredjeblair: like, I don't think we need to have other nodepool backends fully implemented - it's more "before we release a 3.0, can we sanity check that we're not including something that would make doing so extra hard/awkward"21:30
mordredjeblair: but happy to write that up as an email to the list21:30
jeblairmordred: if there is something breaking, i agree we should look at it early.  but i think our goal should be to polish up the thing that we are basically running, and do the minimum to ensure we're not backing ourselves into a corner later, and defer the rest.21:31
mordredjeblair: yes, I agree with that 100%21:31
jeblair(because "we're running a thing but telling no one else to run it because doing so sucks" is not a state we should be in long-term)21:31
mordredjeblair: I mostly want to make sure we're not backing ourselves into a corner on this one21:31
mordredjeblair: ++21:31
clarkbjeblair: I've reviewed the daemon change. I like it but a couple things that I think need to be addressed21:39
jeblairclarkb: thanks, i'll look at that.  i also just noticed it collided with another patch that landed ahead of it.  that's why pep8 failed.  i'll have to untangle that too.21:40
jeblairclarkb: i'm having trouble following your static comment21:42
jeblairclarkb: i call "Executor().main()", so that's an instance method21:43
clarkbjeblair: when you call main() you are doing so as ClassName.main() not ClassObject.main()21:43
jeblairclarkb: nah, there's a () after Executor21:43
clarkboh am I just compeltely blind? because that may be too21:44
jeblairit's just anonymous21:44
clarkbya I'm just blind. So I think the only other thing is making sure you don't need a pidfile when running non daemon mode21:44
jeblairclarkb: cool.  i agree with that and will implement in next rev21:44
clarkband I think that with: pass will lock the file21:44
jeblairclarkb: yeah, it's intended to lock-then-unlock21:45
jlkjeblair: I think the "new thing" that's happened with "how do we container" is that a driver implementation was submitted21:47
jeblairjlk: oh, this is tristanC's change?21:48
jeblair521356?21:49
jlkthat's the one that spurred the conversation. I have the tab open to read it, but I haven't yet. I glean from the conversation that it implements / relies upon sshd inside the container21:49
jeblairjlk: i suspect that's going to spawn the design discussion that we all know we need to have21:50
jeblairso it still may be worth asking ourselves, is it most beneficial to have this now, or would it be better to defer it until we're closer to that point on the roadmap.  or do we need to change the roadmap.21:50
jeblairall of those '.' should be '?'.  :)21:51
jlkwell yeah21:51
jlkI'd like to contribute to this discussion/feature, because I have use cases, and some time to dedicate to it. But if y'all aren't ready for it, then now's not the time21:51
jeblairwell, it's more "us" than "ya'll" i hope :)21:52
jlkthat's... entirely fair. I've been checked out lately and I slipped phrasing21:52
jeblairto be fair "us aren't ready for it" isn't great phrasing either.  heh.  :)21:53
jeblairi'm apparently englishing poorly today21:54
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Normalize daemon process handling  https://review.openstack.org/51738121:55
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114221:59
clarkbmeeting time now right?21:59
jeblairyep!22:04
jeblairin #openstack-meeting-alt22:04
pabelangerI'll only be able to attend first 30mins today22:04
jeblairjlk, SpamapS, leifmadsen, Shrews, dmsimard: ^22:05
*** hashar has quit IRC22:06
jeblairmordred: ^22:07
leifmadsenI can't make 5pm meetings22:09
leifmadsenso won't be there :)22:09
*** jlk has quit IRC22:31
*** jlk has joined #zuul22:32
*** jlk has quit IRC22:32
*** jlk has joined #zuul22:32
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add general sphinx and reno jobs and role  https://review.openstack.org/52114222:38
dmsimardoh hey just got my notification that the zuul meeting is in 10 minutes from now .. :/22:50
openstackgerritMerged openstack-infra/zuul-jobs master: Make build-python-release job  https://review.openstack.org/51392522:56
ianwdmsimard: just be sure not to change the timeline, don't want to have a marty mcfly situation on our hands :)23:00
clarkbjeblair: re zuul config breakages the one I'm most familiar with is the parent to final job one so I'll start with that23:01
clarkbjeblair: one of the release jobs that only runs on tags was marked final. Neutron/Horizon then parented that job in order to add require projects to the job. This merged with no errors. Problem then only came up when tag was made and zuul said no that job is final23:02
clarkbthis is less than ideal because it could be quite a bit of time between jobs. Maybe we handle this not with zuul self checking internally but by running a job that does an actual compile of the jobs?23:02
dmsimardjlk: the support for different version of Ansible doesn't necessarily have to come out of Zuul.. for example for ARA integration testing I just install a "nested" ansible of the desired version on the target nodes23:03
dmsimardjlk: because I want to test that ara works with different version of ansible/py27/py35/etc23:03
clarkbI'm guessing that the "binding" of final happens far too late in the compile process for the simpler syntax checking to catch it23:03
jeblairclarkb: yeah, that's a run-time error which is impossible (or at least very difficult) to detect in advance23:03
dmsimardjlk: however, it sort of sucks because then the "zuul" ara report contains just one large command instead of the granular tasks23:04
jlkdmsimard: right. That was my intent. I don't think you should be able to select which executor your job runs from based on the ansible version on the executor23:04
jlkotherwise I think it feels like being overly dependent on the fact that Zuul runs Ansible under the hood, making it that much harder to change (if we ever changed)23:04
clarkbjeblair: I'm digging through gerrit changes to find the error that prevented zuul from starting now23:05
jeblairclarkb: basically, you can construct combinations of variants that could theoretically trigger or not trigger such problems.  so it's hard to detect in advance.23:05
dmsimardjlk: I still think, however, that "forcing" the upgrade from 2.3 to 2.4 on users is dangerous23:05
clarkbjeblair: https://review.openstack.org/#/c/519949/ is the change23:05
clarkbhttps://review.openstack.org/#/c/520205/ is another that we may want to look at as far as pre merge testing goes23:05
dmsimardjlk: historically there has always been issues between "major" versions (2.0 -> 2.1 -> 2.2 -> 2.3 -> 2.4)23:06
jlkyeah, do you pin that to major zuul versions?23:06
dmsimardpeople usually pin ansible for that reason23:06
clarkbwe've had issues with minor updates too fwiw23:06
dmsimardclarkb: yeah, but they're not as common23:06
jeblairclarkb: there is perhaps a simple case where you could say that all variants for a certain job are final and therefore there would be an error.  but that's hard, and it's half a solution, so i worry about whether it's a good idea.23:06
dmsimardthey're gotten better at testing but their internal API is not stable23:06
jlkyeah there's two concerns at play23:07
jlkwill Zuul's integration with Ansible continue to work23:07
clarkbjeblair: ya thinking it might be better to have a job that compiles zuul outside the running zuul?23:07
clarkbjlk: if that makes sense?23:07
jlkand will end user playbooks continue to work as expected23:07
jeblairclarkb: well, the thing is that the job doesn't exist until it's run.  it's based on variants matching a specific change/ref/etc23:08
dmsimardclarkb: my idea around testing base jobs was along these lines, run a nested zuul (complete with executor, with a second node to be used in the upcoming static nodepool driver) and then do a "manual" zuul enqueue of the real job23:08
jlkHeizenjob23:08
dmsimardclarkb: but the problem with that is that you still need to load like 2000 repositories worth of configuration23:09
clarkbdmsimard: ya23:09
clarkbjeblair: gotcha23:09
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Update fetch sphinx output to use sphinx vars  https://review.openstack.org/52159023:09
dmsimardclarkb: but then the nested zuul also doesn't have the private keys to decrypt the secrets23:09
dmsimardwhich is totally okay, because otherwise that'd allow people to peek at things23:10
clarkbI wonder, could we do a lint type check that just walks up the parent tree for finals?23:10
clarkbjeblair: ^23:10
clarkbmaybe thats the case you mean where all variants are final23:10
jeblairclarkb: that's the thing i've probably been unable to express clearly -- even the parent hierarchy isn't fully determined until the job runs23:11
clarkbjeblair: got it23:11
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add support for warning-is-error to sphinx role  https://review.openstack.org/52161823:11
jeblairif you say "parent: foo" and "foo" has N variants, we don't know which will apply.  it could be anywhere from 0-N.23:11
jeblair(if 0 apply, the job doesn't run)23:12
clarkbok lets ignore that one for now because its reltively minor and not easily fixable. https://review.openstack.org/#/c/519949/1 is the fix for the thing that prevented zuul from starting23:12
clarkbI think ^ is more important as it impacts the ability to run zuul23:12
clarkb(also if anyone is wondering we think we tracked back the source of the OOM to puppet openstack pushing a ton of job changes all at once which I think is a known issue, we asked OSA to push them slowly)23:13
jeblairclarkb: i suspect that error came in due to mordred's git surgery?23:15
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Add support for warning-is-error to sphinx role  https://review.openstack.org/52161823:15
jeblairclarkb: i was wondering, thanks23:15
clarkbjeblair: ya I think .zuul.yaml must've come from os-c-c23:15
clarkbbut that should still have failed pre merge no?23:15
jeblairclarkb: we should be in much better shape than when osa did it originally, but still, lots of job changes can use lots of ram23:15
jeblairclarkb: i think mordred did git surgery to create that branch and push it directly23:16
clarkbah23:16
clarkbthat may be the piece I am missing23:16
jeblairit is likely that caused zuul to get stuck on an old config as well23:16
jeblair(it would have kept running with the last config it was able to fully load)23:16
clarkblesson here then is be very careful with force merges23:16
* mordred reads23:17
jeblairone thing that will make this very specific case better: we plan to drop the project name from in-repo config files23:17
mordredAH. derp23:17
mordredyah. that's my bad for sure23:18
jeblairbut the general problem of dealing with erroneous configs remain.  i think we'll eventually have to do something like have zuul automatically remove projects with broken configs.  once we have the dashboard, there will be a nice place to have a big red warning that has happened.23:18
mordredwe can actually delete that branch already ... it was just thereto prevent gerrit from creating thousands of gerrit changes23:18
jeblair(that could have lots of follow-on effects, like breaking other projects, but i don't think there's anything else that could be done)23:19
mordredin fact, how about I go ahead and delete it now23:19
mordredjeblair: ++23:19
SpamapSsorry to miss the meeting.. had a long drive and we got moving late :-P23:46
jeblairno meeting while driving! :)23:46
SpamapSFYI, I use my zuul to test our ansible deployment stuff, which is pinned at ansible 2.123:56
SpamapSI just install ansible in a virtualenv on a node and run it.23:56
SpamapSWhich is good, because I wouldn't want to mix the concerns of Zuul with the concerns of my k8s deploying ansible code.23:57

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!