Monday, 2018-02-19

*** elyezer has quit IRC00:35
*** elyezer has joined #zuul00:37
*** elyezer has quit IRC00:47
*** elyezer has joined #zuul00:48
*** elyezer has quit IRC01:39
*** elyezer has joined #zuul01:41
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Reorgnize the output bundle split  https://review.openstack.org/54570602:04
mordredSpamapS: is remaning az capacity available to an api user?02:08
mordredSpamapS: ah- I understand now (read more scrollback)02:15
*** dkranz has quit IRC02:30
*** rlandy has quit IRC02:52
dmsimardmordred, andreaf: yes, actually I did release ara 0.14.6 to fix just that: https://github.com/openstack/ara/commit/9b962c5dcb77dc3e7687f75030ea41de30a5474604:50
dmsimardI fixed it from the perspective of ara but I still feel it's a bug. I remember discussing it with #ansible-devel and it was supposed to be fixed in 2.4.3 iirc04:52
tobiashSpamapS: we should probably make the az stickyness configurable per provider05:08
tobiashIn my cloud this also has these disadvantages05:09
tobiashAnd in my cloud inter az bandwith (which probably was a reason for az stickyness) is a non-issue for me05:10
tobiashEspecially in smaller clouds this is even problematic as it can prevent nodes from being created regardless of quota05:12
tristanCtobiash: shouldn't we make az stickyness part of the nodeset definition instead?05:12
tobiashThat depends strongly on the cloud and its inter az bandwidth05:13
tobiashSo I think this should be primarily set by nodepool as the user usually doesn't have the knowledge to judge about that05:14
tristanCtobiash: doesn't it change network configuration when nodes aren't in the same az?05:16
*** elyezer has quit IRC05:18
*** elyezer has joined #zuul05:21
tristanCiiuc, cross az nodeset won't necessarly be in the same tenant network, and maybe some job explicitely need that05:27
tobiashtristanC: no it doesn't necessarily05:34
tobiashThe networks used is also defined by the provider05:35
*** bhavik1 has joined #zuul05:52
*** bhavik1 has quit IRC06:00
*** threestrands has quit IRC06:10
*** threestrands has joined #zuul06:23
SpamapSSo, I'd say that cross-az-stickyness should be a boolean, defaulting to true.06:48
SpamapSIf you know your cloud doesn't need it, and you find yourself wishing it wasn't a thing, you can turn it off.06:48
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: angular controller for status and stream page  https://review.openstack.org/54575506:48
SpamapSFor me, cross-AZ bandwidth is the same as intra-AZ bandwidth when the HV's are in different racks.06:49
SpamapSAnd the networks are all more or less the same.. cross-AZ may go through one extra router, but that may happen intra-AZ too if the nodes are on different subnet.s06:49
*** threestrands has quit IRC07:00
*** AJaeger has quit IRC07:00
*** AJaeger has joined #zuul07:04
tobiashSpamapS: yeah, it should default to true to keep the current behavior07:06
tristanCtobiash: thanks for the ansible-2.4 port, it seems to be working fine07:10
tobiashtristanC: yeah, it worked in my testenv and I'm going to roll this out on production this week :)07:11
tobiashThis is essential for me to get real windows jobs running as with ansible 2.3 almost every win_* module is broken with py3...07:12
tristanCwe also have some playbooks that are >2.4 only, hopefully we'll deploy your patch soon too07:16
tobiashtristanC: we'll land it shortly after 3.0 release07:16
SpamapSwhich is hopefully.. shortly. :)07:18
* SpamapS has been missing meetings.. not sure what's left07:18
tristanCtoo bad, ansible-2.4 support is also blocking fedora packages07:19
SpamapSWhat's the reasoning behind delaying the bump after release?07:19
SpamapS2.4 breaks old stuff?07:19
SpamapS(Does kind of feel like 2.4 was overly aggressive, and introduced some new annoying, meaningless warnings)07:20
tobiashSpamapS: the openstack release process is critical and roughly at the same time as the 3.0 release07:21
tobiashansible 2.4 is a high risk to disrupt the openstack release so it will be merged after that07:22
AJaegerzuul and nodepool cores, ianw and myself fixed the integration tests, could you review the stacks starting at https://review.openstack.org/#/c/545163/ and https://review.openstack.org/#/c/545158/ , please?07:25
*** elyezer has quit IRC07:27
*** elyezer has joined #zuul07:31
SpamapStobiash: ah yeah that makes more sense actually. :)07:42
SpamapSI kind of think we're going to have to support multiple ansible versions at some point.07:43
SpamapSbtw, my AZ problems show nicely here http://paste.openstack.org/show/677466/07:43
SpamapSneed 10 nodes, have 11, building 6 (min-ready is 15)07:45
*** chrnils has joined #zuul07:47
*** sshnaidm|off has quit IRC08:35
*** electrofelix has joined #zuul08:35
*** jpena|off is now known as jpena08:43
*** sshnaidm has joined #zuul08:44
openstackgerritAntoine Musso proposed openstack-infra/zuul master: WIP: ensure that Change.number is a string  https://review.openstack.org/54576809:04
*** sshnaidm has quit IRC09:07
*** sshnaidm has joined #zuul09:08
*** hashar_ has joined #zuul09:24
openstackgerritAntoine Musso proposed openstack-infra/zuul master: WIP: ensure that Change.number is a string  https://review.openstack.org/54576809:46
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: Tenant config dynamic loading of project from external script  https://review.openstack.org/53587809:46
*** elyezer has quit IRC09:58
*** elyezer has joined #zuul09:59
*** kmalloc has joined #zuul10:14
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: zuul web: add admin endpoint, enqueue & autohold commands  https://review.openstack.org/53900410:19
*** tosky has joined #zuul10:29
*** fbo has quit IRC10:40
*** dmellado has quit IRC12:26
*** elyezer has quit IRC12:30
*** elyezer has joined #zuul12:35
andreafcorvus: I've done some tests in https://review.openstack.org/#/c/545696 with group_vars12:35
andreafcorvus: from what I can see devstack_localrc is merged between controller group vars and generic vars (from parent job)12:37
andreafcorvus: but not between subnode group vars and generic vars (from parent job)12:37
andreafcorvus: http://logs.openstack.org/28/545728/4/check/devstack-multinode/f2a68a6/compute1/logs/_.localrc_auto.txt12:38
andreafcorvus: could you have a look into that?12:38
*** jpena is now known as jpena|lunch12:39
*** sshnaidm has quit IRC12:42
*** dmellado has joined #zuul13:07
*** sshnaidm has joined #zuul13:25
*** jpena|lunch is now known as jpena13:31
*** rlandy has joined #zuul13:32
*** rlandy is now known as rlandy|rover13:33
*** dmellado has quit IRC13:38
*** JasonCL has joined #zuul13:41
*** JasonCL has quit IRC13:46
*** JasonCL has joined #zuul13:46
*** JasonCL has quit IRC13:49
*** JasonCL has joined #zuul13:50
*** JasonCL has quit IRC14:10
*** dmellado has joined #zuul14:44
*** sshnaidm has quit IRC14:46
*** sshnaidm has joined #zuul14:47
*** dmellado has quit IRC15:02
*** tosky_ has joined #zuul15:02
*** tosky has quit IRC15:03
*** elyezer has quit IRC15:11
*** JasonCL has joined #zuul15:13
-openstackstatus- NOTICE: Zuul has been restarted to pick up latest memory fixes. Queues were saved however patches uploaded after 14:40UTC may have been missed. Please recheck if needed.15:17
*** swest has quit IRC15:17
*** swest has joined #zuul15:19
*** elyezer has joined #zuul15:24
corvusandreaf: that's strange, i wouldn't expect them to be merged at all -- i'd expect one to take precedence over the other15:25
andreafcorvus: oh.. is it? so what I was trying to do would not work - i.e. defined devstack_conf in vars and then extend it in group vars15:27
corvusSpamapS, pabelanger: we shouldn't need a custom logging configuration to answer questions like SpamapS is asking, so let's figure out what log lines we need to switch to info in order to at least answer basic questions like "is nodepool doing anything"15:27
pabelanger++15:28
andreafcorvus i.e. https://review.openstack.org/#/c/545696/10/.zuul.yaml15:28
andreafcorvus: yeah I think you are right they are basically overriding each other... the subnode only gets what's defined in group vars for the subnode15:29
andreafcorvus: any idea on how to achieve what I'm trying to do? I don't really want to repeat the whole localrc settings for every node group, I only want to add what's relevant15:31
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Remove .json suffix from web routes  https://review.openstack.org/53701015:31
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Port per-change status to zuul-web  https://review.openstack.org/53590315:31
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Re-enable test of returning 404 on unknown tenant  https://review.openstack.org/54564415:31
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Add /info and /{tenant}/info route to zuul-web  https://review.openstack.org/53701115:31
corvusso we could leave this the way it is, which basically means the job author would have to be careful about whether they put a given variable in 'vars', 'host_vars', or 'group_vars'.  or we could merge variables inside zuul and only use host vars when writing the inventory.15:33
mordredSpamapS: yes, I agree - and I have some thoughts on doing that but have been holding off on them until post zuul release15:33
AJaegerzuul and nodepool cores, ianw and myself fixed the integration tests, could you review the stacks starting at https://review.openstack.org/#/c/545163/ and https://review.openstack.org/#/c/545158/ , please?15:34
mordredcorvus, tobiash, tristanC: when you get a chance, could you review the four patches above? (the first four of the javascript topic) I'd like to get them in and deployed (or get them rejected) so that the live draft changes later in the stack will work as we review those patches15:35
mordredandreaf, corvus, SpamapS: also - I just got a patch landed to keystoneauth and released last week that allows us to tell keystoneauth to split its logging across four more specific loggers ... which was all in support of being able to have *some* keystoneauth debug logging in the default nodepool logging config but not *all* of it15:35
mordredandreaf: sorry - I mentoined the wrong people15:36
mordredpabelanger: ^^15:36
corvusandreaf: so i think if we kept this the way we have it now, you'd have to stop using vars at all, and only use group_vars, which means putting the devstack_localrc in both group_vars section.15:37
mordredpabelanger, corvus, SpamapS: https://docs.openstack.org/keystoneauth/latest/using-sessions.html#logging FWIW - I'll get us a patch up so that we can take advantage of that15:37
*** JasonCL has quit IRC15:38
*** JasonCL has joined #zuul15:38
corvusmordred: yeah, that will be nice to have in our debug logs, but i think it's too low-level for the default log (or what SpamapS is asking for)15:38
andreafcorvus: I don't think that's very maintainable, unless we use vars to define a base set of devstack setting, and then extend the dictionaries in group vars15:38
andreafcorvus not sure we can really do that though15:39
corvusandreaf: right, so if that's the goal, we either need to rethink the whole devstack_localrc dictionary thing (maybe split it out so that not everything is in one dictionary), or merge vars, host_vars, and group_vars internal to zuul.15:40
corvusmordred: thoughts ^?15:40
mordredcorvus: hrm. I'm not crazy about merging vars, host_vars and group_vars internal to zuul - that seems like starting to reimplement ansible internals15:41
corvusmordred: well, ansible *doesn't* merge them, so...15:41
andreafcorvus: we could have variables with different names and merge them into the write_devstack_conf module15:41
corvusandreaf: yep, that's a possibility15:42
Shrewsmordred: oh, that logging split is nice15:42
*** JasonCL has quit IRC15:43
mordredcorvus, andreaf: well, it *can*: http://docs.ansible.com/ansible/latest/intro_configuration.html#hash-behaviour15:43
andreafcorvus, mordred: however I fear that would make a poor user interface15:43
andreafmordred: yeah it's configured not to by default  right15:43
mordredbut that's potentially much more widely reaching15:43
mordredand I think has a greater chance to confuse people15:44
corvusmordred: oh neat, i didn't know that.  that does change things.  :)15:44
corvuswe could expose that option, and let the devstack job use it15:44
mordredcorvus: it does- but I think that might break people's roles or playbooks if they're not expecting it - so maybe doing some sort of merge specific to the zuul variables in zuul is fine here15:45
mordredcorvus: oh - that's an idea15:45
mordredcorvus: that way it doesn't mess up variables in roles or playbooks where people aren't expecting it, but devstack can opt-in15:46
corvusfor that matter, if we merged them internal to zuul, we could also make that a per-job flag i think.15:47
mordredyah15:47
AJaegercorvus: and that would apply to all used roles and playbooks? So, could still break some...15:49
AJaegertricky ;(15:49
corvusto summarize: merging internally to zuul is probably the simplest in some ways (we don't add any new zuul options), but the reason not to is because we think someone might prefer the override behavior between host_vars and group_vars at some point, and it alters a standard behavior of ansible.15:49
corvuswhile exposing the hash_behavior option is easy to implement and allows a normal and valid use of ansible.  we just need to tell people about that option when this case comes up.15:50
andreafmordred, corvus: I think the best would be to be able to set the behaviour o specific variables15:51
corvusAJaeger: yes it could -- but that's true anytime you use that option with ansible.  so, as the docs say, don't use it unless you really need it.15:51
andreafmordred, corvus if we remove the test matrix we will have to set list of services and that's going to be dicts that we don't want to merge15:52
corvusAJaeger: i don't anticipate that would be a substantial problem in practice.  much of the time, we don't even need hostvars.15:52
mordredluckily, since it would be a job variable, we can try it and make sure none of hte base roles will be borked15:52
corvusandreaf: the services aren't lists, they are dict entries, specifically so that we can merge them.15:52
mordredcorvus: well, it affects all variable precedence - so it's possible it could introduce a merge where otherwise a replace would have happened15:52
mordredeven i host_vars/group_vars aren't present15:52
mordredcorvus: I agree, I doubt it'll be an issue for us - but it's still *possible* that something in one of the base roles would not be hash-merge clean15:53
corvusmordred: oh.  that might be a bummer for site vars.15:53
mordredyah. I think we want to double-check that it doesn't allow site vars to have local values merged into them15:54
andreafcorvus: yes I know they are dicts - I guess it depends if we want controllers and subnodes  to disable they don't need15:54
andreafcorvus or to define which services they do need15:54
andreafcorvus, mordred: well maybe I should just duplicate devstack_localrc and everything would be easier15:55
andreafcorvus, mordred: would it be possible in the job definition to set the shared part of the dict in a shared_devstack_localrc varilable, and then set devstack_localrc in group vars as the merge of shared plus a custom part?15:57
corvusandreaf: yeah, we could merge them in the role15:59
andreafcorvus, mordred: can we use jinja2 in the job definition? for instance  devstack_localrc: "{{ shared_devstack_localrc |combine(my_devstack_localrc) }}"16:07
corvusandreaf: i don't believe that will work -- zuul isn't going to interpolate it, and i don't think ansible does so for inventory files16:09
mordredcorvus: it doesn't - but it does for vars files in roles - so it's _possible_ it'll do late interpolatio16:12
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Simplify launcher cleanup worker  https://review.openstack.org/54586716:12
mordredcorvus, andreaf: which is to say - I agree with corvus, I think it won't work ... BUT, it's totally worth trying :)16:14
andreafmordred corvus another question - will zuul merge dicts between group_vars in parent and group_vars in children jobs?16:15
corvusandreaf: yes16:15
tobiashmordred: am I right that your per change status change only uses the change number?16:16
tobiashmordred: I'm asking because github has no global change numbers16:16
andreafcorvus ok so if base multinode defines services for controller and services for subnode, a child job can selectively add/remove services on controller and on subnode16:16
corvusandreaf: yep, that should work.16:17
tobiashmordred: if I'm right this won't work for github16:18
mordredtobiash: good point - yes, I believe you are correct about that16:23
mordredtobiash: we probably should add support for filtering by project+change too, yeah?16:23
tobiashmordred: let me check what the 'id' field is in github16:24
andreafmordred, corvus you never know, it may work https://review.openstack.org/#/c/545696/11/.zuul.yaml16:24
tobiashmordred: ok, the id field is something like this: "id": "6,a672a19d197779e4d5410cb25775ee373cf47269"16:25
tobiashand your checking the whole string right?16:25
mordredyah16:25
tobiashassuming the git hash is unique enough this should work equally well in github :)16:26
mordredcool. so there is a gh value that will work - it might not be immediately guessable by an end-user like the gerrit one is16:26
tobiashnot guessable but the intention is probably reporting this via status_url right?16:27
mordredtobiash: but if we're mostly expecting this to get called as the result of links we're putting places, that should be good enough for now - and we can leave figuring out how to get someone from "ansible/ansible/pulls/6" to status/change/6,a672a19d197779e4d5410cb25775ee373cf47269 later16:27
*** JasonCL has joined #zuul16:27
mordredtobiash: yup16:27
mordredhowever - I thnk at some pint having a dashboard page for a change that doesn't change as the change is revised would e nice - perhaps something like 'change/review.openstack.org/501040' and '/change/github.com/ansible/ansible/6' - so that the key is similar to what's used in depends-on lines?16:30
corvus++16:31
mordredbut that dashboard page could use the ansible/ansible/6 to look up 6,a672a19d197779e4d5410cb25775ee373cf47269 and then get the status from status/change/6,a672a19d197779e4d5410cb25775ee373cf4726916:31
mordred(and in fact could potentially look up older builds/status for that change too and show info about them from the build history while showing live status or completed build info for the current change16:32
SpamapScorvus: to be clear, I log at debug level. The only indication I get that nodes are spinning up is this: 2018-02-18 01:04:45,639 DEBUG nodepool.NodeLauncher-0000005071: Waiting for server ded9e08e-311c-4c98-8b1f-1d516148aed9 for node id: 000000507116:32
SpamapSactually I also maybe just missed this one 2018-02-18 01:04:30,230 INFO nodepool.NodeLauncher-0000005071: Creating server with hostname k7-a-0000005071 in a from image centos7-kolla for node id: 000000507116:33
corvusSpamapS: oh, what notification are you missing?16:33
SpamapSBut I'd swear sometimes I don't see those until just before the request completes16:33
*** JasonCL has quit IRC16:33
SpamapSand really, I think the logging is ok. I am just overwhelmed with logs16:34
SpamapS(Info has the creating message, and that should be plenty)16:34
SpamapSThe real problem is the AZ allocation.16:34
SpamapSI haven't checked my math, but from what I can tell, each node added to a nodeset makes it *far far* harder to have enough ready nodes in the chosen AZ ever.16:35
tobiashmordred: +2 with comment on https://review.openstack.org/#/c/54564416:36
SpamapSBecause everything is chosen at random, min-ready may not spin up any nodes in the AZ you need next, but if there are nodeset_size - 1 nodes in that AZ already, then your job queues while a node spins up.16:36
SpamapSAnd then min-ready may come along and not get enough nodes in that AZ again.. and the cycle repeats until you get a request which happens to pull a node from the very-full AZ.16:37
SpamapSSo yeah.. I think the simplest thing to do is let people turn off the nodeset-in-one-AZ for a particular cloud.16:38
mordredtobiash: that's a good point ...16:38
corvusmordred: when did those 4 patches become the first in the series?  i thought 538099 was the first?16:39
corvusmordred: oh, i think you explained that in the commit msg for 537010...16:40
*** jpena is now known as jpena|brb16:42
dmsimardmordred: replied to your email on zuul-discuss with some additional thoughts16:43
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Use a status code to detect unknown vs. missing tenant  https://review.openstack.org/54587916:45
mordredtobiash: ^^ how about if we did that?16:45
tobiashmordred: looks good, but what do you think about just sending the complete payload with 404?16:50
SpamapScorvus: so yeah just to reset.. I think my logging problem was actually that I should have been watching my INFO-only log and I'd have seen the activity just fine.16:50
tobiashthat way the response still would be json (consistent and also machine readable if you don't have the status code like in a shell pipe)16:50
SpamapSbut my "why are my jobs always queueing for 5 minutes" problem is with AZ stickyness meeting large nodeset sizes.16:51
mordredtobiash: wfm - one sec16:51
corvusSpamapS: ok16:51
SpamapSThe latter I intend to submit a patch for.16:51
clarkbfwiw I'm not a huge fan of the random az by default behavior in nodepool and think we should use the default az unless a list of az's is specified16:51
SpamapSWhich would allow turning it off per provider for those who know the cloud that is pointed at should be fine with cross-AZ jobs.16:52
clarkb(and aiui default could be nova giving you whatever it wants to give you)16:52
dmsimardmordred, andreaf: btw just in case you missed my reply for this issue http://eavesdrop.openstack.org/irclogs/%23zuul/%23zuul.2018-02-18.log.html#t2018-02-18T17:43:57 .. see my answer here: http://eavesdrop.openstack.org/irclogs/%23zuul/%23zuul.2018-02-19.log.html#t2018-02-19T04:50:4116:52
SpamapSclarkb: agreed. I didn't know we did that by default.16:52
SpamapSTypically clouds will configure their default AZ to be the one that is most empty.16:52
clarkbSpamapS: yes it surprised me when we broke citycloud due to this behavior16:52
SpamapSIn GoDaddy's case, we catually require explicit AZ because we did routed networks before neutron did them. ;)16:53
SpamapScatually and actually.16:53
* SpamapS meows16:53
clarkbSpamapS: in that case you could use different providers for each az16:54
clarkbexcept now we won't weight the requests by max-server size anymore right?16:54
corvusclarkb: pool, not provider16:54
clarkbso maybe that doesn't help16:54
clarkboh right pool. But still I'm not sure it helps unless we weigh the requests based on max-server size16:55
corvuswhat is max-server size?16:55
clarkbthe max-servers allowed to be booted by each pool/provider16:56
corvusclarkb: pool again16:56
clarkbso that you could give the smaller az a smaller max-servers value and it would be weighed against the whole. This is how old nodepool worked but I think with the new request mechanism we don't do that anymore16:56
corvusclarkb: why would that be necessary?16:56
clarkbcorvus: because aiui SpamapS'16:56
clarkber16:57
clarkbmy understanding of SpamapS' problem is that he has two az's one is smaller than the other. When the small oen gets a large node request say for 5 nodes it will often block waiting for resources to clear up16:57
clarkbif we weighed handling of requests the bigger az would be more likely to handle that request until they would both be in a situation to block and then it doesn't matter16:57
*** JasonCL has joined #zuul16:58
mordredtobiash: actually, the exception there doesn't take that parameter at all - and takes code and message explicitly16:58
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Use a status code to detect unknown vs. missing tenant  https://review.openstack.org/54587917:01
corvusclarkb, SpamapS: yeah, there are a lot of improvements that can be made to the launchers if they are aware of each other.  right now, the algorithm assumes no information sharing between them.  i think it's worth evolving it in the future to share info, but it needs to be done carefully.17:01
clarkbanother way to appraoch this is to let the cloud put instances in whichever AZ is most appropriate. Though sounds like that won't entirely work for godaddy's cloud17:01
clarkbI do think it is worth figuring out if we can remove the random az by default behavior in nodepool as it is suprising to users of openstack clouds imo17:01
corvusclarkb: i believe the reason we always specify an az is because otherwise we can't guarantee that we'll get nodes in the same az.17:02
clarkbyes, but it is also not how all the existing clients and tools work if you don't specify an az so you will get different potentially unexpected behavior17:02
clarkb(I know I hadn't expected that when we ran into it)17:02
corvusclarkb: sure, but, i mean, if this existed we wouldn't have had to write nodepool. :)17:03
corvuslet's make this argument less circulare17:03
corvuswe need to (in some cases) create 2 nodes that are guaranteed to be in the same az17:03
corvusis there another way to do that?17:03
clarkbinstruct the users that if they have more than one AZ to configure them?17:04
clarkbI think this is a case where ^ is less harm than booting instances in AZs you shouldn't be booting in17:04
clarkbI have yet tofind a cloud where the default az is actually more than one az17:04
clarkbeven though it is theoretically possible17:04
corvusclarkb: can you ask the cloud what the default az is?17:05
clarkbI am not sure17:05
mordredno17:05
corvusthat would be the simplest solution.  perhaps we should start the very long process of adding that as an openstack api feature?17:06
mordredwell - it's not really ... it would be exceptionally hard to do - because an AZ isn't actually a thing17:06
mordredso, with regions, you can query the keystone catalog - because region is an aspect of the endpoint registration record17:07
mordredbut azs exist independently in nova, neutron and cinder17:07
corvusmordred: well, we only care about regions in nova, so it could be a nova api call17:07
corvuser17:08
corvusaz not region17:08
clarkbworth pointing out that az foo in nova may require volumes from az bar in cinder and a different combo of az's won't work. So you'll have to configure nodepool for cases like that anyways (random won't necessarily work)17:08
clarkb(though we aren't doing cinder volumes with nodepool so mostly a non issue)17:09
mordredright - I was actually just checking floating ips to make sure we weren't accidentally doing something incorrect there17:09
corvusclarkb: perhaps a way to use the default az would be to do have nodepool rememer what value or values it has gotten back, and then discard errant nodes or specify that value going forward17:10
*** dmellado has joined #zuul17:10
mordredbut I believe we're fine due to how we create floating ips directly on a port17:10
*** myoung is now known as myoung|bbl17:11
corvusclarkb: that would probably make things better in the citycloud case.  i don't think anything short of information sharing between launchers would help SpamapS's case.17:11
clarkbyes, I think SpamapS' case is exceptional in its corner case qualities :)17:12
*** JasonCL has quit IRC17:12
mordredclarkb: SpamapS is exceptional at producing corner cases17:12
clarkband ya could remember check the az value of the first node that comes back using the default no value az reuqest then explicitly specify that going forward17:12
*** JasonCL has joined #zuul17:21
*** jpena|brb is now known as jpena17:22
*** JasonCL has quit IRC17:25
*** JasonCL has joined #zuul17:37
corvustobiash: did you want to review https://review.openstack.org/537011 ?17:37
tobiashcorvus: oh yes, forgot about that between discussion with mordred and dinner ;)17:39
andreafdmsimard thanks17:40
corvustobiash, mordred: tristanC reviewed previous versions of all those changes positively, so i went ahead and approved all of them except 537011.17:40
mordredcorvus, clarkb SpamapS: I have verified with the nova team that there is no way to find out what the default az is17:41
*** JasonCL has quit IRC17:42
tobiashlgtm17:42
corvusmordred: looks like you're all clear on those 4 patches17:43
mordredcorvus: woot! thanks!17:43
corvusover in #openstack-infra, we found that my memory improvement is using a bit too much cpu, so i'm going to switch to working on a fix to that.17:44
mordredcorvus: ++17:44
mordredcorvus: I'll wait to suggest a scheduler/zuul-web restart on your cpu fix17:44
corvussounds good17:45
mordredanybody remember who it was that had a console app that shows zuul status? was that dansmith?17:53
mordredclarkb: ^^?17:53
clarkbI think harlowja and dansmith had their own17:54
*** sshnaidm is now known as sshnaidm|off17:54
*** JasonCL has joined #zuul17:59
*** tosky_ is now known as tosky18:04
*** JasonCL has quit IRC18:05
*** tosky has quit IRC18:07
dmsimardmordred: harlowja yeah18:12
dmsimardI have a link for it somewhere...18:12
*** hashar_ is now known as DockerIsA_Sin18:12
dmsimardmordred: https://github.com/harlowja/gerrit_view/blob/master/README.rst#czuul18:14
*** openstackgerrit has quit IRC18:18
*** jpena is now known as jpena|away18:19
*** JasonCL has joined #zuul18:26
mordreddmsimard: thanks. pull request submitted18:28
dmsimard\o/18:29
*** chrnils has quit IRC18:31
*** JasonCL has quit IRC18:32
*** DockerIsA_Sin is now known as hashar18:33
*** openstackgerrit has joined #zuul18:34
openstackgerritMerged openstack-infra/zuul master: Remove .json suffix from web routes  https://review.openstack.org/53701018:34
openstackgerritMerged openstack-infra/zuul master: Port per-change status to zuul-web  https://review.openstack.org/53590318:40
openstackgerritMerged openstack-infra/zuul master: Re-enable test of returning 404 on unknown tenant  https://review.openstack.org/54564418:40
openstackgerritMerged openstack-infra/zuul master: Add /info and /{tenant}/info route to zuul-web  https://review.openstack.org/53701118:40
openstackgerritJames E. Blair proposed openstack-infra/zuul master: Avoid creating extra project schemas  https://review.openstack.org/54595318:45
clarkbcorvus: is ^ the performance fix?18:46
corvusclarkb: yep18:46
corvusthat was not what i was expecting, fwiw.18:46
clarkbok reviewing now, then breakfast18:46
corvusbut in my local tests, that restored (and actually slightly improved with the extra changes i put in there) the parse time of a large configuration to the same as before the whole rework stack18:47
corvusthe difference was about double (ie, https://review.openstack.org/545448 caused the time to double, and then https://review.openstack.org/545953 halved it again).  that doesn't match the 6x difference we see in openstack-infra.  i can't really account for that other than to say that my local config doesn't mirror it exactly.18:49
corvusso i think we should try that in prod to see whether that's a complete or partial fix.18:49
corvusi started a local test run when i pushed it up -- it looks like it's going to fail unit tests, but i think it may only be due to test infrastructure issues.  i think it's worth reviewing it for substance now while i fix up the test failures.18:51
AJaegercorvus: did you see mordred's change https://review.openstack.org/#/c/545610 ? That touches similar lines as yours...18:51
clarkbcorvus: substance wise it looks fine to me18:51
corvusAJaeger: i like mordred's change -- but let's rebase it after this one since it's not as urgent18:52
fungiyeah, i approved it before i saw you mention local test failures. curious to see where it breaks18:57
openstackgerritJames E. Blair proposed openstack-infra/zuul master: Avoid creating extra project schemas  https://review.openstack.org/54595318:58
openstackgerritJames E. Blair proposed openstack-infra/zuul master: Allow test_playbook to run long  https://review.openstack.org/54595518:58
dmsimardDo we have a strategy for updating Ansible in Zuul without potentially breaking everyone's jobs ?18:59
corvusfungi, clarkb, mordred: ^ local test failures fixed.  i also stacked it on top of a test timeout bump.18:59
fungithanks, i was wondering whether the change in timeout handling was to blame for the rise in unit tests timeouts for zuul changes18:59
corvusdmsimard: no, i think what we're going to do is land the 2.4 update, deal with breakage, and after that work out a way to support multiple versions.19:00
corvusdmsimard: (then, of course, changing the default version will break everyone, but that's an easier fix)19:00
fungibut i guess the tests themselves outgrew the timeout some too19:00
corvusfungi: yeah, i think that particular test timeout is due to our adding new jobs to that test for the post-timeout and hostvars19:00
fungik19:01
fungii see, that's the per-test timeout, not the job/playbook timeout19:01
dmsimardcorvus: yeah Ansible unfortunately doesn't have a very good reputation at being "stable" (at least in the semver sense), typically breaking roles/playbooks. One has to be careful about using the word "regression" in #ansible-devel, often times it's a fix that broke an "unexpected behavior" or something like that :)19:01
dmsimardSince we have hundreds of jobs, upgrading Ansible might be quite a sensitive task19:02
*** harlowja has joined #zuul19:03
fungiand no sane way to exercise them all against new ansible versions other than just taking the plunge and seeing what breaks19:04
clarkbcorvus: the two changes are approved now19:06
*** nhicher has quit IRC19:06
*** myoung|bbl is now known as myoung19:07
*** nhicher has joined #zuul19:07
dmsimardfungi: that's scary19:08
dmsimardfungi: I wonder if there could be a job that'd run a job in another version of Ansible somehow19:08
clarkbdmsimard: aiui the idea is that with zuul testing ansible we can avoid some of those issues, but yes. Upgrading zuul's ansible likely to be an adventure19:08
dmsimardclarkb: I'm not worried so much about Zuul than I am about the users' jobs19:09
clarkbdmsimard: sure thats what I mean. the ansible zuul runs19:09
corvusdmsimard: well, i mean, if you do that, you've solved the problem of running multiple versions of ansible.  which we plan to do.19:09
corvusso i mean, i guess we could stay on 2.3 until we do that.19:09
clarkbat the very least we need to publish what version of ansible a job ran under to aid in debugging (I think mordred was going to add that, not sure if it got added)19:09
corvusbut i feel like we're early enough in the process that we (and i mean the collective zuul's of the world here, not just openstack) can deal with a cutover to 2.4.19:10
Shrewspretty sure the logging of ansible version is already landed19:10
corvusthen solve for multiple versions of ansible later19:10
dmsimardclarkb: That's https://review.openstack.org/#/c/514489/ and https://review.openstack.org/#/c/532304/ which were blocked by a weird include_role issue, I think that's fixed now. I'll look.19:11
fungidmsimard: we could set up some system to run preliminary builds of select jobs against newer ansible or something, but my point was we have tons of jobs which run infrequently under specific situations and have side effects we cant fully exercise outside production without a lot of extra effort to do things like set up our own mock pypi and whatnot (and even that doesn't 100% guarantee it'll accurately19:11
fungirepresent a production run)19:11
dmsimardOH, no, it was a random fedora26 failure I could not understand19:11
dmsimardI'll debug that.19:11
Shrewsdmsimard: oh, that still hasn't landed. thought it had19:12
dmsimardfungi: yeah I don't have any great ideas right now... at least as far as ARA is concerned, there are different integration jobs which tests that ara works with different versions of Ansible (2.2, 2.3, 2.4, devel). We don't really have the luxury of doing that with Zuul and the hundreds of jobs our users have19:13
dmsimardand then again, it needs to be tested under *Zuul's* Ansible19:13
openstackgerritDavid Moreau Simard proposed openstack-infra/zuul-jobs master: Revert "Revert "Add zuul.{pipeline,nodepool.provider,executor.hostname} to job header""  https://review.openstack.org/51448919:14
openstackgerritDavid Moreau Simard proposed openstack-infra/zuul-jobs master: Add Ansible version to job header  https://review.openstack.org/53230419:14
*** jpena|away is now known as jpena|off19:14
clarkbis the ansible version something we want the jobs to provide or something that zuul should just write out?19:17
clarkb(I worrythat that is a feature that everyone should have and if it is in job ocnfig many will miss it?)19:17
corvuswell, this is the way we provide it to everyone -- a role in zuul-jobs that anyone can (should) use19:18
clarkbya I guess the intent is for people to be using these roles. Maybe info about these roles can go into the bootstrapping doc when writing a base job19:20
*** JasonCL has joined #zuul19:20
corvusclarkb: yeah, that's sort of where the zuul-from-scratch doc leaves off right now.  but i believe we're all agreed it should be included there.19:22
*** JasonCL has quit IRC19:25
dmsimardclarkb: I believe we discussed in the past that if users need advanced customization of their Ansible (ex: OSA might need a specific version of Ansible to test things.. different modules/callbacks, etc.), they would need to rely on installing a "nested" version of Ansible and running with that19:47
corvuswell, that's the story now.  but just so we're clear: after we upgrade to ansible 2.4, we are going to figure out a way for zuul to support more than one version of ansible, and people will be able to specify which version of ansible to run in their jobs.19:51
dmsimardcorvus: that sounds amazing -- hard, but amazing19:51
corvusit will not be easy, but it is required in order for us to achive our objective of people using the same ansible in test and production.19:52
rcarrillocruzwow19:52
rcarrillocruzThat would be amazing indeed19:52
*** JasonCL has joined #zuul19:53
clarkbseems like the biggest hurdle there is the security monkey patching?19:53
fungiseems like we're going to become increasingly aware of the nuanced differences between each minor ansible release19:53
clarkbotherwise you can just install a virtualenv for each ansible and fork ansible out of them?19:54
fungiansible seemed amenable to upstreaming some of the security improvements, right?19:54
fungii mean, after the got over the "what? you're going to run arbitrary user-supplied playbooks on your servers!?!"19:55
fungiafter they19:55
dmsimardclarkb: yeah, we need a clean way/interface to patch the stuff that we do in Zuul (security is one, streaming is another, etc.)19:57
corvusi actually think that's not going to take much more work.  we already need to do work for each new version of ansible.  so it's *mostly* a matter of not deleting the old versions, but with some minor-to-moderate extra maintenance work to deal with stable point releases.  most of that can be automated (this stuff doesn't generally change a lot between point releases)20:00
corvusthe thing we have to sort of invent is how to have multiple versions of ansible installed that works with all the ways people want to deploy zuul20:01
*** JasonCL has quit IRC20:02
dmsimardI have some quite dirty code to handle forward/backward compat in ARA20:10
dmsimardi.e, https://github.com/openstack/ara/blob/dd37d4566bb4a17871e55e4ba50e17b46079efc2/ara/config.py#L41 "if LooseVersion(ansible_version) >= LooseVersion('2.3.0')" etc20:10
clarkbcorvus: my naive assumption here is that zuul can sort of internally use the new python3 virtualenv stdlib tooling20:10
dmsimardIf we end up finding a better way to handle this sort of stuff in Zuul, I'm all ears20:11
clarkbhttps://docs.python.org/3/library/venv.html#api20:11
dmsimardclarkb: so the executor would bootstrap the appropriate version of Ansible after creating the bwrap or something like that ?20:12
clarkbdmsimard: I think it can do it before bwrapping then bind mount into it and just keep a bunch of premade ansible installs20:12
clarkbbut ya something like that20:12
clarkbprobably make them on demand the first time20:12
dmsimardah I see what you mean.. yeah, that could make sense20:13
corvusclarkb: ah yes, that sounds like just the thing.20:20
mordredclarkb, corvus: yah - I like the venv idea and think that's a good way to move forward on that20:41
corvusdmsimard: what's the status of https://review.openstack.org/543633 ?20:43
corvusdmsimard: (that's the fix for https://storyboard.openstack.org/#!/story/2001329 right?)20:44
mordredcorvus: so - I just noticed (because I looked at docs online and the docs lied to me) we have an upper-bounds cap on aiohttp of <3 ... but 3.0.1 is out now20:57
corvusmordred: yeah, they broke something, commit should have info20:58
mordredcorvus: it's mostly the same, although there is at least one thing that be nice: https://aiohttp.readthedocs.io/en/stable/web_advanced.html#graceful-shutdown20:58
corvusmordred: https://review.openstack.org/54354620:58
mordredcorvus: yah - that was what I was about to say - glad to see that was the reason for th ecap20:59
mordredso we can update to 3.0 when we've migrated to bionic20:59
corvusmordred: oh, any chance 3.0.1 fixed the issue?20:59
Shrewsodd. i wonder what 3.5.3 has that they needed21:00
SpamapSclarkb: "my understanding of SpamapS' problem is that he has two az's one is smaller than the other." -- that's a misunderstanding. I have 3 AZ's, and they are all relatively the same size.21:00
dmsimardcorvus: I rebased it today, it's on top of my todo list right now. I plan on reproducing it with https://review.openstack.org/#/c/530642/12/playbooks/devstack-tempest.yaml21:00
dmsimardcorvus: it's probably the fix, yes, but I want a test that is able to demonstrate that it actually does fix it21:01
dmsimardcorvus: er, I rebased something else that was *not* that -- I don't actually need to rebase that one.21:01
clarkbSpamapS: ah ok, I'm not sure what nodepool would do differently then? maybe weight not based on max-servers but by available quota?21:05
corvusdmsimard: can you add a test for it in zuul?21:09
dmsimardcorvus: That's what I attempted to do with the functional integration test https://review.openstack.org/#/c/543633/2/playbooks/zuul-stream/functional.yaml but it doesn't seem to reproduce the issue, I'm doing some testing locally to try and understand why21:11
mordredclarkb: I think what I was hearing from SpamapS yesterday is that we do the az randomization before we check for ready nodes - so he had an az with ready nodes, but the az balancer chose a different az and proceeded to spin up the nodes there21:12
clarkbah21:13
corvuser, that's not how that works21:14
corvusif there's a ready node, we'll grab it and use its AZ to spin up any more nodes if necessary21:14
mordredoh - yah. I just re-read21:15
pabelangerI'm just looking at https://review.openstack.org/545953/ and py35 seems to have failed21:15
mordredin his case it was that min-ready was also working on spinning up nodes and he was near capacity21:15
corvusspectacularly.. i'll look into that21:16
mordredpabelanger: oh - wow - that's a lot of red21:16
pabelangerlooks like threads didn't shutdown propelry21:17
pabelangerproperly*21:17
clarkbbut we also can't really fix that without having handler coordination like what corvus described I think21:18
clarkbbecause it requires handlers to have a rough idea if any other handler is in a better spot to handle a request21:18
mordredclarkb: yah21:18
corvusthose tests work in isolation...21:18
corvusi'm wondering if one test went awry and messed up the state for other tests21:19
openstackgerritMerged openstack-infra/zuul master: Allow test_playbook to run long  https://review.openstack.org/54595521:19
corvusoh i think i know which :)21:21
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Hack for test_delete_now rare failures  https://review.openstack.org/54598221:22
corvusmordred: is there a way i can have ttrun do 2 different classes?21:22
Shrewsi have been pondering the test_delete_now failures today. i'd like to see the above hack go in until a real solution can be generated21:23
corvusmordred: like i want to run: "ttrun -epy35 tests.unit.test_model" and ttrun -epy35 tests.unit.test_requirements" in one solution?21:23
corvusShrews: oh is *that* the thing we added recursive locking for?21:23
corvusShrews: that merged in kazoo, we can probably use it if we can remember how it was going to solve it.21:23
Shrewscorvus: i spent a great deal today trying to see how that could help (i totally forgot the solution we discussed)21:24
Shrewscorvus: i did not see how it would help at all since ReadLock and WriteLock still create znodes under the thing we intend to delete21:24
corvusShrews: heh, i may have to ask eavesdrop to figure out what we were thinking21:24
mordredcorvus: not really, no - I think stestr can though21:25
clarkbtox -epy36 -- '(TestClass1|TestClass2)' ?21:25
mordredcorvus: actually, I think testr itself can too ... ^^ like that21:26
mordredyou may also want to add --parallel=1 or whatever the concurrency arg is21:26
openstackgerritJames E. Blair proposed openstack-infra/zuul master: Avoid creating extra project schemas  https://review.openstack.org/54595321:26
mordredconcurrency21:26
mordredtox -epy35 -- --concurrency=1 'tests.unit.test_(model|requirements)'21:27
mordredshould do the trick21:27
corvusoh ok, well i just started running the whole squite locally21:27
mordredk21:27
corvusthough i'm pretty sure that should fix it21:27
mordredcorvus: should we go ahead and re-promote in gate?21:29
corvusmordred: i should have a local test run in just a minute21:30
clarkbdoes nothing have to call readConfig?21:30
corvusclarkb: the scheduler now calls both read and load config21:30
clarkbwas that git added?21:31
clarkb(I'm not seeing it but also could just be looking poorly)21:31
corvusclarkb: i may need more context for your question21:31
clarkbI'm not seeing where the scheduelr calls read config. But I'm about to fetch change locally and grep better21:31
mordredcorvus: gerrit is showing something strange in the interdiff21:31
corvusclarkb: you're talking about what change?  mordred's config changes?21:32
mordredcorvus, clarkb: that readConfig  change is from here: https://review.openstack.org/#/c/545610/ ... but gerrit is showing it in the 'show me differences between patchset 2 and 3" view21:32
clarkb545953 I think I'm just not seeing an existing api due to the restricted scope of that change21:32
corvusmy latest patchset is a fix+rebase21:32
mordredoh - it's a rebase - so it's just the werid gerrit rebase thing21:32
corvussorry, i don't normally like to do that, but i did so this time because the test failures were hard to repro locally and i thought it might be due to a change in the tree21:33
mordredand sorry - it wasn't from that patch, it was from a different one21:33
corvusi can un-rebase that if you'd like21:33
mordredcorvus: ++21:33
mordredcorvus: nah - I was just confused because I thought it was content from an un-landed change :)21:33
corvusi only made one change, and the rebase should (in retrospect) not be necessary.  the change is to add a cleanup in test_model.21:33
clarkbcorvus: ya cleanup in test_model makes sense to me21:33
mordred++21:34
corvus(so probably easiest to just look at the actual diff from base and know that single line is the only change in the latest patchset)21:34
clarkbok I've approved it21:35
clarkbgit log -p straightened me out21:36
mordred\o/21:36
*** threestrands has joined #zuul21:37
corvuslocal tests pass, i'll re-enqueue in gate21:37
fungithe 545956 update included a rebase i guess21:38
fungiahh, scrollback confirms21:39
corvusdmsimard: 543633 has a unit test in it -- does it not fail without the fix?21:39
corvussorry, i should have undone that but forgot21:39
SpamapSSo .. sort of?21:39
* SpamapS is laggy today sorry21:40
dmsimardcorvus: that's not unit, that's integration -- the reason I'm confused is that this exact same patch used to fail if you look at the prior history in https://review.openstack.org/#/c/504238/21:40
corvusdmsimard: oh, it doesn't have a unit :(21:40
dmsimardcorvus: it suddenly stopped failing and I was never able to understand why21:40
SpamapSMy issue is that sometimes I routinely have 10+ ready nodes in an AZ, and because requests randomly choose a ready node, instead of a ready set, it's entirely possible we never choose one of those 10, and thus, end up needing to spin up nodeset_size - 1 nodes on every single request.21:41
mordreddmsimard: maybe point-release of ansible?21:41
dmsimardI'm working on setting up a reproducer right now, mordred's test-logs ( https://github.com/openstack-infra/zuul/blob/master/tools/test-logs.sh ) is quite useful -- I'll send some improvements for it21:41
SpamapSI have min-ready: 15, and a nodeset size of 5, and 3 AZ's.21:41
clarkbSpamapS: so we probably want to make ready nodes a heap sorted by nodes per az21:41
clarkbSpamapS: and pull off in that order21:41
mordreddmsimard: \o/ I'm useful :)21:42
SpamapSI had hoped that meant we'd have a high chance of picking an AZ with an entire set in it.21:42
SpamapSWhat actually happens is, we pick the right one almost never.21:42
corvusmordred, dmsimard: well part of mordred's work on logging is that we should not need test-logs anymore21:42
corvuswe need *actual* tests for this stuff, and i believe mordred is adding those21:42
dmsimardmordred: The fix for the issue we're talking about landed in 2.4.3 so unless Zuul is suddenly using 2.4.3, something's up21:42
corvusso let's not spend too much time on improving test-logs21:42
mordredcorvus: I don't think we're looing at improving it at all21:43
mordredcorvus: I think he's just using it to help figure out why things suddenly started working21:43
dmsimardcorvus: sure -- it's a way for me to run the test playbooks with the callbacks and stuff without having access to a sandbox/test zuul environment21:43
dmsimardmordred: yeah that21:43
corvusmordred: dmsimard said he was improving it21:43
SpamapSIt gets really bad when 2 or 3 req's come in at once.21:43
dmsimardcorvus: 2 minutes worth of improvement to make my life easier :P21:43
mordredcorvus: not related to that- do all of the parsers do self.schema = self.getSchema now?21:44
mordredcorvus: nm - they don't21:44
corvusmordred: i think only tenantparser doesn't... but i think it probably could?21:44
SpamapSclarkb: I think we'd also need to make min-ready's do the same in reverse. When a min-ready req's comes in, choose the AZ with the least ready nodes.21:45
corvusjust wasn't important to me earlier because it's only called once either way21:45
corvusdmsimard: so what happens if you add that test without the fix?21:45
SpamapSOr, just allow disabling az-sticky21:45
mordredcorvus: I actually tried generalizing the schema bits when I made the base class patch - but it got to weird and I figured I'd do it as a follow up21:45
mordredcorvus: (there was one of the classes that called getSchema with arguments)21:45
SpamapSTake nodeset_size nodes from the provider and ignore AZ.21:45
dmsimardcorvus: I'll have an answer in a few minutes21:45
corvusmordred: yeah, nodeset has an anonymous argument, so we store 2 variations21:46
mordredyah. that was the one that made it seems less nice21:46
corvusmordred: could throw a layer of indirection in there... self.createSchema() and override that in nodeset....21:47
mordredcorvus: oh - wait - you've solved the issue I had already I think21:48
mordredyup.21:49
mordredcorvus: thanks! your latest patches remove the issue I had putting that in the base :)21:49
corvusyay21:50
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Extract an abstract base Parser class  https://review.openstack.org/54561021:53
mordredcorvus: there ya go ^^21:53
clarkbSpamapS: in your config you have a single pool for all three AZs and nodepool is randomly assigning nodes to each AZ right?21:56
*** hashar has quit IRC21:56
dmsimardcorvus: can't reproduce it locally again... with or without the patch. I'll send a dummy patch with a generic version of what andreaf reported last time it occurred.21:57
andreafmordred I started documenting a bit https://review.openstack.org/#/q/status:open+project:openstack-dev/devstack+branch:master+topic:ansible_docs21:58
corvusandreaf: \o/22:00
mordredhttp://logs.openstack.org/52/545952/1/check/build-openstack-sphinx-docs/478cade/html/roles.html <-- looks great!22:00
clarkbmeeting time right?22:00
corvusyep, meeting time in #openstack-meeting-alt22:00
mordredandreaf: there's a similar thing for jobs too22:00
mordredandreaf: .. zuul:autojobs:: IIRC22:01
mordredandreaf: the second patch - I think you missed a git add22:02
andreafmordred: heh ok thanks22:14
*** JasonCL has joined #zuul22:16
mordredandreaf: exciting to see that though22:17
openstackgerritMerged openstack-infra/zuul master: Avoid creating extra project schemas  https://review.openstack.org/54595322:18
*** openstackgerrit has quit IRC22:18
*** JasonCL has quit IRC22:20
*** openstackgerrit has joined #zuul22:38
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Use a status code to detect unknown vs. missing tenant  https://review.openstack.org/54587922:38
openstackgerritMonty Taylor proposed openstack-infra/zuul master: Remove Paste from the dependencies  https://review.openstack.org/54600022:38
mordredcorvus, pabelanger, clarkb: ^^ 545879 fixes an api mismatch - and the follow up just cleans up Paste - neither are urgent, but both should be easy +As22:54
*** JasonCL has joined #zuul22:54
dmsimard(╯°□°)╯︵ ┻━┻22:57
dmsimardcorvus, mordred: https://review.openstack.org/#/c/545994/ didn't trigger the json issue22:57
mordreddmsimard: :(22:59
mordreddmsimard: maybe a 2.3 point release fixed the issue22:59
dmsimardmordred: as far as I know, the fix is https://github.com/ansible/ansible/commit/c30ee42fe1f0a9666a90f4d63121780f2a186c54 and it didn't land until 2.4.3, it's not backported to 2.3.x which is what we're using23:02
mordreddmsimard: yah23:02
dmsimardmordred: yeah 2.3 has the bug: https://github.com/ansible/ansible/blob/stable-2.3/lib/ansible/executor/task_executor.py#L45923:03
dmsimardthis is annoying23:03
dmsimardandreaf: do you have a some sort of reference on the failure from the include_role in https://review.openstack.org/#/c/530642/12/playbooks/devstack-tempest.yaml ?23:04
dmsimardandreaf: which job ran that code ?23:05
andreafdmsimard devstack-tempest and tempest-full and tempest-full-py323:07
*** JasonCL has quit IRC23:07
dmsimardandreaf: I can't find any signs of the error I'm looking for, could you point it out to me ?23:10
corvusdmsimard: well, according to the bug in storyboard, we saw the issue in build e4fdb1c004fe47478a5468bb6318ad6623:23
corvusdmsimard: http://logs.openstack.org/42/530642/11/check/tempest-full/e4fdb1c/job-output.txt.gz#_2018-02-09_18_10_40_31496823:24
corvusdmsimard: there is no output in the console log about that23:24
corvusthe behavior is just start followed by an immediate end23:25
corvusdmsimard: and you'll note there's no entry in ara for the devstack-tempest playbook23:25
corvusdmsimard: it's patchset 11 that hit that error, not the latest version.23:26
corvusdmsimard: patchset 11 doesn't use include_role... it actually looks like it's invalid ansible, but that should have been reported.23:27
corvusdmsimard: that's in executor-debug.log.10.gz on ze03 if more logs would help23:31
dmsimardcorvus: I'll check it out, thanks23:45

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!