Monday, 2018-01-15

openstackgerritTristan Cacqueray proposed openstack-infra/nodepool feature/zuulv3: config: add statsd-server config parameter  https://review.openstack.org/52896900:33
*** dtruong2 has joined #zuul01:27
*** threestrands_ has joined #zuul01:40
*** threestrands has quit IRC01:43
*** dtruong2 has quit IRC01:45
*** dkranz has quit IRC02:03
*** jappleii__ has joined #zuul03:12
*** jappleii__ has quit IRC03:13
*** jappleii__ has joined #zuul03:13
*** threestrands_ has quit IRC03:15
*** JasonCL has quit IRC03:48
*** bhavik1 has joined #zuul04:10
*** bhavik1 has quit IRC04:16
*** xinliang has quit IRC05:46
*** xinliang has joined #zuul06:01
*** xinliang has quit IRC06:01
*** xinliang has joined #zuul06:01
openstackgerritTristan Cacqueray proposed openstack-infra/zuul feature/zuulv3: ansible: honor command no_log module attribute  https://review.openstack.org/53350906:49
*** jappleii__ has quit IRC07:02
*** saop has joined #zuul07:39
saophello all07:39
saopI setup zuul v3 in my CI, and zuul-web is not showing tenants.html page, any reason , what i need to do?07:52
tristanCsaop: have you checked zuul/web/static/README ?08:06
saoptristanC, yes08:06
saoptristanC, I did same thing in apache config08:07
saoptristanC, and it also showing in status.html page status.json not found08:07
saoptristanC, Thanks i found one error08:10
saoptristanC, it solved08:10
*** hashar has joined #zuul08:13
openstackgerritTristan Cacqueray proposed openstack-infra/zuul-jobs master: Add buildset-artifacts-location and fetch roles  https://review.openstack.org/53067908:25
*** sshnaidm is now known as sshnaidm|afk08:41
*** jpena|off is now known as jpena08:49
*** jaianshu has joined #zuul08:59
*** xinliang has quit IRC09:52
*** sshnaidm|afk has quit IRC09:55
*** xinliang has joined #zuul10:05
*** xinliang has quit IRC10:05
*** xinliang has joined #zuul10:05
*** sshnaidm|afk has joined #zuul10:52
*** sshnaidm|afk is now known as sshnaidm10:58
*** electrofelix has joined #zuul11:06
*** JasonCL has joined #zuul12:05
*** jkilpatr has joined #zuul12:14
*** jkilpatr has quit IRC12:19
*** jkilpatr has joined #zuul12:19
*** saop has quit IRC12:30
*** jpena is now known as jpena|lunch12:38
*** nhicher has joined #zuul12:57
ShrewsFYI, I am unavailable today.13:02
*** openstackgerrit has quit IRC13:18
*** jpena|lunch is now known as jpena13:29
*** jaianshu has quit IRC13:31
*** lennyb_ has joined #zuul13:33
*** lennyb_ has quit IRC13:34
*** rlandy has joined #zuul13:41
*** jkilpatr has quit IRC14:03
mhuHello, is it possible to find out the nodepool id(s) of the node(s) used by zuul to build a specific job? I've looked at status.json but I can only find references to zuul executors. The use case would be, for example, given that I have access to nodepool, I could hold a job's node(s) before the build completes so I can investigate the build env post mortem14:04
*** jkilpatr has joined #zuul14:06
*** JasonCL has quit IRC14:11
*** JasonCL has joined #zuul14:11
*** JasonCL has quit IRC14:11
*** JasonCL has joined #zuul14:12
*** JasonCL has quit IRC14:12
*** JasonCL has joined #zuul14:12
*** JasonCL has quit IRC14:19
*** JasonCL has joined #zuul14:24
*** JasonCL has quit IRC14:25
lennybmhu, we usually check in Jenkins where job is executed and then hold it14:40
mhulennyb, I forgot to mention I was talking about zuulv314:40
*** jkilpatr has quit IRC14:54
*** jkilpatr has joined #zuul14:55
lennybmhu, sorry, I cant help you with this one yet14:58
corvusmhu: check out the 'zuul autohold' command15:12
corvusmhu: in v3, the hold functionality has been moved to zuul15:13
*** dkranz has joined #zuul15:14
mhucorvus, ok, looking, thx15:19
*** AJaeger has joined #zuul15:34
AJaegerzuul team, the zuul tests fail on zuul-jobs, see http://logs.openstack.org/36/531936/6/check/tox-py35-on-zuul/f5f94b0/job-output.txt.gz#_2018-01-15_15_25_49_59333115:34
*** dkranz has quit IRC15:50
*** dkranz has joined #zuul16:03
corvusAJaeger: https://github.com/aio-libs/aiohttp/issues/2662 is the issue16:28
corvuswe may need to add yarl to requirements ahead of aiohttp until they fix that16:29
AJaegercorvus: I see ;(16:31
corvusadding "yarl>=0.11,<1.0" above aiohttp in requirements.txt should do it i think16:32
corvusi'll do that real quick16:32
*** openstackgerrit has joined #zuul16:34
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Temporarily pin yarl while aiohttp is broken  https://review.openstack.org/53372616:34
corvusAJaeger: ^16:34
AJaegerthanks, corvus16:35
*** jpena is now known as jpena|brb16:44
*** jpena|brb is now known as jpena17:27
clarkbpabelanger: testing this little bit of code is proving to be difficult17:37
clarkbpabelanger: the problem is we hand node boot failures just fine and there is a bit of an assumption that all external failures would happen there. Problem is when the handler itself fails (due to quota or whatever) is the place where we are fixing17:37
clarkband there isn't any way to order these node request handler run_handlers in a reliable way that I have found17:38
clarkboh wait there is a paused_handler. Maybe that is the magic I need17:39
clarkboh except thats an actual state and not just a flag17:39
clarkbI think there may be a bug in quota checking too17:43
clarkbOpenStackNodeRequestHandler._waitForNodeSet() checks if it should pause itself after we check the quota in run_handler() so we can bail out early if we are at quota instead of pausing17:45
clarkbcorvus: ^17:45
clarkbthough I'm not entirely sure how the state machine would properly represent that situation and keep jobs happy17:46
*** jkilpatr has quit IRC18:05
*** jpena is now known as jpena|off18:18
*** jkilpatr has joined #zuul18:18
*** kmalloc has joined #zuul18:19
*** corvus is now known as jeblair18:21
*** jeblair is now known as corvus18:21
*** myoung|pto is now known as myoung18:39
mnasercorvus: in oom related stuff, im not sure if this data point helps but i approved a large set of patches today in puppet just before things fell through (and i wonder if the first time a few days  ago it fell over was because of the large amounts of patches pushed at once)18:45
mnaserhttps://review.openstack.org/#/q/status:open+topic:use_journal18:45
corvusmnaser: it should only significantly affect zuul memory usage if it changes zuul configuration, those don't look like they should, so probably not a direct cause18:46
mnasercorvus: yeah did notice that they were not zuul reconfigs but i figured i'd let you know as a data point just in case18:46
corvusmnaser: thx18:49
*** corvus is now known as jeblair18:58
*** jeblair is now known as corvus18:58
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Temporarily pin yarl while aiohttp is broken  https://review.openstack.org/53372619:04
pabelangerhttps://review.openstack.org/511986/ and https://review.openstack.org/532615/ cleans up old zuulv3-dev.o.o and zuul-dev.o.o servers in site.pp19:07
*** electrofelix has quit IRC19:17
openstackgerritDavid Moreau Simard proposed openstack-infra/zuul-jobs master: Revert "Revert "Add zuul.{pipeline,nodepool.provider,executor.hostname} to job header""  https://review.openstack.org/51448919:31
corvuslet's try to keep openstack-infra specific topics in #openstack-infra19:37
SpamapSOy, porting pyre2 to Py3 is going to be a bit of a bear19:50
SpamapSunicode.. bytes... lots of assumptions. :-P19:50
SpamapSI'm about 30% of the way through I think, but having to double back to make it py2+py3. Also the original author is MIA, so we may have to adopt it if we want to use it. But it might be worth it.. really nice RE library.19:51
SpamapSEven better would be if we could convince *python* to adopt it :)19:51
*** jkilpatr has quit IRC19:55
openstackgerritClark Boylan proposed openstack-infra/nodepool feature/zuulv3: Add test_launcher test  https://review.openstack.org/53377119:56
clarkbpabelanger: ^ thats a half working test for the thing you wanted. It doesn't work reliably enough to merge it though and I really have no idea how to make it reliable without putting a bunch of test specific code in the launch handler.19:57
clarkbThe problem here is we basically need to pause the main run loop of the second provider until after the first provider fails19:58
clarkbBut using the built in pause mechanism doesn't seem to work because a paused request doesn't update its config (so if you use max-servers to pause a handler it won't see any max-server updates to unpause it)19:58
clarkbthe other I dea I have that I haven't had a chance to test yet is getting both providers to pause due to max-servers being met. Then make a third node request. Delete provider1's existing node now there is room but it will fail first, then delete provider2's existing node which should succeed20:01
tobiashSpamapS: what's the problem with regular expressions in python?20:02
clarkbthe problem with this is that third request could be in a paused state on either provider. Which means I'm back to the same problem as before of other provider getting it first20:03
clarkbShrews: ^ any ideas on how to test this?20:06
clarkbthe real problem here seems to be that paused handlers are in a super exceptional state where we can't really affect them20:07
corvusclarkb: Shrews is afk today20:07
clarkbanyways, I believe that code does work as intended. The test passes when the assignHandlers race goes in the order we want20:07
clarkbit fails when it goes in the other order20:08
SpamapStobiash: try timing the regex '(x+x+)+y' against an ever increasing string of 'x' ...20:08
SpamapStobiash: around 100 x's the permutations exceed seconds until the sun will supernova ;-)20:09
SpamapSre2 doesn't allow backrefs20:10
SpamapSBut it's a C lib and needs glue. The glue is 2.7 only at the moment20:10
tobiashah, ok20:13
tobiashthat would be a great way to dos zuul...20:13
openstackgerritMerged openstack-infra/zuul-jobs master: Use setup.py if we can't get the name from setup.cfg  https://review.openstack.org/53193620:14
openstackgerritClark Boylan proposed openstack-infra/nodepool feature/zuulv3: Add test_launcher test  https://review.openstack.org/53377120:14
clarkbnow with missing file add20:14
dmsimardbtw I formalized something I may have discussed before -- needing a feature to "autokeep" build directories: https://storyboard.openstack.org/#!/story/200147020:16
SpamapStobiash: have done it to my stage env. 😋20:19
SpamapSwill hopefully have re2 ported soon, but with author MIA.. it gets complicated20:20
dmsimardIs the procedure to enable/disable keep on a live executor documented somewhere ? I think it can be enabled through the command socket without restarting the executor but I can't find where20:22
SpamapSIt can and I think it is in the help for zuul-executor20:23
dmsimardyeah I see the "keep" and "nokeep" commands, as well as the --keep-jobdir arg20:26
dmsimardJust not 100% positive that running "zuul-executor keep" and "zuul-executor nokeep" (while zuul-executor is running) does what I expect20:26
dmsimardI need to add RAM to my home lab so I can set up zuul/nodepool :/20:27
tobiashdmsimard: it does what you expect20:33
tobiashthe same works with verbose20:33
dmsimardtobiash: ok, so if I want to troubleshoot something, I can safely use "zuul-executor keep" to toggle keep on and then when I'm done "zuul-executor nokeep".20:35
tobiashdmsimard: yes20:38
dmsimardGreat, thanks.20:38
tobiashbut you have to delete the dirs manually afterwards20:38
*** jkilpatr has joined #zuul20:53
*** sshnaidm is now known as sshnaidm|afk21:23
rcarrillocruzheya folks21:24
rcarrillocruzso looking at post queue on zuul github21:24
rcarrillocruzchecking docs, it's not clear to me what event is triggered when a PR is merged21:24
rcarrillocruzi.e. what's the gerrit equivalent ref-updated on GH21:26
*** jappleii__ has joined #zuul21:32
openstackgerritMerged openstack-infra/zuul-jobs master: Add consolidated role for processing subunit  https://review.openstack.org/52933921:33
corvusrcarrillocruz: oh i think it's 'push'21:36
rcarrillocruzreally? gah, was assuming it would be an action on the pull_request event21:39
rcarrillocruzk thx corvus21:39
dmsimardZuul meeting in 5 minutes21:55
clarkbwoo I think I finally figured out a test for this cloud error issue22:00
clarkbjust in time for the meeting22:00
openstackgerritClark Boylan proposed openstack-infra/nodepool feature/zuulv3: Add test_launcher test  https://review.openstack.org/53377122:07
clarkbI think ^ should work now22:08
clarkbthat was brain melting22:08
*** dkranz has quit IRC22:17
*** hashar has quit IRC22:18
dmsimardRight now we whitelist "/var/lib/zuul/builds/uuid/work" (my understanding of '~'): http://git.openstack.org/cgit/openstack-infra/zuul/tree/zuul/ansible/paths.py?h=feature/zuulv3#n2522:44
dmsimardBut we implicitely prepare roles in /var/lib/zuul/builds/uuid/ansible, which is not allowed by that directive22:44
dmsimardExample of role preparation: 2018-01-15 21:20:15,318 DEBUG zuul.AnsibleJob: [build: 406ded37f2f5469fbe0ca8af3011fd5d] Adding role path /var/lib/zuul/builds/406ded37f2f5469fbe0ca8af3011fd5d/ansible/playbook_0/role_2/zuul-jobs/roles22:45
dmsimardThis seems to be problematic when loading content from within the /var/lib/zuul/builds/uuid/ansible, see here for example: http://logs.openstack.org/89/514489/3/check/openstack-infra-base-integration-opensuse423/406ded3/job-output.txt.gz#_2018-01-15_21_20_48_12076222:46
dmsimardIs the fix to authorize "../ansible" (/var/lib/zuul/builds/uuid/ansible) in http://git.openstack.org/cgit/openstack-infra/zuul/tree/zuul/ansible/paths.py?h=feature/zuulv3#n25 ?22:47
pabelangerwhy does an untrusted job need access to ansible folder?22:49
dmsimardpabelanger: that's an include_vars task which wants to include vars from that role...22:51
dmsimardpabelanger: i.e, configure-unbound is in /var/lib/zuul/builds/uuid/ansible/roles/configure-unbound and it wants to include /var/lib/zuul/builds/uuid/ansible/roles/configure-unbound/vars/something.yml22:51
dmsimardI think it has to do with using "{{ role_path }}" instead of doing a relative include22:52
dmsimardSee the include here http://git.openstack.org/cgit/openstack-infra/openstack-zuul-jobs/tree/roles/configure-unbound/tasks/main.yaml#n4122:53
dmsimardThe role_path is required in the include_vars because Ansible uses precedence when finding relative paths -- so if you have a "vars" folder inside your playbook directory, it's going to look there before looking inside your role's vars directory.22:54
corvusdmsimard: let's chat with mordred about this when he's back23:11
mordredcorvus, dmsimard: ohai! sorry - was stuck in a dark hole all day - just now coming up for air - looks like I missed a fun day though :(23:12
* mordred reads scrollback23:12
dmsimardmordred: I have a kept build of our role path issue, summary is above23:14
mordreddmsimard: yah, thank you for the writeup - that at least makes it make more sense...23:14
mordreddmsimard, corvus: as for what to do about it ... I'm not super-crazy about opening up that dir explicitly - since whitelisting it whitelists it for all the things23:17
corvusthat could allow untrusted playbooks to overwrite trusted ones.  we would need to be very careful with that.23:18
mordredyah23:18
mordredI think it's a topic worthy of some serious pondering ... I think that's our only current use of explicit {{ role_path }} at the moment23:19
corvusmordred: dhellman may have a use case, but it may not be necessary23:20
dmsimardfwiw we can work around the issue by naming one of the directories something else than "vars"23:22
dmsimardBut it's a workaround23:22
mordreddmsimard: yah - I think that's a good workaround for now until we have time to sort out a more fundamental approach we're all happy with23:22
dmsimardEither that or we convince upstream that their precedence for include_vars is wrong (i.e, when used from inside a role, the role should have precedence)23:23
dmsimardWhich I could try but that's not going to help us in the short term23:23
mordredno, and I don't think we're likely to convince them - I think the current precedence is that way to allow folks to have playbook associated overrides23:24

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!