Thursday, 2018-12-13

ianwok, it must have been similar enough for me to be confused ;)  i think if have them both as something obviously different that's great00:01
dmsimardianw: for the key, I'm thinking we can probably set it up in an unrelated location -- it's just for the duration of the installation00:04
dmsimardAs long as the location of the key matches in the .repo file00:04
dmsimardI was renaming the key but it felt... weird00:05
dmsimardLike, we could put it in /tmp and it wouldn't matter00:05
ianwthat seems fine too00:06
openstackgerritDavid Moreau Simard proposed openstack-infra/zuul-jobs master: Vendor the RDO repository configuration for installing OVS  https://review.openstack.org/62481700:18
dmsimardclarkb, ianw: ^ this sets up the key temporarily in /tmp/ to ensure it doesn't clash with the key installed by packages00:18
*** dkehn has joined #zuul01:42
*** dkehn has left #zuul01:42
*** dkehn has joined #zuul01:42
*** ssbarnea|rover has quit IRC02:06
openstackgerritIan Wienand proposed openstack-infra/nodepool master: Switch devstack jobs to Xenial  https://review.openstack.org/62485503:47
openstackgerritIan Wienand proposed openstack-infra/nodepool master: Switch devstack jobs to Xenial  https://review.openstack.org/62485503:48
openstackgerritIan Wienand proposed openstack-infra/nodepool master: Switch devstack jobs to Xenial  https://review.openstack.org/62485503:51
*** bhavikdbavishi has joined #zuul04:16
*** swest has joined #zuul05:42
*** swest has quit IRC05:48
*** swest has joined #zuul05:50
*** swest has quit IRC05:55
*** bhavikdbavishi has quit IRC06:00
*** swest has joined #zuul06:12
*** quiquell|off is now known as quiquell06:59
*** bhavikdbavishi has joined #zuul06:59
*** rlandy|bbl has quit IRC06:59
*** smyers has quit IRC07:03
*** bhavikdbavishi has quit IRC07:12
*** pcaruana has joined #zuul07:12
*** smyers has joined #zuul07:25
*** hashar has joined #zuul08:02
*** goern has joined #zuul08:09
*** Alvass has joined #zuul08:23
*** themroc has joined #zuul08:28
AlvassHi, I'm having some problems with nodepool configuration when trying to use winrm for the host08:42
Alvassthe zuul-executor seem to crash in server.py, row 1262 with 'NoneType is not iterable'. When trying to get the host key to set know_hosts for ssh. Not sure but this seems like a bug.08:44
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: refactor build page to use a reducer  https://review.openstack.org/62489408:55
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: refactor build page using a container  https://review.openstack.org/62489508:55
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add errors from the job-output to the build page  https://review.openstack.org/62489608:55
*** jpena|off is now known as jpena09:04
tristanCmordred: corvus: using the "local" tenant of the zuul-build-dashboard-multi-tenant shows the failed task output on the build page of ^ To make this work with openstack, we'll need cors header from the logserver.09:10
tristanCactually, the zuul-build-dashboard also works with logs.openstack.org because it's already serving from there09:20
*** gtema has joined #zuul09:25
*** ssbarnea|rover has joined #zuul09:34
*** sshnaidm has quit IRC09:37
*** ssbarnea|rover has quit IRC09:53
*** bhavikdbavishi has joined #zuul09:54
*** gtema has quit IRC10:01
*** electrofelix has joined #zuul10:04
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: web: add errors from the job-output to the build page  https://review.openstack.org/62489610:16
*** sshnaidm has joined #zuul10:20
*** bhavikdbavishi has quit IRC10:34
*** ssbarnea|rover has joined #zuul10:43
*** sshnaidm has quit IRC10:50
*** hashar has quit IRC11:07
*** dkehn has quit IRC11:17
*** Alvass has quit IRC11:17
*** hashar has joined #zuul11:25
*** tobias-urdin is now known as tobias-urdin_afk11:41
*** tobias-urdin_afk is now known as tobias-urdin11:42
*** tobias-urdin is now known as tobias-urdin_afk11:43
*** dkehn has joined #zuul12:04
*** sshnaidm has joined #zuul12:12
*** bhavikdbavishi has joined #zuul12:20
*** gtema has joined #zuul12:36
*** tobias-urdin_afk is now known as tobias-urdin12:53
*** gtema has quit IRC13:00
*** jpena is now known as jpena|lunch13:01
*** rlandy has joined #zuul13:03
*** themroc has quit IRC13:05
*** themroc has joined #zuul13:06
*** themroc has quit IRC13:06
ssbarnea|roverhow can I find what happens with zuul on a job? it seems stuck at https://zuul.openstack.org/stream/0dc38407c7c94063afc53ad126cfe932?logfile=console.log -- and 624381 seems stuck13:12
ssbarnea|roverthis was supposed to be last task on post, but even after 40mins, zuul is still waiting for it to finish13:13
*** hashar has quit IRC13:18
*** bhavikdbavishi has quit IRC13:25
pabelangerssbarnea|rover: which review was that for, it is done now, but if nothing is showing in logs a zuul admin may need to look at executor logs13:57
*** irclogbot_3 has quit IRC14:02
ssbarnea|roverpabelanger: this is serious problem because we have high number of POST_FAILUREs, and we assumed that something wrong was happening with "du", and this was an attempt to fix it. Guess what, it didn't and my impression is that may not be related to that.14:03
ssbarnea|roverpabelanger: the behaviours was always the same: post was failing and last thing seen was this du command which happens to be the last command in the last task.14:03
ssbarnea|rovernow I am 100% sure that du executes before that and zuul gets stuck.14:04
ssbarnea|roverthe mentioned bug has a link to logstash: http://logstash.openstack.org/#/dashboard/file/logstash.json?query=(message:%20%5C%22FAILED%20with%20status:%20137%5C%22%20OR%20message:%20%5C%22FAILED%20with%20status:%20143%5C%22%20OR%20message:%20%5C%22POST-RUN%20END%20RESULT_TIMED_OUT%5C%22)%20AND%20tags:%20%5C%22console%5C%22%20AND%20voting:114:06
*** smyers has quit IRC14:08
*** hashar has joined #zuul14:08
ssbarnea|roveri checked and I was not able to find any builder, branch or job pattern. it just happens, and often enough to be of concern. 177 counts in one week.14:09
*** irclogbot_3 has joined #zuul14:12
*** smyers has joined #zuul14:12
pabelangerssbarnea|rover: lets move to #openstack-infra to debug14:13
ssbarnea|roversure, good idea. i was not sure where it fits.14:13
*** swest has quit IRC14:23
*** albin_vass has joined #zuul14:47
*** quiquell is now known as quiquell|lunch14:58
*** panda|off is now known as panda14:58
*** quiquell|lunch is now known as quiquell15:10
*** andreykurilin has joined #zuul15:21
andreykurilinhi folks! Can someone help me to solve the issue? I'm trying to access the variable defined by 'parent' job, but i'm getting strange error15:22
pabelangersure, if you want to link your log file15:23
andreykurilinhttps://review.openstack.org/#/c/624409/5/tests/ci/playbooks/run-rally-task.yaml (originally, I tried "{{ devstack_localrc.get('USE_PYTHON3') or False }}", but simplified this temporary)15:23
andreykurilinand here is a usage of my local variable https://review.openstack.org/#/c/624409/3/tests/ci/playbooks/roles/snapshot-available-os-resources/tasks/main.yaml15:24
andreykurilinerror: http://logs.openstack.org/09/624409/5/check/rally-task-simple-job/1971602/job-output.txt.gz#_2018-12-13_15_00_15_67151115:24
andreykurilinps: I do not need a local variable and would like to use directly the parent one from my roles, but it raised the same issue and I tried to play with accessing it15:25
pabelangerandreykurilin: lets move this to #openstack-infra15:29
andreykurilinok15:29
andreykurilinpabelanger: it looks like I cannot send messages there which is strange since I have registered nick. hm15:32
pabelangerandreykurilin: ah, yes. you need to register nick with freenode15:33
pabelangerit is needed now to cut down on spam15:33
pabelangerbut, thought it was also enabled here15:33
andreykurilinpabelanger: i already have registered nick. strange, will recheck everything15:36
*** albin_vass has quit IRC16:02
*** jpena|lunch is now known as jpena16:08
*** morgan is now known as kmalloc16:09
*** andreykurilin has quit IRC16:12
*** andreykurilin has joined #zuul16:12
*** quiquell is now known as quiquell|off16:22
*** hashar has quit IRC16:50
dmsimardclarkb: btw I just did a recheck on https://review.openstack.org/#/c/624526/ for https://review.openstack.org/#/c/624817/16:50
clarkbdmsimard: thanks16:51
clarkbdmsimard: if the tripleo jobs come back happy I think we should merge your change. That should be a decent reliability improvement17:08
*** hashar has joined #zuul17:09
*** hashar has quit IRC17:14
*** arxcruz is now known as arxcruz|off|next17:34
*** arxcruz|off|next is now known as arxcruz|next_yea17:34
*** arxcruz|next_yea is now known as arxcruz|next_yr17:35
*** ianychoi has quit IRC17:42
*** panda is now known as panda|off17:42
*** hashar has joined #zuul17:51
*** sshnaidm is now known as sshnaidm|off18:01
*** rlandy is now known as rlandy|biab18:09
*** hashar has quit IRC18:22
*** electrofelix has quit IRC18:34
*** jpena is now known as jpena|off18:38
*** electrofelix has joined #zuul18:44
*** electrofelix has quit IRC18:51
SpamapSI dare you to read this Ansible task and not get a little angry at the universe in general for having made me do this: http://paste.openstack.org/show/737251/19:21
*** rlandy|biab is now known as rlandy19:23
dmsimardSpamapS: that would be more readable in python :/19:43
*** pcaruana has quit IRC19:47
*** hashar has joined #zuul19:48
*** hashar has quit IRC19:48
SpamapSdmsimard: yep19:55
SpamapSBut it shouldn't have to be like this19:55
SpamapSit's just "put this map in this other map at this location"19:55
SpamapSSometimes I wish there was a more lightweight "put python here".19:56
*** gtema has joined #zuul20:12
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Extract out common config parsing for ConfigPool  https://review.openstack.org/62164220:23
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Extract common config parsing for ProviderConfig  https://review.openstack.org/62509420:23
*** ianychoi has joined #zuul20:59
*** gtema has quit IRC21:11
*** tobias-urdin has quit IRC21:32
SpamapSHm I just had something super weird happen.22:39
SpamapSLanded changes to my post jobs..22:39
SpamapSbut it's still running the old ones on commit to that branch.22:39
SpamapSmerger fail?22:39
SpamapSLike, it's still running the old playbooks, the old content, despite a new sha.22:40
clarkbSpamapS: or an ordering/race problem. If the trigger for post is processed before the code is merged event that updates config it may kick off jobs with the old config?22:40
clarkbhrm I would've expect the merger to pull in the up to date playbooks in that case though22:41
clarkb(since it should be running against $sha1 with thoseplaybooks)22:41
SpamapSIt has happened twice now.22:41
SpamapSI was sort of in the "oh maybe.. sequence..." but now..22:42
clarkbare there config loading errors? that may possibly force it to fallback to the last successfully loaded config22:42
clarkb(though I think it would actually result in jobs not running instead)22:43
SpamapSYou can see logs and stuff here https://zuul.gdmny.co/builds22:43
SpamapS(though the sha's are in private repos so you can't see those)22:43
SpamapSno, gate landed the changes.22:43
clarkbno alarm bell for broken configs22:43
clarkbSpamapS: in your job logs, there should be the inventory file that zuul logged for you (this comes from our base jobs though so you may not have that I guess). In that inventory should be the job inheritance path for that job22:44
clarkbSpamapS: does that offer any clues like maybe the job is actually defined elsewhere22:44
SpamapSI do have the inventory file22:44
clarkbone thing that can happen like for tag events is those don't map to branches so if you have multiple branches with overlapping config for those pipelines zuul gets confused and runs the first found iirc22:45
clarkbI don't think this is uspposed to be an issue for post since post should be able to be associated with a branch. But maybe that is github vs gerrit difference22:45
SpamapShttp://paste.openstack.org/show/737271/22:45
SpamapSand I have `branches: master` in that job22:46
clarkbthat may confuse it in post depending on the available branch context data on that trigger (I have no idea if that works with the github triggers)22:46
clarkbthough I'd expect the confusion to be job doesn't run at all because branch doesn't match22:47
clarkbSpamapS: tobiash had all these paged in recently (maybe look at recent changes from tobiash too?) but aiui github sends and event that says this merged. Zuul is then supposed to send the please update my repo event to the mergers. THis moves master ahead and results in new config being loaded22:48
SpamapSThe trigger is on push22:50
SpamapS    trigger:22:50
SpamapS      github:22:50
SpamapS- event: push22:51
clarkbthere was a similar thing that led to me writing https://review.openstack.org/621665 which ended up being unnecessary after tobiash dug more. This was to handle deleted branches on the mergers (which weren't being updated)22:51
SpamapSbut the weird thing is that this has run like, 50 times in the last 10 days.22:51
clarkbthere was another change that I thought improved that specific issue but having a hard time finding that, but wonder if it is related to this22:51
SpamapSWe've been using it to push our site live. It's been nice. :-P22:51
SpamapSto me it looks like the job itself just isn't updated in the config22:52
SpamapSI wonder if the scheduler had a problem.22:52
clarkbthe sequencing there (from memory) is that the scheduler will ask the mergers to give it the new updates for .zuul.yaml etc, then it compiles that and applies to subsequent runs of jobs22:53
clarkbso ti could be in the scheduler or the merger potentially22:53
clarkbI would look for logs around when the push trigger happens on the scheduler and merger to see if it says anything about loading its configs22:57
SpamapSthing is, that might explain 122:59
SpamapSbut this is multiple fails in a row. :(22:59
clarkbunless its a persistent failure to update22:59
clarkbthen each push would fail to update22:59
clarkb(potentially)22:59
*** rlandy has quit IRC23:03
SpamapSfailure to update what?23:03
clarkbthe repo for .zuul.yaml config updates23:04
* SpamapS also has just discovered that his dev ELK instance filled up about 6 hours ago.. whoops23:04
clarkbelasticsaerch will do that for you openstack runs 6 * 1TB disk instances and keeps 10 days of data23:04
SpamapSclarkb: Not sure I follow.. but.. I'm going to try restarting the scheduler next. :-P23:04
SpamapSthe weird thing is, the *contents*.. the working dirs.. are all correct.23:05
clarkbSpamapS: the way zuul updates its configuration is the scheduler asks a merger to cat the contents of the various possible zuul config files. If that fails the scheduler won't know what the new config is23:05
SpamapSjust not the zuul config or playbooks.23:05
clarkb(and may keep running with the old config)23:05
SpamapSah ok23:06
SpamapSI'll dig around in the logs for something like that23:06

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!