Sunday, 2017-10-15

*** Douhet has joined #openstack-infra00:02
*** yamahata has quit IRC00:06
*** cody-somerville has quit IRC00:12
*** thorst has joined #openstack-infra00:13
*** armax has quit IRC00:13
*** thorst has quit IRC00:13
*** cody-somerville has joined #openstack-infra00:17
*** cody-somerville has joined #openstack-infra00:17
*** Shrews has quit IRC00:18
*** yamahata has joined #openstack-infra00:19
*** bobh has joined #openstack-infra00:30
*** Shrews has joined #openstack-infra00:32
*** bobh has quit IRC00:35
*** smatzek has quit IRC00:36
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: legacy-trove-functional-dsvm-mysql requires trove-integration  https://review.openstack.org/51206400:45
pabelangermordred: something does look right on trove legacy jobs, I think we might have duplicated them00:47
pabelangerhttp://git.openstack.org/cgit/openstack-infra/openstack-zuul-jobs/tree/zuul.d/zuul-legacy-jobs.yaml#n14131 - legacy-trove-functional-dsvm-mysql00:47
pabelangerhttp://git.openstack.org/cgit/openstack-infra/openstack-zuul-jobs/tree/zuul.d/zuul-legacy-jobs.yaml#n14145 - legacy-trove-legacy-functional-dsvm-mysql00:47
pabelangerwill have to confirm with trove PTL in the morning00:49
clarkbpabelanger: aiui they have their legacy jobs for older branches and non legacy for master00:50
clarkband that is independent of zuulv3 migration00:50
*** Shrews has quit IRC00:51
pabelangerclarkb: oh, how confusing00:51
*** Shrews has joined #openstack-infra00:57
*** yamamoto has joined #openstack-infra00:58
*** yamamoto has quit IRC00:59
*** yamamoto has joined #openstack-infra01:01
*** owalsh_ has joined #openstack-infra01:01
*** owalsh has quit IRC01:04
*** yamamoto has quit IRC01:12
*** smatzek has joined #openstack-infra01:14
*** yamamoto has joined #openstack-infra01:20
*** gouthamr has quit IRC01:20
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Add horizon templates  https://review.openstack.org/51201301:22
*** yamamoto has quit IRC01:25
*** yamamoto has joined #openstack-infra01:30
*** bobh has joined #openstack-infra01:31
*** yamamoto has quit IRC01:35
*** bobh has quit IRC01:35
*** thorst has joined #openstack-infra01:59
*** thorst has quit IRC02:04
*** smatzek has quit IRC02:06
*** xarses has joined #openstack-infra02:07
*** cody-somerville has quit IRC02:14
*** bobh has joined #openstack-infra02:32
*** yamamoto has joined #openstack-infra02:33
*** bobh has quit IRC02:36
*** dhinesh has joined #openstack-infra02:36
*** yamamoto has quit IRC02:38
*** gouthamr has joined #openstack-infra02:39
*** edmondsw has joined #openstack-infra02:40
*** edmondsw has quit IRC02:44
*** thorst has joined #openstack-infra03:00
*** thorst has quit IRC03:04
*** xarses has quit IRC03:17
*** thorst has joined #openstack-infra03:18
*** thorst has quit IRC03:18
*** smatzek has joined #openstack-infra03:26
*** smatzek has quit IRC03:32
*** bobh has joined #openstack-infra03:33
*** bobh has quit IRC03:37
*** nikhil_k has quit IRC03:40
*** dhinesh has quit IRC04:00
*** yamamoto has joined #openstack-infra04:15
*** yamamoto has quit IRC04:23
*** gongysh has joined #openstack-infra04:25
*** edmondsw has joined #openstack-infra04:28
*** edmondsw has quit IRC04:32
*** bobh has joined #openstack-infra04:33
*** bobh has quit IRC04:38
*** liusheng has quit IRC04:41
*** dimak has quit IRC04:45
*** psachin has joined #openstack-infra04:45
*** ari[m] has quit IRC04:47
*** dimak has joined #openstack-infra04:47
*** ari[m] has joined #openstack-infra04:47
*** liusheng has joined #openstack-infra04:54
*** thorst has joined #openstack-infra05:19
*** chlong has joined #openstack-infra05:20
*** thorst has quit IRC05:24
*** armaan has joined #openstack-infra05:36
*** snapiri has joined #openstack-infra05:52
*** gouthamr has quit IRC06:12
*** claudiub has quit IRC06:13
*** bobh has joined #openstack-infra06:35
*** bobh has quit IRC06:40
*** pcaruana has joined #openstack-infra06:42
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Convert publish-api-ref for service name  https://review.openstack.org/51203106:52
openstackgerritEmilien Macchi proposed openstack-infra/tripleo-ci master: Support pike->master undercloud upgrades  https://review.openstack.org/51188506:52
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Fix translations: Use zanata_client_version  https://review.openstack.org/51207406:55
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Fix translations: Use zanata_client_version  https://review.openstack.org/51207406:55
*** markvoelker has joined #openstack-infra06:57
*** markvoelker_ has joined #openstack-infra06:59
*** markvoelker has quit IRC07:02
*** markvoelker has joined #openstack-infra07:03
*** markvoelker_ has quit IRC07:06
*** slaweq has joined #openstack-infra07:07
*** slaweq has quit IRC07:09
*** thorst has joined #openstack-infra07:21
*** CHIPPY has joined #openstack-infra07:21
*** thorst has quit IRC07:25
*** bobh has joined #openstack-infra07:36
*** jtomasek has joined #openstack-infra07:40
*** liusheng has quit IRC07:40
*** bobh has quit IRC07:40
*** liusheng has joined #openstack-infra07:41
*** ssbarnea has quit IRC07:43
*** wznoinsk has quit IRC07:44
*** wznoinsk has joined #openstack-infra07:46
*** ssbarnea has joined #openstack-infra07:46
*** armaan has quit IRC07:50
*** armaan has joined #openstack-infra07:50
*** jtomasek has quit IRC07:51
*** edmondsw has joined #openstack-infra08:04
*** edmondsw has quit IRC08:08
*** liusheng has quit IRC08:14
*** liusheng has joined #openstack-infra08:14
*** armaan has quit IRC08:20
*** armaan has joined #openstack-infra08:21
*** slaweq has joined #openstack-infra08:26
*** bobh has joined #openstack-infra08:37
*** bobh has quit IRC08:41
*** armaan has quit IRC08:44
*** armaan has joined #openstack-infra08:44
*** armaan has quit IRC08:45
*** armaan has joined #openstack-infra08:45
*** bhavik1 has joined #openstack-infra08:48
*** bhavik1 has quit IRC08:49
*** CHIPPY has quit IRC08:50
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Use horizon templates  https://review.openstack.org/51201708:50
*** slaweq has quit IRC08:59
*** tosky has joined #openstack-infra09:08
*** jascott1 has quit IRC09:28
*** jascott1 has joined #openstack-infra09:29
*** jascott1 has quit IRC09:33
*** persia has quit IRC09:35
*** persia has joined #openstack-infra09:36
*** bobh has joined #openstack-infra09:37
*** chlong has quit IRC09:38
*** armaan has quit IRC09:40
*** armaan has joined #openstack-infra09:40
*** bobh has quit IRC09:42
snapiriAJaeger: Is there a way I can tell if I am running under zuulv3 or zuulv2? I need it to remove the zuul-clone in the zuulv3 path...09:48
*** edmondsw has joined #openstack-infra09:52
*** edmondsw has quit IRC09:56
snapiriHi infra: any answer for ^^ ?10:08
*** pbourke has quit IRC10:21
*** pbourke has joined #openstack-infra10:23
*** yamamoto has joined #openstack-infra10:35
*** bobh has joined #openstack-infra10:38
*** bobh has quit IRC10:44
*** claudiub has joined #openstack-infra10:51
*** thorst has joined #openstack-infra10:54
*** jkilpatr_ has joined #openstack-infra10:57
*** thorst has quit IRC10:59
toskysnapiri: jobs that vote as "Jenkins" are zuul v2; jobs that votes as "Zuul" are zuul v311:01
*** yamamoto has quit IRC11:02
toskyright now both are leaving a vote, but the vote from Jenkins (aka zuul v2) is still the vote which drives the merge11:02
toskywith the exceptions of few infra repositories which are zuulv3-only11:02
toskythe switch will be later today, according the last emails11:02
*** armaan has quit IRC11:03
*** yamamoto has joined #openstack-infra11:03
*** armaan has joined #openstack-infra11:03
snapiritosky: thanks for your reply. What I am trying to figure out is how to modify my scripts, so tox_install will not use zuul-clone when using zuulv3. When zuul-clone is used the build fails :(11:07
*** yamamoto has quit IRC11:08
*** jkilpatr_ has quit IRC11:08
*** abelur_ has quit IRC11:11
toskysnapiri: I would rather understand why zuul-clone fails11:21
*** askb has quit IRC11:24
*** abelur_ has joined #openstack-infra11:29
*** markvoelker has quit IRC11:32
*** askb has joined #openstack-infra11:37
*** bobh has joined #openstack-infra11:39
*** edmondsw has joined #openstack-infra11:40
*** jascott1 has joined #openstack-infra11:40
*** bobh has quit IRC11:44
*** edmondsw has quit IRC11:44
*** AJaeger has joined #openstack-infra11:45
AJaegerinfra-root, could you review https://review.openstack.org/512074 for the next translation fix, please?11:45
AJaegerinfra-root, and let's merge https://review.openstack.org/#/c/512030/ - to not publish api-ref/api-guide to wrong locations and then forced to cleanup. This gives us time to find a proper solution. https://review.openstack.org/512031 is one way forward but needs followup for additional projects11:47
AJaegerthere're a couple more fixes in the review queue - I suggest to get everthing in to be in a good state but those thow above are critical for me11:48
toskyAJaeger: aaaand if you have some time to extend https://review.openstack.org/#/c/512017/ for sahara-dashboard too, I would be happy :)11:49
AJaegersnapiri: then let's fix it - did you read the docs we have? If you show me a review, I might guide you for a fix11:49
AJaegertosky: looking...11:49
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Use horizon templates  https://review.openstack.org/51201711:53
AJaegertosky: done ^11:53
toskythanks!11:54
AJaegersnapiri: I expect that you miss a few "required-projects" lines, this was all explained in the mails by Monty. Do you have a log file to look at?11:54
toskyit was pointless to create yet-another-review for that11:54
AJaegertosky: you could have just updated mine, would have been ok ;)11:54
toskyAJaeger: oh, ok, but it's better that you did it: you added also two additional templates that I would have missed11:55
AJaegerinfra-root, the one cross-repo job we should fix IMHO is requirements-check - see etherpad. That will hit quite a few repos11:55
AJaegertosky: ;)11:56
toskyAJaeger: just to be sure: the changes in project-config were not applied immediately to jobs for v2; if I understand it correctly, this is the same now, and changes are applied immediately only for the other repositories; is it correct?11:56
AJaegertosky: you mean with depends-on?11:58
AJaegeryeah, depends-on a project-config change will not update the jobs - since it's a trusted repo.11:58
toskyack, so I've got it correctly11:59
*** thorst has joined #openstack-infra11:59
toskynow let's hope that infra-root will merge 512017 too before The Switch™12:01
AJaegertosky: would be great indeed...12:02
*** markvoelker has joined #openstack-infra12:08
toskyAJaeger: would a recheck on https://review.openstack.org/#/c/508906/ make sense? There is only one error which is seems infra-related12:11
openstackgerritMerged openstack-infra/tripleo-ci master: Support pike->master undercloud upgrades  https://review.openstack.org/51188512:11
*** ccamacho has quit IRC12:11
AJaegertosky: we hit this "devstack finger://ze08.openstack.org/d3619f975aea4d45b3fb0de4eea1e448 : RETRY_LIMIT in 1m 26s" already twice - not sure what's going on there. Let's try...12:12
AJaegerjeblair, mordred ^12:13
*** markvoelker_ has joined #openstack-infra12:14
toskybut what is the failed job? It would be useful to have it printed instead of the link to the output stream12:15
toskyoh, the "devstack" job, of course12:16
*** markvoelker has quit IRC12:17
*** boden has joined #openstack-infra12:17
toskyif I add that devstack job to my project, should add my project the "required-project" field or is it implicit?12:18
* tosky should probably RTFM more12:18
openstackgerritboden proposed openstack-infra/project-config master: zuul v3 gate changes for vmware-nsx  https://review.openstack.org/51183612:21
*** yamamoto has joined #openstack-infra12:21
*** slaweq has joined #openstack-infra12:22
*** esberglu has joined #openstack-infra12:24
*** thorst has quit IRC12:24
snapiriAJaeger: thanks for the replies. I have read the migration instructions, but maybe I missed something.12:24
snapiriI am a bit lost here... This is a relevant output: http://logs.openstack.org/89/509189/11/check/legacy-dragonflow-dsvm-fullstack/c349087/job-output.txt.gz12:25
snapiriThe relevant patch is: https://review.openstack.org/#/c/508761/12:25
*** armaan has quit IRC12:27
*** armaan has joined #openstack-infra12:27
*** slaweq has quit IRC12:27
*** esberglu has quit IRC12:28
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Remove api-guide remains  https://review.openstack.org/51192412:29
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Add integration tests for multi-node-firewall  https://review.openstack.org/50478812:29
lennybHello, Is there any work going on BlackDuck Hub support for JJB?12:32
AJaegersnapiri: you're playing around with tox_install.sh - I have not looked further into this, I suggest to discusss this some more with tonyb who wanted to look into it. Is your job working without your change? If not, let's fix that first...12:36
openstackgerritAndreas Jaeger proposed openstack-infra/openstack-zuul-jobs master: Add integration tests for multi-node-bridge role  https://review.openstack.org/50478912:37
*** dave-mccowan has joined #openstack-infra12:39
snapiriAJaeger: the log I sent is of a different patch, all patches fail with the same log - permission denied when trying to perform hard-link by zuul-clone. This is the reason I am trying to play with the tox_install12:39
snapiriAJaeger: here is another log if it is relevant: http://logs.openstack.org/34/507034/2/check/legacy-dragonflow-dsvm-fullstack/2b6ba59/12:40
*** bobh has joined #openstack-infra12:40
*** bobh has quit IRC12:45
*** liusheng has quit IRC12:46
*** liusheng has joined #openstack-infra12:47
*** yamamoto has quit IRC12:49
bodenhi, if someone gets a chance it’d be nice to land https://review.openstack.org/#/c/511836/  so I can verify in the respective project before the v3 switch later12:57
*** thorre_se has joined #openstack-infra13:05
*** thorre has quit IRC13:08
*** thorre_se is now known as thorre13:08
AJaegersnapiri: please link to the line that shows the problem that makes it far easier13:12
AJaegerboden: commented13:12
AJaegerbbl13:12
*** yamamoto has joined #openstack-infra13:18
snapiriAJaeger: I believe it is here: http://logs.openstack.org/34/507034/2/check/legacy-dragonflow-dsvm-fullstack/2b6ba59/job-output.txt.gz#_2017-10-15_12_20_16_71027113:24
snapiriTo the best of my understanding zuul-clone tries to create hard-links and fails13:26
*** edmondsw has joined #openstack-infra13:28
fungisnapiri: ahh, yes, there are some errors just before that in the log as well about "ERROR: invocation failed (exit code 1), logfile: /opt/stack/new/dragonflow/.tox/fullstack/log/fullstack-1.log" but we don't appear to collect that logfile13:31
*** edmondsw has quit IRC13:33
mordredfungi: oh - that looks likea permissions issue13:34
fungimordred: well, my expectation is that it's a cross-device hardlink13:34
snapirifungi: any ideas? I thought that all the errors above are actually the content of the file...13:34
jeblairi just woke up and tried to figure out what ikea permissions were13:34
mordredfungi: we should be testing for that in zuul-cloner shim13:35
mordredjeblair: :)13:35
fungijeblair: you can only adjust them with an allen key?13:35
jeblairhehe13:36
fungibut yeah, now i'm thinking it's a permissions issue because iirc we don't have /opt and /home on separate filesystems in ovh, where that example build happened13:38
fungiconfirmed: http://logs.openstack.org/34/507034/2/check/legacy-dragonflow-dsvm-fullstack/2b6ba59/logs/df.txt.gz13:39
mordredfungi: my guess is that that tox command is running as the stack user and is trying to make hardlinks to the git dirs in ~zuul13:40
snapirimordred: I believe you are right13:40
snapiriThe tox is being run as stack user13:40
*** bobh has joined #openstack-infra13:41
*** sbezverk has quit IRC13:41
mordredlonger-term there's a nicer pattern that can be used where the tox can be run using the tox role from the job as the stack user ... but shorter term we probably need to just chmod -R the src tree to be writable13:41
snapirihttp://logs.openstack.org/34/507034/2/check/legacy-dragonflow-dsvm-fullstack/2b6ba59/job-output.txt.gz#_2017-10-15_12_20_08_51953613:41
* mordred is looking at the requirements issue btw ...13:42
*** bobh has quit IRC13:42
fungimordred: world-writable you mean? (o+w/a+w?)13:42
*** bobh has joined #openstack-infra13:42
mordredfungi: yah13:43
*** yamamoto has quit IRC13:45
*** bobh has quit IRC13:45
jeblairAJaeger: i'm leaving https://review.openstack.org/495426 for you to +313:45
mordredjeblair: this requirements issue doesn't make sense me to - do you have a sec to look at it with me?13:47
*** slaweq has joined #openstack-infra13:47
jeblairmordred: i can give it a shot.  i'm still warming up.  :)13:47
mordredjeblair: so - if you look at http://logs.openstack.org/05/510105/16/check/requirements-check/7256969/ara/13:48
mordredyou'll see "SystemError: fatal: Not a git repository (or any of the parent directories): .git"13:48
openstackgerritMerged openstack-infra/project-config master: v3 cleanup: Use openstack-python-jobs  https://review.openstack.org/51180313:48
jeblairya13:48
mordredthat comes from here: http://git.openstack.org/cgit/openstack/requirements/tree/playbooks/files/project-requirements-change.py#n17113:49
mordredthing is - cwd should be the git repo: http://git.openstack.org/cgit/openstack/requirements/tree/playbooks/requirements-check.yaml#n613:49
mordredof the project in question13:49
jeblairmordred: docs say chdir added in 2.413:51
jeblairhttp://docs.ansible.com/ansible/latest/script_module.html13:51
jeblair(we're still running 2.3)13:54
openstackgerritMerged openstack-infra/project-config master: Stop translation of trove  https://review.openstack.org/51181313:55
openstackgerritMerged openstack-infra/project-config master: Added initial jobs for python-tempestconf  https://review.openstack.org/51119413:55
jeblairmordred: so maybe we should copy+command ?  or else give the script an extra argument of a directory to change to at the start.13:55
mordredjeblair: wow - I just had a weird networking event - can't ssh directly to my irc bouncer14:01
mordredjeblair: https://review.openstack.org/512083 Move chdir into script itself14:01
mordredjeblair: I agree with your assessment14:01
jeblairdmsimard: 511953 +2 with comment14:03
jeblairmordred: +214:04
mordredjeblair: I have pushed up https://review.openstack.org/510105 with a depends-on14:04
toskyI don't want to ruin your Sunday, but the new devstack job is failing with an hard-to-debug error: https://review.openstack.org/#/c/508906/14:04
mordredfungi: mind +3ing https://review.openstack.org/#/c/495426/ ?14:05
openstackgerritMerged openstack-infra/project-config master: Remove legacy magnum jobs from pipeline  https://review.openstack.org/50880414:05
jeblairtosky: i'll track that down14:06
toskythanks14:06
jeblairtosky, mordred: http://paste.openstack.org/show/623656/  are the 2 errors (a job bug and a zuul bug)14:09
toskyjeblair: out of curiosity, in case it happens again: how did you get the log? From the finger:// link?14:11
*** psachin has quit IRC14:11
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Fix python3 unicode errors in executor server  https://review.openstack.org/51208414:12
jeblairtosky, mordred: ^14:12
jeblairtosky: root access on the zuul executor -- the exception is a bug in zuul that prevented the log uploading functions from working.14:13
*** yamamoto has joined #openstack-infra14:13
toskyoh, I see14:13
openstackgerritMerged openstack-infra/project-config master: Add required projects for networking-cisco  https://review.openstack.org/51200414:13
openstackgerritMerged openstack-infra/project-config master: Remove training-guides legacy jobs  https://review.openstack.org/51201014:13
openstackgerritAndreas Florath proposed openstack/diskimage-builder master: [WIP] disk-usage refactoring  https://review.openstack.org/51208514:14
jeblairfungi: can you also review 512084 -- that's a pretty serious error we should restart the executors for14:14
openstackgerritMerged openstack-infra/project-config master: Use horizon templates  https://review.openstack.org/51201714:15
pabelangermorning14:16
pabelanger+3 on 51208314:17
jeblairtosky, mordred: i pushed up a new revision of remote:   https://review.openstack.org/508906 Add devstack base job for zuul v314:17
jeblairtosky, mordred: i had forgotten to git-add the configure-swap role :(14:17
toskyoh, that was easy to solve luckily :)14:18
toskyso I can recheck my job creating another 11-patches stack14:18
jeblair\o/14:19
pabelangerlooking at zuulv3-issues on etherpad14:19
*** slaweq has quit IRC14:20
bodenAJaeger left you a response in https://review.openstack.org/#/c/511836… I don’t understand one of your comments14:20
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Use os-service-types to get api-ref destination  https://review.openstack.org/51208614:21
mordredAJaeger, jeblair: ^^ that is an alternate suggestion for 51203114:21
openstackgerritMerged openstack-infra/project-config master: Mark publish-openstack-artifacts and pipelines as post-review  https://review.openstack.org/49542614:22
*** slaweq has joined #openstack-infra14:22
jeblairmordred: we should consider pointing our executors at our pip mirrors14:23
openstackgerritboden proposed openstack-infra/project-config master: zuul v3 gate changes for vmware-nsx  https://review.openstack.org/51183614:24
mordredobviously that pre-task section can be cleaned up and put into a role - but I figured that would be the shortest way to talk about it14:24
jeblairmordred: (or maybe making a sort of configure-mirrors role for tasks that run on the executors)14:24
mordredjeblair: yah - alternately we could shift that task to be a post-task on the remote node14:24
jeblairor that14:24
*** thorst has joined #openstack-infra14:25
*** yamamoto has quit IRC14:25
jeblairmordred: this probably isn't a big deal for this job, but generally speaking, we'd probably trust the executor to produce the correct publish location more than the node.14:26
jeblairmordred: i'd probably rank solutions as: 1) executor-configure-mirrors role 2) post-task on node  (and i don't think we should do my first suggestion of setting a mirror for the executors at the system level)14:27
*** modrednetfail has joined #openstack-infra14:27
mordredjeblair: there's a third option - which is fetching the service-types json file rather than consuming the library14:28
jeblairthat's published by its own job, yeah?14:28
jeblairso there's an initial bootstrapping sequence, but no catch-22.14:28
jeblairthat would work for me too (it's eventually consistent if you add a new type)14:29
*** logan- has quit IRC14:29
pabelangerzuulv3-issues looks to be in great shape, over all. Going to see if I can find any unknown issues on zuulv3.o.o status page14:29
*** logan- has joined #openstack-infra14:29
*** thorst has quit IRC14:29
*** bobh has joined #openstack-infra14:31
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Use service-types.openstack.org to get api-ref destination  https://review.openstack.org/51208614:33
mordredjeblair: like that ^^14:33
pabelangerclarkb: who is our citycloud contact again? We have 8 stuck nodes in citycloud-kna1 we should ask to see if they can help clean them up14:36
mordredjeblair: if we like that I can make it nicer14:36
pabelangeroff topic, but does any other infra-root get 2.9M emails from root@cacti01.openstack.org about /usr/local/bin/create_graphs.sh failing?14:39
pabelangerI'm on limited bandwidth over the next month or so, and its killing my quota now14:40
*** chlong has joined #openstack-infra14:47
pabelangerah, think I see it14:48
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Use service-types.openstack.org to get api destinations  https://review.openstack.org/51208614:53
openstackgerritPaul Belanger proposed openstack-infra/system-config master: Fix grep error when listing ifName values  https://review.openstack.org/51208814:55
openstackgerritPaul Belanger proposed openstack-infra/system-config master: Fix grep error when listing ifName values  https://review.openstack.org/51208814:56
pabelangerjeblair: mordred: clarkb: fungi: ^ when you have spare time, should help cut down on email size from cacti.14:56
openstackgerritMonty Taylor proposed openstack-infra/openstack-zuul-jobs master: Add role to provice a service_type_data fact  https://review.openstack.org/51208914:58
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Use service-types.openstack.org to get api destinations  https://review.openstack.org/51208614:59
mordredjeblair, pabelanger, AJaeger: ^^ ok - I think that's the clean version15:00
jeblairdmsimard: i'm happy to discuss the ara changes whenever our schedules next sync up.  :)  i'm happy to change the approach, or discard them entirely if you have other suggestions.  just trying to be helpful.  at the very least, i'm learning more about ara.15:02
*** wolverineav has joined #openstack-infra15:04
dmsimardjeblair: I'm sort of there :p15:05
*** armax has joined #openstack-infra15:05
*** armaan has quit IRC15:05
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Use propose-update-constraints instead of legacy  https://review.openstack.org/51209015:05
*** armaan has joined #openstack-infra15:06
pabelangerfungi: Thanks! I'll push up an update now15:07
fungiany guesses what caused the memory usage for the scheduler to nearly double around 13:30z?15:07
*** armaan has quit IRC15:07
fungier, more than double even15:08
*** armaan has joined #openstack-infra15:08
*** armaan has quit IRC15:08
openstackgerritPaul Belanger proposed openstack-infra/system-config master: Fix grep error when listing ifName values  https://review.openstack.org/51208815:08
pabelangerjeblair: fungi: ^updates15:08
dmsimardjeblair: on the topic of playbook paths and how they're displayed, last time we touched the topic we settled on being able to provide the ability to simply give "names" to playbooks -- so that if you wanted to give your playbook the "name" '/home/zuul/src/git.openstack.org/zuul-jobs/playbooks/foo/playbook.yml', you could do just that, but if you also happened to want to name it "foo mysql setup" you15:08
dmsimardcould do that in the same way15:08
pabelangerty for fast reviews, my ISP bill will thank you15:09
jeblairdmsimard: should we add a new field in the db for that?15:09
*** armaan has joined #openstack-infra15:09
snapiriHi Infra, continuing the problem above, I have noticed something strange, in patch the following https://review.openstack.org/#/c/509189/ I have two zuul tasks that are basically the same, one is from project-config and one is from the project itself (legacy-dragonflow-dsvm-fullstack & dragonflow-dsvm-fullstack). I have noticed that the first fails while the second succeeds.15:09
openstackgerritMerged openstack-infra/project-config master: Fix translations: Use zanata_client_version  https://review.openstack.org/51207415:09
toskyjeblair: how to debug "This change depends on a change that failed to merge." like in https://review.openstack.org/#/c/508906/ ? The only dependency (https://review.openstack.org/#/c/512027/) seems to apply cleanly to devstack-gate15:10
*** armax has quit IRC15:10
snapiriI also noticed that the filesystems are different in both cases...15:10
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Use propose-updates for requirements-constraints  https://review.openstack.org/51209015:12
dmsimardjeblair: It'd require a new field in the database yeah. The only thing I haven't quite figured out was what would be the most convenient way to provide this name. Env vars ? extra vars ? an ansible module (not unlike ara_record, could we even just use ara_record outright?) etc15:12
pabelangersnapiri: looking, but just left a comment about syntax error in .zuul.yaml15:13
dmsimardjeblair: ideally, it would not be too far off from another feature we'd like -- being able to group playbooks in a logical group (in our context, pre/run/post phases)15:14
openstackgerritMonty Taylor proposed openstack-infra/openstack-zuul-jobs master: Remove legacy proposal jobs  https://review.openstack.org/51209115:14
jeblairdmsimard: oh interesting -- an ansible module might actually be really convenient because it would let us move the specification of the name into job content rather than hard coded in zuul.15:14
pabelangersnapiri: tempest-dsvm-dragonflow tempest-dsvm-dragonflow : ERROR No valid playbook found (non-voting)15:14
pabelangersnapiri: that is likey the issue, and my comments will help fix that15:14
jeblairtosky: i went back to 512027 and did a recheck, and it now has a very strange configuration error15:15
*** armax has joined #openstack-infra15:15
jeblairmordred, pabelanger: ^ we should take a look at the error in 51202715:15
pabelangerlooking15:15
dmsimardjeblair: if you look here and expand the 'records' tab, that's freeform key/value things associated with the playbook through the ara_record module http://logs.openstack.org/07/489207/5/check/gate-ara-integration-py35-latest-fedora-26/363923b/logs/build/reports/0b4a6c53-74e6-4994-bc0e-a6585f70ab12.html15:15
mordredjeblair: looking15:16
dmsimardjeblair: http://ara.readthedocs.io/en/latest/usage.html#using-the-ara-record-module15:16
jeblairmordred: i wonder if perhaps the change to add the post-review stuff broke some of those jobs15:16
*** edmondsw has joined #openstack-infra15:16
pabelangerFirst time I've seen that error15:16
jeblairdmsimard: oh neat15:16
mordredjeblair: oh. fascinating15:16
dmsimardjeblair: ara_record can be used from inside a playbook (and will automatically tie data back to that particular playbook) or it can also be used as an 'ad-hoc' task to associate data to a playbook if you provide the id of the playbook you want to associate data with15:16
jeblairdmsimard: hrm... as i think of this further though, we really only control the base job playbooks at the level we'd need.  we can't really ask folks to put ara_record tasks in job playbooks...15:17
mordredjeblair: I don't understand why it thinks legacy-tripleo-ci-centos-7-containers-multinode is a post-review job15:18
jeblairdmsimard: i really like the ara_record idea, but i can't think of a way to use it right now... we may need to use env vars?15:18
dmsimardjeblair: yup. That's why I'm wondering what is the best way to do this from a user perspective15:18
dmsimardjeblair: Otherwise, they can be implicit tasks ran by the executor so that users don't have to worry about it.15:19
dmsimardNot unlike how we run network tests implicitely15:19
jeblairmordred: does that job inherit from publish-openstack-artifacts?15:19
jeblairthat's the thing we added post-review to15:19
modrednetfailyup15:19
modrednetfailthat would be it15:19
jeblairhttp://logs.openstack.org/27/512027/1/check/legacy-tripleo-ci-centos-7-containers-multinode/ed0e5a6/zuul-info/inventory.yaml15:20
jeblairya ^15:20
jeblairyay the system works15:20
modrednetfail\o/15:20
modrednetfailso - this should be better with the base job rework15:20
*** edmondsw has quit IRC15:21
snapiripabelanger: thanks. fixed, hope it will solve it...15:21
jeblairoh, the logs/docs/artifacts publish proposal15:21
jeblairmordred: ^?15:21
mordredyah15:21
pabelangermordred: one thing I was confused about that, we still publish to logs.o.o (for example) from executor, right?15:22
dmsimardjeblair: and then, the other thing is that ara is a bit in the same state as zuul v3 right now, as in I'm trying to keep the master branch as frozen as possible while the work goes in feature/1.0. 1.0 is not quite production ready yet though.15:22
mordredsince I belive legacy-tripleo-ci-dsvm is inheriting from publish-opnestack-artifacts so that it can do temporary artifacts or something (I'll need to go look at the why)15:23
mordredpabelanger: yes, that's right15:23
mordredpabelanger: but if we wanted to change that - or some other zuul deployer wanted to publish directly from their executors and still use zuul-jobs- that would be possible15:23
pabelanger++15:24
pabelangerokay, switching back to zuul-jobs: http://logs.openstack.org/88/512088/3/check/legacy-openstackci-beaker-ubuntu-trusty/2597cf4/job-output.txt.gz#_2017-10-15_15_18_53_40595915:24
jeblairdmsimard: implicit tasks -- how could we have the executor run a task in the context of a playbook?  surely that would be a separate run and not possible to associate with the real user-supplied playbook?15:24
pabelangerwe had SSh error mid run15:24
pabelangerlooking why now15:24
*** chlong has quit IRC15:24
pabelangerlooks to be vexxhost15:25
mordredpabelanger: I had vexxhost network issues this morning with my IRC bouncer15:25
mordredpabelanger: they seem to now be resolved - but they were bad for a while15:25
*** thorst has joined #openstack-infra15:26
pabelangerkk15:26
pabelangerI'll recheck15:26
openstackgerritMerged openstack-infra/system-config master: Fix grep error when listing ifName values  https://review.openstack.org/51208815:26
jeblairi'm sure mnaser has an actual nagios, but just in case human nagios helps here, i have said his name :)15:26
pabelangermordred: actually, just thinnking, would ssh pipe still be open? between playbook runs?15:27
mordredpabelanger: it should be - unless the network itself ceasees being able to route packets (which is what I was seeing this morning)15:27
pabelangerYah, okay15:27
pabelangermordred: jeblair: BTW: found our first use-case for zuul inline comments :) https://review.openstack.org/#/c/509189/11/.zuul.yaml15:29
pabelangerwill be awesome the day zuul is able to do that15:29
smcginnis:q15:30
*** thorst has quit IRC15:30
smcginnisBah!15:30
dmsimardjeblair: None of the ways I'm thinking about are super clean actually.. so I don't have any great suggestions right now.15:31
jeblairdmsimard: maybe both then.  :)15:32
dmsimardjeblair: oh, definitely, being able to supply this kind of information through an env variable or through a module do not need to be mutually exclusive15:32
dmsimardthere'd be ara_playbook_name env var, or something, and then ara_playbook_group for our phases15:33
mordreddmsimard: that's what we do for zuul_stream today ...15:34
mordreddmsimard: we pass a name and a phase env var15:34
dmsimardyeah, I just feel it's a bit out of scope of the callback to be doing that kind of stuff, but I can't think of another way that is better or more convenient right now so whatever :)15:35
mordredoh - scuse me - we pass those as ansible vars - zuul_execution_phase and zuul_execution_canonical_name_and_path15:35
mordredwe could pass them as env vars instead pretty easily of course15:36
dmsimardbah it can be like os.getenv('ara_playbook_name', vars['ara_playbook_name']) or something15:36
dmsimardthe callback has access to either15:36
dmsimardok, I'll go back to try and figure out why my sqlite middleware doesn't work for css and js files..15:37
mordreddmsimard, jeblair: I don't think that tripleo base job needs to be parented on legacy-openstack-publish-artifacts15:39
mordredI believe that's there for the legacy-periodic-tripleo-* jobs15:39
dmsimardmordred: yeah they were originally migrated that way, I'm not sure why. All I was told is that jobs that were found to be publishing to tarballs.o.o were automatically parented to that job.15:40
dmsimardI don't know enough about tripleo jobs to know for sure but I can dig if need be15:40
mordredah - yes - I see it in the job content now15:40
mordredit does seem that there is an artifact publisher on normal jobs15:41
*** dhinesh has joined #openstack-infra15:41
mordredplaybooks/legacy/tripleo-ci-centos-7-ovb-1ctlr_1comp_1ceph-featureset024-puppet-pike/post.yaml - for instance - is set up to copy things from workspace/images/*15:41
*** armax has quit IRC15:42
jeblairhttp://tarballs.openstack.org/tripleo-ci/images/ is empty15:46
mordredjeblair: yah - I think it's a bogus publisher - likely a macro copy-pastad on the old jjb jobs that isn't actually used15:47
jeblairi just double checked that *is* the publish location of the old version of that job15:48
jeblairhere is a list of old jobs publishing to that location: http://paste.openstack.org/show/623657/15:50
pabelangerI don't think tripleo ever switch to using tarballs.o.o for publishing yet15:50
mordredawesome15:50
AJaegermordred: thanks for working on the service-type. I like your ozj change, just one question on https://review.openstack.org/#/c/512089/1/roles/set-service-type-data-fact/README.rst15:50
mordredlet's reparent the tripleo base job off of openstack-publish-artifacts15:50
pabelangerthey still using a server in tripleo-test-cloud-rh1 for some images, IIRC15:50
pabelangermordred: +115:51
jeblairthere's a single template that all the tripleo jobs use that has that publisher15:51
jeblairmordred: ++15:51
*** slaweq has quit IRC15:52
*** liusheng has quit IRC15:52
jeblairlegacy-tripleo-ci-dsvm i think is the job in question15:52
openstackgerritMonty Taylor proposed openstack-infra/openstack-zuul-jobs master: Remove openstack-publish-artifacts from tripleo-base-dsvm  https://review.openstack.org/51209415:52
openstackgerritMonty Taylor proposed openstack-infra/openstack-zuul-jobs master: Remove openstack-publish-artifacts from legacy-tripleo-ci-dsvm  https://review.openstack.org/51209415:53
*** liusheng has joined #openstack-infra15:53
mordredAJaeger: looking15:53
mordredAJaeger: BWAHAHAH. thanks15:53
openstackgerritMonty Taylor proposed openstack-infra/openstack-zuul-jobs master: Add role to provice a service_type_data fact  https://review.openstack.org/51208915:56
mordredAJaeger: thereyago15:56
AJaegerthanks, mordred15:56
AJaegermordred: I see an extra space, let me fix...15:57
openstackgerritAndreas Jaeger proposed openstack-infra/openstack-zuul-jobs master: Add role to provice a service_type_data fact  https://review.openstack.org/51208915:57
pabelangermordred: did we ever land your patched for post-run extra debug logs?15:58
toskyprovice?15:58
pabelangerhttp://logs.openstack.org/50/511950/1/check/legacy-trove-scenario-dsvm-mariadb-single/6be1f2e/ showing POST_FAILURE and job passed15:58
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Install git-review for translation proposal job  https://review.openstack.org/51209516:00
AJaegermordred, similar to your proposal job change ^16:00
pabelangermordred: Oh, we did!16:01
jeblairmordred: can you +3 https://review.openstack.org/512084 ?  we should restart the executors with that before The Switch.16:01
jeblairi have to go sunday some things, back later16:02
pabelangergrep 6be1f2ea1439428f936b6a926673e48f /var/log/zuul/executor-debug.log16:02
mordredAJaeger: we shouldn't need thatone - the translation proposal job has propose-update as a base job - and the other patch installs git-review in a pre-playbook16:02
pabelangermordred: jeblair: ^running that on ze01 will expose a failure in logstash playboods16:02
pabelanger2017-10-15 15:48:07,395 DEBUG zuul.AnsibleJob: [build: 6be1f2ea1439428f936b6a926673e48f]         details: NoConnectedServersError('No connected Gearman servers',)16:03
AJaegermordred: ah!16:03
AJaegermordred: abandoned16:04
AJaegerfedora26 has problems "dnf.exceptions.Error: Some packages from local repository have incorrect checksum"16:05
AJaegerSee http://logs.openstack.org/89/504789/17/infra-check/multinode-integration-fedora-26/1784853/ara/16:05
AJaegerinteresting, base-integration-fedora-26 succeeded16:05
pabelangerso, looks like error is comming from gear directly16:06
pabelangerAJaeger: same error on zuulv2? Maybe our mirror our of sync16:06
AJaegerpabelanger: no idea, only happened with multinode - this is projec-tconfig, so no v2 for this job16:07
AJaegeror openstack-zuul-jobs is change 50478916:07
pabelangerlooking16:07
AJaegeranybody for quick review of adding required-repos for vmware-nsx and networking-odl? https://review.openstack.org/#/c/511836/ and https://review.openstack.org/#/c/511836/16:08
*** e0ne has joined #openstack-infra16:09
*** pcaruana has quit IRC16:10
pabelangerAJaeger: I'm going to add an autohold and see if we can debug the node16:10
mordredAJaeger: that uses system-required - do we have a patch up to change all of the mergecheck to system-required already?16:11
AJaegermordred: already in...16:11
mordredAJaeger: ah - great16:11
AJaegermordred: it's based on old version. Should we rebase first to double check?16:11
mordredAJaeger: nah - I think it's great16:12
mordredAJaeger: you pasted the same link twice above - any chance you have the other link handy?16:12
AJaegerhttps://review.openstack.org/#/c/509139/16:12
AJaegermordred: ^16:12
AJaegersorry!16:12
AJaegerhttps://review.openstack.org/509341 is now also ready - removes magnum jobs from openstack-zuul-jobs16:13
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Fix python3 unicode errors in executor server  https://review.openstack.org/51208416:14
AJaegerJust got "Could not resolve host: mirror.ca-ymq-1.vexxhost.openstack.org" - http://logs.openstack.org/94/512094/2/infra-check/multinode-integration-opensuse423/593529d/ara/16:14
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: Limit base-integration jobs  https://review.openstack.org/50755816:15
pabelangerAJaeger: mordred: dmsimard: lets see about landing ^ to help cut down on curn for ozj. We can do the same for multinode jobs too16:15
dmsimardpabelanger: oh, yeah, mordred mentioned we can filter "cross repo" too.16:16
pabelangerdmsimard: right, but we need to also run jobs on zuul-jobs16:16
AJaegerpabelanger: thanks16:16
pabelangerwhich, that patch doesn't do16:17
pabelangerso, we can follow up on that16:17
dmsimardpabelanger: what do you mean on zuul-jobs ? we already have jobs on zuul-jobs16:17
dmsimardpabelanger: but they're defined in project-config16:17
pabelangermordred: jeblair: https://review.openstack.org/511656/ could also use a +3, adds launch-timeout for citycloud, which mostly happens when we upload new images16:17
pabelangerdmsimard: oh, I thought it was removed for some reason16:17
pabelangerAJaeger: re: vexxhost DNS, I wonder if that is related to unbound16:18
openstackgerritAndreas Jaeger proposed openstack-infra/project-config master: Remove infra specific pipelines and enable pipelines  https://review.openstack.org/51089216:19
AJaegerpabelanger: yes, that might be. Not sure whether it's worth investigating, just wanted ot mention it...16:20
*** claudiub has quit IRC16:20
AJaegerpabelanger: care to +2A https://review.openstack.org/512089 , please?16:20
pabelangerAJaeger: /etc/resolv.conf on opensuse looks right16:21
AJaegerand also https://review.openstack.org/#/c/512090/ , please16:21
AJaegerpabelanger: good!16:21
pabelangerAJaeger: lets keep an eye on it, until mnaser reports back about networking issue16:21
AJaegerpabelanger: ok16:21
*** thorst has joined #openstack-infra16:21
*** thorst has quit IRC16:22
dmsimardpabelanger: added a comment on the base-integration patch16:22
*** Goneri has joined #openstack-infra16:23
openstackgerritMerged openstack-infra/project-config master: zuul v3 gate changes for vmware-nsx  https://review.openstack.org/51183616:23
dmsimardman, I can't figure out this wsgi middleware issue :(16:23
*** e0ne has quit IRC16:24
pabelangerlooking at fedora-26 issue now16:24
openstackgerritMerged openstack-infra/project-config master: networking-odl: Add required-projects for tox jobs  https://review.openstack.org/50913916:24
*** ykarel has joined #openstack-infra16:26
*** e0ne has joined #openstack-infra16:26
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Remove legacy training-guide jobs  https://review.openstack.org/51201116:29
pabelangerdmsimard: in 504789, we shouldn't be hit rdoproject.org directly. We can use the reverse proxy cache in apache16:29
*** e0ne_ has joined #openstack-infra16:30
pabelanger    "rdo_repository": "https://rdoproject.org/repos/openstack-ocata/rdo-release-ocata.rpm"16:30
pabelangerbut, that's why the fedora node is failing, just something I noticed16:30
AJaegerdmsimard: did you see jeblair's comment on https://review.openstack.org/#/c/511953/ ?16:31
pabelangerdmsimard: in fact, we likey can stop doing RPM install of the release rpm, and just create the repo list ourself16:31
pabelangerthen we don't need to test we can unintsall it16:31
*** ykarel has quit IRC16:32
*** e0ne_ has quit IRC16:32
dmsimardAJaeger: I have earlier, will be able to fix it later16:32
*** slaweq has joined #openstack-infra16:32
AJaegerdmsimard: shall I +A?16:32
AJaegerdmsimard: or do you want to wait?16:32
*** e0ne has quit IRC16:33
dmsimardAJaeger: it can be a follow-up16:33
*** e0ne has joined #openstack-infra16:34
AJaegerok, I'll +A16:34
*** markvoelker_ has quit IRC16:34
AJaegermordred: what do you think of https://review.openstack.org/507558 ?16:36
dmsimardAJaeger: in the meantime, it'll at least prevent us from breaking the gate by merging something bad in fetch-zuul-cloner16:36
pabelangerAJaeger: dmsimard: I've left a comment on 504554 too, about rdo-release rpm file also.16:36
pabelangerwhich has already been merged16:36
dmsimardpabelanger: let me see16:37
*** jtomasek has joined #openstack-infra16:37
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Add required repos for legacy-check-osc-plugins  https://review.openstack.org/51202616:40
dmsimardpabelanger: hm, why would fedora be setting up the rdo repos in the first place, that's probably part of the issue16:40
dmsimardthat's a weird failure http://logs.openstack.org/89/504789/17/infra-check/multinode-integration-fedora-26/63202cd/ara/result/92aedddb-4ac6-41cb-a739-0836f4b45783/16:40
dmsimardbut I know where that's coming from16:40
pabelangeryah, that is what I am looking at now16:41
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Add integration tests for fetch-zuul-cloner  https://review.openstack.org/51195316:42
pabelangerI think the best step is, template repo list for openvswitch repo, disable it by default, then have package task use --enable-repo when intalling openvswitch, then we don't have to worry about removing .list, as it is disabled16:42
pabelangerwe do that with epel and haveged today16:43
dmsimardpabelanger: so the issue is that https://github.com/openstack-infra/zuul-jobs/blob/master/roles/multi-node-bridge/tasks/common.yaml#L17 should not be os_family RedHat16:43
dmsimardpabelanger: it should probably be os_distribution CentOS, right ?16:44
pabelangerdmsimard: is openvswitch in fedora?16:44
pabelangercan't remember16:44
pabelangerif so, they yah, we just want it for centos16:44
dmsimardI agree that we'll want to stop relying on a package to set up the repos16:46
dmsimardpabelanger: fedora has ovs, yeah: https://koji.fedoraproject.org/koji/packageinfo?packageID=1345516:47
dmsimardpabelanger: let me submit a stopgap fix to address fedora, we can think about templating it after16:48
*** e0ne_ has joined #openstack-infra16:49
pabelangergreat16:49
*** kiennt26 has joined #openstack-infra16:50
dmsimardpabelanger: ohh, I know why I did it that way..16:51
*** e0ne has quit IRC16:51
*** e0ne_ has quit IRC17:01
*** e0ne has joined #openstack-infra17:03
*** slaweq has quit IRC17:03
*** edmondsw has joined #openstack-infra17:04
*** jtomasek has quit IRC17:04
dmsimardpabelanger: is there a bug in gerritbot's config ? I saw a zuul-jobs patch trigger a notification in #zuul but not #openstack-infra but openstack-infra/zuul-jobs is indeed there: https://github.com/openstack-infra/project-config/blob/master/gerritbot/channels.yaml#L106017:05
AJaeger"No connected Gearman servers" - is that known?17:06
*** e0ne_ has joined #openstack-infra17:06
AJaegermordred: yeah, requirements check worked - http://logs.openstack.org/82/509782/7/check/requirements-check/be20757/ but has the above post_failure17:07
pabelangermordred: jeblair: AJaeger: ^one way to help mitigate the no connected gearman servers17:07
pabelangerremote:   https://review.openstack.org/512098 Try 3 times to submit results to logstash.o.o17:07
pabelangerI'm unure how to manually test our ansible library from an executor17:08
*** kzaitsev_pi has quit IRC17:08
pabelangerdmsimard: not sure, would need to check logs for gerritbot17:09
*** edmondsw has quit IRC17:09
AJaegerpabelanger: where does it try 3 times in the change?17:09
pabelangerAJaeger: defaults from ansible, until keyword17:09
*** e0ne has quit IRC17:09
AJaegerpabelanger: Ah, thanks17:10
pabelangerAJaeger: we can pass retries: 3 too17:10
*** kzaitsev_pi has joined #openstack-infra17:10
AJaegerpabelanger: no, I'm fine..17:11
pabelanger2017-10-15 17:02:21,774 INFO gerritbot: Sending "David Moreau Simard proposed openstack-infra/zuul-jobs master: Don't set up RDO repositories on Fedora  https://review.openstack.org/512097" to #openstack-infra17:12
pabelangerdmsimard: ^17:12
pabelangerthat is in gerritbot logs17:12
pabelangerpossible network issue?17:13
dmsimardpabelanger: -1 with comment on https://review.openstack.org/#/c/512098/17:13
dmsimardpabelanger: yeah maybe17:13
AJaegerpabelanger: could you +2A https://review.openstack.org/#/c/512089/ to move us forward with api-ref jobs, please?17:13
pabelangerdmsimard: ah, ya. I expected that17:14
pabelangerAJaeger: looking17:14
pabelangerAJaeger: Hmm, we should see why emacs was missing on trusty17:15
SamYapleno one needs emacs17:15
dmsimardpabelanger: I don't think the module exposes something that we can use to retry other than the task status itself (fail_json vs exit_json). Maybe, just maybe, we can do until result | succeeded.. but I'm not sure, let me test it17:16
pabelangerAJaeger: +317:17
AJaegerthanks17:17
pabelangerdmsimard: cool, thanks17:17
pabelangerAJaeger: also, 510982 is plan B to work around POST_FAILURES for successful jobs17:18
AJaegerironic that https://review.openstack.org/#/c/512098/ has a post_failure ;)17:18
AJaegerpabelanger: yes, understood17:18
pabelangerAJaeger: but think we are on track for fixing logstash tasks, with gearman connection errors17:18
* tosky wonders why https://review.openstack.org/#/c/512094/ is so unstable17:19
pabelangerlooks17:19
pabelangerTimeout (32s) waiting for privilege escalation prompt:17:19
pabelangerthat doesn't look right17:19
toskybefore it failed twice for other reasons (missing emacs package on trusty)17:20
pabelangeryah17:20
toskyAJaeger already retriggered it more times17:20
pabelangeryah, lets start tracking these failure in etherpad17:20
pabelangerwe shouldn't need to recheck these integration jobs17:20
dmsimardpabelanger: we can test the task status, added a comment with example17:20
pabelangerI'm looking at trusty emacs issue now17:20
pabelangerokay, trusty emacs failure was on vexxhost17:21
dmsimardemacs, for the configure_mirror integration tests ? I think I remember troubleshooting that17:21
dmsimardpabelanger: I have a e-r for privilege prompt issues https://review.openstack.org/#/c/505233/17:22
*** thorst has joined #openstack-infra17:22
pabelangerI'm starting to wonder if we should stop trying to install a package, we mostly want to see if we can update indexes properly with configure-mirror17:22
dmsimardpabelanger: the idea is that we're interested in testing that we're able to install a package, thus making sure the mirrors were configured successfully and work. If there's a failure to install a package in the integration tests, there would have probably been one had it been a legitimate job (like devstack or whatever)17:24
*** ykarel has joined #openstack-infra17:24
dmsimardWe can arguably settle for another package than emacs, I just took that because the package is available over all distros we support under the same name and it's not something installed by default17:24
openstackgerritAndreas Jaeger proposed openstack-infra/openstack-zuul-jobs master: enhance integration test files section  https://review.openstack.org/51210017:27
pabelangerdmsimard: sure, but at the rate we launch jobs, we are going to see mirror issues from running jobs way before testing configure-mirror role.  I think I'd be fine with setting up mirrors, and running update to fetch indexes. I don't think we should be testing that the mirror is actually valid17:27
*** thorst has quit IRC17:27
AJaegerpabelanger, dmsimard ^ I'll rebase once pabelanger's change is in - but this is updates the list of files for triggering the integration tests17:27
pabelangerdmsimard: mostly because, if the mirror isn't valid, we cannot fix it in confgure-mirror role17:28
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Add role to provice a service_type_data fact  https://review.openstack.org/51208917:28
*** lin_yang has quit IRC17:29
dmsimardpabelanger: if someone sends a patch to configure-mirrors that screws up the mirror configuration, you're not going to be able to tell until you try to install a package17:29
dmsimardpabelanger: the integration test is meant to make sure we don't merge something that breaks that, especially since it's included in all base jobs17:29
pabelangerdmsimard: I'd expect us not to be able to download indexes17:30
AJaegerpabelanger: now we can merge https://review.openstack.org/512086 - to use the service_type_data fact. Could you +2A, please?17:31
dmsimardpabelanger: what's 'indexes' ? The equivalent of apt-cache update/dnf makecache/yum makecache and whatever is the zypper equivalent ?17:31
pabelangerdmsimard: if we install new mirrors, in configure-mirror, we first would apt-get update, dnf update, zypper update? If that passes, to me that is good enough17:33
pabelangerinstalling emacs, doesn't really add anything more, expect validating repo state is good17:33
dmsimardpabelanger: I think that's a good thing ? I mean, the problem can be either an infrastructure one or a role one17:34
dmsimardyou don't know which kind of problem it is until you look17:34
AJaegerwow, far too many post_failures in zuulv3.o.o ;(17:34
*** slaweq has joined #openstack-infra17:34
dmsimardpabelanger: I'm not trying to be annoying here, if we want to take emacs out that's fine, but that's not going to change how legit integration jobs are going to end up failing to install packages17:36
*** kiennt26 has quit IRC17:36
dmsimardwhether that's a tripleo job, a devstack job, or whatever17:36
pabelangerdmsimard: right, but that is not a failure of configure-mirror role. It is likely an infrastrucuture issue, like you said. Which is just stopping us from landing code in ozj.17:38
pabelangerlets see what others say, but I don't really see any extra value installing emacs every ozj job run17:38
pabelangerAJaeger: looking17:39
AJaegerpabelanger: thanks. Could you also check https://review.openstack.org/#/c/512090/ ?17:39
dmsimardpabelanger: would adding a retry on the package install alleviate the issue ?17:39
*** ykarel has quit IRC17:40
*** ykarel has joined #openstack-infra17:41
clarkbcan you set files on that job to only be for the mirror role?17:42
dmsimardclarkb: yeah we're already working on it17:43
pabelangerclarkb: yah, we have already17:43
AJaegerclarkb: see https://review.openstack.org/#/c/512100/1/zuul.d/jobs.yaml - I'll update that once the parent has landed17:43
pabelangerwhich will minimize runs17:43
AJaeger(I'll restrict multinode-integration as well)17:43
* AJaeger needs to rebase 512100 first...17:43
AJaegerpabelanger: https://review.openstack.org/512097 is the RDO on Fedora setup  change if you want to +2A17:44
*** bobh has quit IRC17:44
pabelangerSo, it does look like emacs issue is related to vexxhost17:45
*** ykarel has quit IRC17:46
openstackgerritMerged openstack-infra/project-config master: Use service-types.openstack.org to get api destinations  https://review.openstack.org/51208617:46
pabelangerAJaeger: +317:46
dmsimardpabelanger: should we disable vexxhost until mnaser gets a chance to look at the issue ?17:46
*** Goneri has quit IRC17:46
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Limit base-integration jobs  https://review.openstack.org/50755817:46
dmsimardI have to afk for a while, have people over -- I can respond to pings if need be17:46
AJaegerpabelanger: https://review.openstack.org/#/c/512090/ as well, please17:47
pabelangerdmsimard: yah, we can see about making that call for zuulv3-rollout17:47
pabelangerlooking17:47
*** markvoelker has joined #openstack-infra17:48
pabelangerAJaeger: mordred: +3 with comment17:49
pabelangeron 51209017:50
*** markvoelker has quit IRC17:51
AJaegerclarkb, pabelanger, dmsimard ^17:51
AJaegerwait, out bot was silent ;(17:52
AJaegerclarkb, pabelanger, dmsimard : https://review.openstack.org/512100 Limit multinode-integration, update base-integration17:52
*** markvoelker has joined #openstack-infra17:53
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Remove openstack-publish-artifacts from legacy-tripleo-ci-dsvm  https://review.openstack.org/51209417:53
openstackgerritMerged openstack-infra/zuul-jobs master: Don't set up RDO repositories on Fedora  https://review.openstack.org/51209717:54
openstackgerritMerged openstack-infra/project-config master: Use propose-updates for requirements-constraints  https://review.openstack.org/51209017:56
*** salv-orlando has joined #openstack-infra18:00
* AJaeger will be back later - open review requests are https://review.openstack.org/#/c/512100/ and https://review.openstack.org/51209118:02
*** Hal has joined #openstack-infra18:03
*** AJaeger has quit IRC18:03
*** Hal is now known as Guest587318:03
*** slaweq has quit IRC18:08
*** liusheng has quit IRC18:08
*** armaan has quit IRC18:12
*** armaan_ has joined #openstack-infra18:12
*** liusheng has joined #openstack-infra18:16
*** modrednetfail has quit IRC18:16
*** slaweq has joined #openstack-infra18:17
*** sbezverk has joined #openstack-infra18:18
pabelangerinfra-root: I am thinking our post playbook failure task issue (POST_FAILURES on successful jobs) might be a top issue to address before rolling out this evening. This is just going to result in an infulx of rechecks, and user confusion.18:21
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Try 3 times to submit results to logstash.o.o  https://review.openstack.org/51209818:22
*** thorst has joined #openstack-infra18:24
*** yamahata has quit IRC18:28
*** thorst has quit IRC18:30
openstackgerritAndreas Florath proposed openstack/diskimage-builder master: Test elements have README.rst file  https://review.openstack.org/17783218:36
*** baoli has joined #openstack-infra18:40
*** e0ne_ has quit IRC18:41
*** thorst has joined #openstack-infra18:45
*** thorst has quit IRC18:45
*** AJaeger has joined #openstack-infra18:47
*** e0ne has joined #openstack-infra18:48
AJaegerpabelanger: I agree, the POST_FAILURES are currently our biggest - and only known - problem.18:48
*** gouthamr has joined #openstack-infra18:49
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: Legacy ansible-func jobs require openstack/requirements  https://review.openstack.org/51210218:57
*** dhinesh has quit IRC19:01
*** erlon has quit IRC19:02
*** thorst has joined #openstack-infra19:05
bodenhi… I’ve recently started getting a POST_FAILURE in our openstack-tox-pep8 job https://review.openstack.org/#/c/511542   any ideas why would be apprecaited19:10
pabelangerboden: yah, we are aware of the issue. Working towards resolving it19:11
bodenpabelanger ah ok.. thought maybe it was something I did19:11
bodenthx19:11
pabelangerboden: see 512098 and 510982 for workaround19:12
*** Goneri has joined #openstack-infra19:13
* AJaeger calls it a day, sorry, won't be around for the fun of switching ;) Hope it works out fine...19:16
*** AJaeger has quit IRC19:16
clarkbthat runs on the executors right?19:17
clarkbcan we double check all executors can connec to the gear server on logstash.o.o?19:18
*** slaweq has quit IRC19:18
pabelangerclarkb: yah, I haven't confirmed that yet19:18
*** e0ne has quit IRC19:19
SamYapleare e still t minus 3 hours from zuulv3? or have issues popped up that block that19:20
clarkbI don't think this blocks us, worst case we just ignore failures on that task while we debug it19:20
*** slaweq has joined #openstack-infra19:21
*** dhajare has joined #openstack-infra19:21
pabelangerclarkb: yah, I suggest we land 510982, then work to revert it. Unless we can figure out why we fail on connecting to gearman19:22
fungiyeah, we seem to have confirmed the post failures at least mostly stem from logstash indexing, so worst case we make it always claim to have succeeded and live with gaps in logstash until we work out the cause19:22
*** Goneri has quit IRC19:27
pabelangerclarkb: fungi: do we have any logs on logstash.o.o we can check for gearman errors?19:28
fungii guess log-gearman-client.py is the gearman server?19:31
pabelangeryah, thought so too, but seems to be empty19:31
pabelangerwill see if I can find out why19:31
pabelangerfungi: clarkb: if you are also in the mood to review: https://review.openstack.org/512102/ will help zuulv3 jobs for ansible projects19:32
fungipabelanger: i was just looking at a legacy-ansible-func-centos-7 failure over missing openstack/requirements in its required-projects19:33
fungiwas about to try and find whether someone had already claimed it in the issues pad19:33
pabelangeryah19:33
pabelangerI haven't added it to zuulv3-issue, I'll do that now19:34
fungiso anyway, approved!19:34
fungiit's like i imagine the solutions to these problems and suddenly patches materialize19:34
pabelangerwouldn't that be the best19:34
fungiyeah, too bad it doesn't occur with greater frequency19:38
fungiinfra-root: reminder, the rollout/rollback plan for tonight is at https://etherpad.openstack.org/p/zuulv3-rollout if anyone else wants to double-check it before we start19:39
clarkbit forks a geard like zuul does iirc19:39
clarkbso should have its own logs19:40
jeblairi don't see anything about the gearman error in the etherpad19:40
fungiwe should probably send a one-hour warning as a #status notice in about an hour and 20 minutes19:40
jeblairis it still an issue?19:40
fungijeblair: i think pabelanger only just confirmed it from the executor debug log a couple hours ago19:40
fungibut yeah, there were a rash of post_failure states on otherwise successful jobs just a little bit ago19:41
jeblairwe don't even need to confirm things before adding them to the etherpad -- we have a triage section19:41
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Legacy ansible-func jobs require openstack/requirements  https://review.openstack.org/51210219:41
*** armaan_ has quit IRC19:42
jeblairi will start adding things19:42
jeblairbut if we're going to be an effective distributed team, we need to record *everything*19:42
pabelangeragree, I've only logged them in IRC currently. Will review / update fungi entries19:43
openstackgerritDavid Moreau Simard proposed openstack-infra/openstack-zuul-jobs master: Add integration tests for multi-node-bridge role  https://review.openstack.org/50478919:43
jeblairugh, apparently we get an exception but no trace back19:45
openstackgerritDavid Moreau Simard proposed openstack-infra/openstack-zuul-jobs master: Add integration tests for multi-node-bridge role  https://review.openstack.org/50478919:45
fungiany objection to me archiving the current mass of fixed issues in the zuulv3-issues pad to the zuulv3-fixed-issues pad?19:45
fungiit's taking up nearly half the issues pad again already at this point19:46
pabelangerwfm19:46
fungiaha, and in so doing i see the current concern with the post_failure results got moved to the fixed issues section after a debugging patch was merged. not sure that's the sanest way of tracking things19:47
dmsimardpabelanger: +1 with a comment on https://review.openstack.org/#/c/512098/19:49
pabelangerdmsimard: yah, you are not wrong19:50
pabelangerdmsimard: feel free to -1 until I fix19:50
dmsimardok19:50
*** pcaruana has joined #openstack-infra19:51
*** slaweq has quit IRC19:54
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Remove legacy proposal jobs  https://review.openstack.org/51209119:55
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Remove legacy magnum jobs  https://review.openstack.org/50934119:55
jeblairthe previous errors all seemed to happen around in a cluster around the same time  (twice)19:56
jeblairthat corresponds with an increased load average on the logstash server -- it almost reached 1.19:56
pabelangerYah, I was thinking maybe it was server side also. I haven't been able to see where our gearman server logging goes19:57
jeblairthis is still pretty perplexing because it should be waiting for a connection and only proceeding when it gets one19:57
jeblairpabelanger: you proposed one or more mitigation patches, right?  can you link them?19:58
jeblairi've started a tcpdump on the logstash server on the gear port; hopefully that will shed light on it if it happens again19:58
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Try 3 times to submit results to logstash.o.o  https://review.openstack.org/51209819:59
pabelangerjeblair: ^ is 119:59
pabelangerjeblair: https://review.openstack.org/510982/ is 2nd19:59
jeblairpabelanger: i'm in favor of the 2nd now20:00
*** pcaruana has quit IRC20:00
pabelangeryah, worth a try20:00
jeblairpabelanger: i don't think i want to do the first -- i think it may actually make the situation worse.  and if we've managed to submit some jobs before failing, we could end up submitting duplicates which could be weird20:00
pabelangerkk20:01
pabelangerjeblair: shall I remove WIP on 1st?20:01
jeblairpabelanger: ya, maybe just go ahead and +W it20:02
pabelangersure20:02
pabelangerdone20:02
pabelangerI'll WIP the other20:02
*** slaweq has joined #openstack-infra20:04
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Add 90s connect timeout for submit log jobs  https://review.openstack.org/51210520:04
jeblairthis is *not* a solution to the problem, but while i was looking at it, i thought about that ^.  we may appreciate having that in there some day in the future.20:04
jeblairwe could also shorten that -- i was just thinking about maybe a really overburdened executor.20:05
clarkbcould it be remote running out of file descriptors?20:05
jeblairclarkb: yes.20:05
clarkbis it connection per file or per job?20:05
jeblairconnection per job, but maybe too many jobs finished at once20:06
jeblairhrm, we should see this on the connection graph though20:06
jeblairthere are only tiny spikes on the tcp connection graph around the times we got errors20:06
*** baoli has quit IRC20:08
jeblairit's possible that they were very brief, since the spike time periods were long enough to show up on the graph at all, i would expect to see the count get significantly higher20:08
jeblairso, possible but lack of evidence for the moment20:09
openstackgerritMerged openstack-infra/project-config master: base: ignore_errors for logstash processing  https://review.openstack.org/51098220:09
ianwo/20:10
jeblairi'm going to restart the executors now, to pick up the unicode fix20:10
pabelanger++20:12
dmsimardianw: hey o/20:14
jeblairfascinating20:14
jeblair2 executors did not shutdown cleanly20:14
dmsimardianw: I have my middleware thing half working20:14
jeblairone of them has an open finger connection20:14
jeblairi wonder if that's keeping it alive20:14
jeblair(it's not so much the executor itself as a subprocess i think)20:15
jeblairi'm going to stop zuul-web and see if this dies20:15
jeblairyes, both gone now20:16
jeblairadding to etherpad20:16
pabelangergood to know20:16
jeblairexecutors have been restarted (also zuul-web)20:17
*** d0ugal_ has joined #openstack-infra20:20
*** d0ugal has quit IRC20:21
*** claudiub has joined #openstack-infra20:32
*** slaweq has quit IRC20:36
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: Remove legacy-windmill jobs  https://review.openstack.org/51211020:39
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: Remove windmill from zuul-legacy-project-templates  https://review.openstack.org/51211420:46
*** slaweq has joined #openstack-infra20:47
*** felipemonteiro has joined #openstack-infra20:49
*** felipemonteiro_ has joined #openstack-infra20:51
*** felipemonteiro has quit IRC20:54
*** nunchuck has joined #openstack-infra20:54
clarkblogstash's gear process seems to have up to 8192 fds and currently has a couple hundred so I doubt that is the issue21:02
*** e0ne has joined #openstack-infra21:02
*** lennyb has quit IRC21:02
pabelangerwe have a few locked ready nodes, over 3hours and counting, 2 of them are over 1 day21:07
pabelangertrying to see why that would be21:07
clarkbI've tested telnet to logstash:4730 on all executors, ze10 felt more laggy but it eventually gets the status output21:07
*** baoli has joined #openstack-infra21:08
pabelanger2 ubuntu-trusty nodes, which are ready and locked for 1d8hrs seem to be from periodic wheel-build jobs for openstack/requirements21:09
pabelanger300-0000423830 is request ID, but not sure why still locked21:09
*** hongbin has joined #openstack-infra21:10
pabelangeradding it to etherpad zuulv3-issues21:10
pabelangerthey are in citycloud-kna1, so it is possible related to quote21:11
pabelangerquota*21:11
pabelangerI've added them to issues with nodepool21:11
fungistatus notice Zuul v3 rollout maintenance begins at 22:00 UTC (roughly 45 minutes from now): http://lists.openstack.org/pipermail/openstack-dev/2017-October/123618.html21:12
pabelangergoing to grab food with family quickly before rollout starts21:12
fungithat look good to send in a few minutes?21:12
pabelangerfungi: ++21:12
*** baoli has quit IRC21:13
clarkbfungi: ya21:15
clarkblooks like es07 is out to lunch21:16
clarkbes logs show it happened on the 13th21:16
clarkbI can't hit it via ssh21:16
clarkbgoing to try a reboot via openstack ap21:16
fungi#status notice Zuul v3 rollout maintenance begins at 22:00 UTC (roughly 45 minutes from now): http://lists.openstack.org/pipermail/openstack-dev/2017-October/123618.html21:17
openstackstatusfungi: sending notice21:17
openstackgerritIan Wienand proposed openstack-infra/openstack-zuul-jobs master: Remove bashate legacy jobs  https://review.openstack.org/51211721:17
clarkb(noticed when digging into logstash related stuff)21:17
Shrewsoh my. lots of weekend scrollback21:18
-openstackstatus- NOTICE: Zuul v3 rollout maintenance begins at 22:00 UTC (roughly 45 minutes from now): http://lists.openstack.org/pipermail/openstack-dev/2017-October/123618.html21:18
openstackstatusfungi: finished sending notice21:20
*** felipemonteiro_ has quit IRC21:20
*** slaweq has quit IRC21:20
clarkbit pings but doesn't ssh guessing due to fsck21:20
* clarkb tries to be patient21:20
evrardjpthanks for the notice and happy rollout!21:21
*** e0ne has quit IRC21:21
*** slaweq has joined #openstack-infra21:23
*** e0ne has joined #openstack-infra21:24
*** bobh has joined #openstack-infra21:24
*** armax has joined #openstack-infra21:26
openstackgerritIan Wienand proposed openstack-infra/project-config master: Move bashate jobs in-tree  https://review.openstack.org/51211821:26
clarkbssh -vvv seems to indicate port 22 is listening21:27
clarkbjust no one is home21:27
clarkbanyone else have better luck than me to elasticsearch07.openstack.org?21:27
ianwclarkb: it does not seem to like australians either21:30
clarkbwill probably need to dig in via the console once zuul things are done21:31
*** e0ne has quit IRC21:31
*** wolverineav has quit IRC21:31
pabelangersame, on es07, down21:32
pabelangerShrews: I've added an issue to nodepool section, FYI21:33
pabelangerShrews: I _think_ it might be quota related, but not 100% sure21:33
Shrewspabelanger: possibly. how many nodes did the request those are for request?21:34
ianwhmm, when you watch this console scroll past more slowly you see things ... all the java deps, generating a bunch of odd locales, etc.  few things to dig into there21:34
pabelangerShrews: where should I look again for that?21:34
Shrewspabelanger: nodepool request-list21:34
clarkbianw: that for es07?21:34
clarkbianw: or bashate?21:35
ianwclarkb: sorry, no, just noticing things on zuulv3 jobs that i probably didn't before as i never really watched the console scroll by21:35
clarkbah21:35
*** lennyb has joined #openstack-infra21:35
pabelangerShrews: okay, so i did check that, but I don't see a matching request ID21:35
* mordred waves to all the nice people21:36
pabelangermordred: ^521:36
Shrewspabelanger: hmm21:36
Shrewspabelanger: if the request is gone, then it's not a quota issue21:37
Shrewswe need to see who has those locked21:37
pabelangerI think zuul did21:40
pabelangerI can see the request ID in zuulv3 debug.log.1.gz21:40
pabelangerbut not sure I understand what is happening21:40
clarkbthere are hung process timeouts from the kernel on es07's console21:41
clarkbits up and login prompt is there but it looks like maybe an unhappy hypervisor21:41
clarkbI'm going to leave it there though and focus on zuul nowish21:41
*** eumel8 has quit IRC21:44
jeblairo/21:51
fungii have basically finished my baking tasks for today (and also recompiled a kernel package to enable a staging tree driver for an sdio-based wifi interface)21:51
fungiready to get down and get zuul21:51
fungii guessing no objections/concerns with https://etherpad.openstack.org/p/zuulv3-rollout21:52
jeblairfungi: lgtm21:52
fungithe pipeline export/import stuff was more relevant when this was a thursday plan. may not be super relevant at today's volume21:53
clarkbmy only suggestion would be to maybe move the nodepool quotas first as that should help things get merged more quickly21:53
clarkbbut either order should work fine21:53
pabelanger+121:54
mordred++21:54
fungiyeah, i suppose shifting quota first sounds better21:54
fungiand we could get that rolling now if we wanted?21:54
jeblairfungi: ++21:55
mordred++21:55
pabelanger++21:55
jeblairthat's https://review.openstack.org/51106921:55
mordred+2 from me21:55
fungi+321:56
mordredI've got "/opt/system-config/production/tools/kick.sh 'nodepool.openstack.org:nl0*.openstack.org'" prepped and ready to go21:58
Shrewspabelanger: i'm confused as to why those nodes are still locked22:00
* fungi pulls up browser and watches virtual paint peel22:00
pabelangerShrews: yah, me also22:00
openstackgerritMerged openstack-infra/project-config master: Revert "Revert "Shift nodepool quota from v2 to v3""  https://review.openstack.org/51106922:01
clarkbmordred: ^22:01
mordredwoot. on it22:01
fungii guess we can status notice that we're starting maintenance now... probably no need for alert since we're not really taking stuff down in ways that will be overly apparent to users (other than people staring at the old status screen)?22:01
pabelangerShrews: 0000230528 is another node that I don't know why it is locked and ready22:02
pabelangerShrews: again, nothing in nodepool request-list22:02
clarkbfungi: ya22:02
jeblairfungi: ++22:02
pabelangerfungi: wfm22:02
fungistatus notice Zuul v3 rollout maintenance is underway, scheduled to conclude by 23:00 UTC: http://lists.openstack.org/pipermail/openstack-dev/2017-October/123618.html22:02
fungithat look good?22:02
jeblairfungi: ++22:03
fungi#status notice Zuul v3 rollout maintenance is underway, scheduled to conclude by 23:00 UTC: http://lists.openstack.org/pipermail/openstack-dev/2017-October/123618.html22:03
pabelanger++22:03
openstackstatusfungi: sending notice22:03
-openstackstatus- NOTICE: Zuul v3 rollout maintenance is underway, scheduled to conclude by 23:00 UTC: http://lists.openstack.org/pipermail/openstack-dev/2017-October/123618.html22:04
mordredkick is done22:04
* smcginnis grabs popcorn for Zuul 3: The Awakening22:05
pabelangerupdating zuulv3-rollout22:05
fungismcginnis: well, more like the reanimation22:05
smcginnis;)22:05
toskysmcginnis: I thought it was Zuul 3: Return of the Job22:06
fungiheh22:06
* mordred has +2d the infra-pipeline revert patches22:06
smcginnisZuul 3, A New Hope?22:07
mordredwouldn't that be 4? ;)22:07
fungiso for line 4, are we missing changes for zuul-jobs or project-config (just infra-gate and infra-post dropping from the latter)?22:07
pabelangerconfirmed, both nodepool and nodepool-launcher appear correct22:07
smcginnismordred: Depends if you actually want to count all of the first three. ;)22:08
fungiahh, 510892 is project-config, down on line 722:08
jeblairi saved the queues22:08
jeblairi'll work on porting the gate script to v322:08
jeblairer it's one line22:08
jeblairdone22:08
fungijeblair: for the record, "porting is just appending --tenant=openstack right? or is there more?22:09
pabelangerhow did we handle ready nodes in nodepool.o.o last time?22:09
jeblairfungi: that's it22:09
pabelangerthey should get released once zuul-launcher is stopped right?22:09
fungik\22:09
jeblairall the patches we need to land are gated by v3 now, right?22:10
jeblairso i can stop v2?22:10
*** baoli has joined #openstack-infra22:10
fungiyes22:10
*** wolverineav has joined #openstack-infra22:10
jeblairdone22:10
mordredbuh-bye22:10
pabelanger\o/22:10
clarkbI'm confused by 0892 still has infra-check in it22:11
clarkbis that a bug?22:11
jeblairclarkb: we're keeping infra-check around so we can land project-config changes quickly22:11
fungiclarkb: we're retaining infra-check solely for project-config22:11
fungi(temporarily)22:11
mordredclarkb: it's not a bug, we're keeping it :)22:11
jeblairclarkb: (it is high-priority, so will return faster than normal check pipeline)22:11
clarkbgotcha22:11
clarkbwe didn't want to leave the other repos on that pipeline too? ozj zuul etc?22:12
jeblairclarkb: nah, depends-on works with those, so we're not in such a bad place22:12
*** bobh has quit IRC22:12
jeblair(and we could still force-enqueue in gate if needed urgently for something widespread)22:12
clarkbare we ready to start approving these changes then?22:13
fungialso things landing to those other repos will mostly be under the guidance of people with the ability to enqueue them directly to the gate pipeline if there is an urgent reason to do so22:13
*** dhajare has quit IRC22:13
jeblairclarkb: i think so22:13
fungier, what jeblair said22:13
*** baoli has quit IRC22:14
* clarkb goes to approve things22:14
*** wolverineav has quit IRC22:14
pabelangerdo we need to do anything to nodepool.o.o to clean out ready nodes?22:14
jeblairpabelanger: yeah, deleting those at this point is a good idea22:14
clarkbI approved the first 322:14
pabelangerjeblair: okay, on it22:15
fungii went ahead and put old zuul.o.o in the emergency disable list so we don't need to wait for puppet to finish a round later22:16
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Revert "Use new post pipelines"  https://review.openstack.org/51088722:17
mordredfungi: +22:17
*** rkukura_ has quit IRC22:17
clarkbthe zuul changes don't appear to want to queue22:17
clarkbI'm guessing because we told them to use gate but the gate criteria are not currently satisfiable22:17
openstackgerritJames E. Blair proposed openstack-infra/puppet-zuul master: Move statsd to correct zuul.conf file  https://review.openstack.org/51212022:17
mordredclarkb: maybe recheck?22:17
clarkbmordred: no I think its not going to work bceause gate needs +322:18
clarkbI think we need https://review.openstack.org/#/c/510892/ in first so the depends on order is wrong22:18
clarkbpabelanger: ^ fyi22:18
clarkbshould I go ahead and remove the depends on to the zuul changes and approve 0892?22:19
mordredclarkb: I think we need to split https://review.openstack.org/#/c/510892/22:19
clarkbI don't think we need to split it22:19
clarkbwe just need to merge the zuul changes afterwards22:20
mordredclarkb: we need to do the bit with the pipeline definition - then the other 3 patches, then https://review.openstack.org/#/c/510892/22:20
pabelangeryah, I orginally left a comment to split 510892 up22:20
clarkbmordred: why?22:20
mordredclarkb: we can't delete infra-gate and infra-post until we've landed the other 3 changes22:20
*** sdague has joined #openstack-infra22:20
clarkbmordred: I don't think that is tre22:20
*** armax has quit IRC22:21
clarkbone of the three already merged and it is the only one that needs to go first22:21
clarkbthe other two should work with gate once criteria are upated (since that is why they are not gating now)22:21
mordredright - but the criteria update patch also deletes the infra-gate - which is a syntax error until the patch stopping using infra-gate lands22:22
clarkbmordred: right the other two patches stop using infra-gate22:22
clarkbwhich is why they won't gate now22:22
jeblairwe should have a change which only re-enables the pipelines22:23
jeblairthen we should switch everything to use those pipelines22:23
mordredyes. I'm working on that now22:23
jeblairthen we should delete the unused pipelines22:23
clarkbI'm still not understanding why that is necessary22:23
clarkbthe two zuul changes should work once 0892 is in because they switch to using the normal gate pipeline22:23
mordredno - it'll be a broken zuul config22:24
clarkbthis is why they can't gate right now22:24
jeblairclarkb: except that zuul currently uses those pipelines22:24
mordredbecause those configs refernece infra-gate22:24
clarkbbceause they switch to gate22:24
jeblairclarkb: 0892 removes them22:24
clarkbmordred: they delete infra-gate22:24
*** markvoelker has quit IRC22:24
jeblairclarkb: right, which is why 0892 depends on them22:24
mordredyes. but they still use infra-gate before theyland22:24
mordredthe repos do22:24
clarkbhttps://review.openstack.org/#/c/510886/3/.zuul.yaml and https://review.openstack.org/#/c/510885/2/.zuul.yaml22:24
mordredthe consistent global zuul config22:24
jeblairclarkb: if we were to hypothetically force-merg 510892 right now, zuul's config would be broken22:24
mordredlanding the project-config patch is a syntax error22:24
jeblairclarkb: because the zuul repo has a .zuul.yaml which references a pipeline which no longer would exist22:25
clarkbI see. Is that something we can make zuul handle more gracefully? eg if a trusted repo redefines things then just ignore the untrusted?22:25
jeblairclarkb: if you want to confirm, you can go ahead and drop the depends-on from 510892 and zuul should report that error in just a couple seconds22:25
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Re-enable zuul v3 pipelines  https://review.openstack.org/51212122:25
clarkbSeems like we would be in a really weird spot if we ever tried to move things around in the future otherwise22:26
mordredclarkb: I think what zuul is doing is exactly correct22:26
jeblairclarkb: zuul won't let us22:26
mordredanyway - that patch ^^ should work22:26
clarkbright we'd have to go hunt down all the random places things are used in potentially thousands of repos22:26
jeblairit is not possible to break zuul's config as long as zuul is gating22:26
clarkbya the weird spot isn't zuul would be broken its affecting the change you want in the system22:27
mordredclarkb: yes. zuul will not allow us to break people's job configs in their repos - that's a good thing - it means we physically cannot land a change to the global config that would break the job config that is in someone's repo ... if there are 1000s of repos using a thing, we likely shouldn't be breaking it22:27
jeblairlarge systemic changes will require more planning and communication in the future.22:27
mordredyup22:27
jeblairthat's not bad, it's just different.  :)22:28
SamYapleso zuul actually loads all the configs from openstack/* repos everytime it runs?22:28
clarkbI can imagine situations where not being able to update config without first updating thousands of repos is a bad thing :)22:28
jeblairSamYaple: yes.  it uses its fleet of mergers and executors to grab every config file from every branch of every project.  it currently takes 6 minutes from a cold start.22:29
jeblairSamYaple: (that will get faster after we cannibalize some mergers from v2)22:29
jeblairclarkb: like?22:29
SamYapleyowza22:29
clarkbjeblair: mostly worried about needing to deal with all the dead repos out there22:29
jeblairclarkb: if they're dead, drop them from zuul.22:29
mordredyah22:29
SamYapleall this talk about unique named jobs and global namespaces makes alot more sense now22:29
jeblairSamYaple: the cool thing is -- it *only* takes 6 minutes!  there are about 6000 project-branches in the system.22:30
mordredSamYaple: the config is also updated automatically every time a change lands to any of the zuul configs in any of the repos - plus there is the speculative config bulding22:30
clarkbjeblair: mordred so if you don't update within X time you will be removed from the system?22:30
pabelangerSamYaple: Yah, that was part of the work jeblair did this past week to make reloads faster! Last roll out it was taking about 50secs pre reload, not I think it is down to a few seconds22:30
jeblairSamYaple: yeah, once loaded, zuul config is incrementally updated in about 6 seconds22:30
SamYaplejeblair: yea im not saying thats a long time. im just impressed by the scale22:30
mordredclarkb: yah- but this is also for systemic things that should be very infrequent ... such as "remove the gate pipeline" or "rename the base job"22:30
*** esberglu has joined #openstack-infra22:31
clarkbmordred: I think a good example in the past was how we redefined release and tag pipelines at one point but ya22:31
mordredclarkb: that won't be a problem22:31
mordredclarkb: it's only the name reference that's an issue - so deleting things from project-config will take effort22:32
SamYaplei cant imagine python is nearly the fastest (or memory effiecent) way of pulling in all that stuff either22:32
mordredclarkb: so redefining the qualities of the release pipeline - no issue - deleting the release pipeline - hard22:32
*** markvoelker has joined #openstack-infra22:32
SamYaplethe mind boggles22:32
clarkbalso removing eg silent (which has been talked about, but maybe if people are using it then we should just keep it around)22:32
mordredclarkb: yah - we a) should totally remove it... as well as experimental22:33
jeblairclarkb: zuul answers the question "is anyone still using this?" very quickly :)22:33
SamYaplehaha22:33
pabelangerindeed22:33
mordredbut b) we just have to clean house first22:33
mordredalso - yah - we should delete experimental pretty soon after this dust settles22:33
jeblairmordred: i'm not sure we should remove those22:33
mordredno?22:33
SamYaplewhy remove experimental?22:34
*** salv-orlando has quit IRC22:34
mordredexperimental has no purpose in a world with speculative job configs from my pov -  just propose a patch with the job you want to be experiemntal22:34
mordredand depends-on/recheck it as needed22:34
jeblairmordred: it's not just an on-ramp for new jobs, it's also "on-demand check jobs".22:34
clarkbmordred: I think its purpose would be having predefined useful jobs you want to run occasionally22:34
jeblairmaybe it needs a new name22:34
fungithe experimental pipeline is arguably a bit more convenient than reparenting a change that tests adding the jobs you're curious to see whether or not work22:34
SamYaplefungi: ++22:34
mordredyah - I can see on-demand check jobs you want to use occasionally - the 'experimental' use case has a whole different mechanism now22:35
jeblairmordred: i agree we should drop it from the process for adding new jobs though22:35
mordredjeblair: ++22:35
jeblairmaybe call it 'extra' or something22:35
fungiexperimental is less useful for testing out new jobs now, but it's nice for jobs you want to run infrequently on-demand for specific changes22:35
jeblairor frank22:35
jeblair"check more"22:35
SamYaplecheck moar22:35
openstackgerritMerged openstack-infra/project-config master: Re-enable zuul v3 pipelines  https://review.openstack.org/51212122:35
fungicheck harder22:35
jeblairit can have more than one trigger even :)22:35
* clarkb rechecks the two zuul changes22:36
SamYaplei would like to purpose aliases as "harder, better, faster, stronger"22:36
* jeblair plays 4 non blondes22:37
pabelangerwe're up to 234 nodes 'in-use' on nodepool-launchers now22:37
*** markvoelker has quit IRC22:37
SamYaplejeblair: heh. theres something i havent heard in a while22:37
jeblair"bigger, better, faster, more"22:37
mordredclarkb: I see them in the gate now22:37
SamYapleya22:37
pabelangerand load on zuulv3.o.o is nice and low22:38
pabelangerYah, i see 510886,3 in gate pipeline22:39
clarkbmordred: did you want to rebase 0892 as it should confict with 12121 now22:39
clarkbI guess we want zuul changes to merge first22:39
clarkbso that it can check validity of new ocnfig properly?22:39
mordredclarkb: yah - let's rebase when they land22:41
*** markvoelker has joined #openstack-infra22:41
openstackgerritMerged openstack-infra/zuul master: Revert "Use new infra pipelines"  https://review.openstack.org/51088622:44
pabelangerwoot22:44
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Revert "Use new infra pipelines"  https://review.openstack.org/51088522:45
clarkbthats both of them22:46
*** markvoelker has quit IRC22:46
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Remove infra specific pipelines and enable pipelines  https://review.openstack.org/51089222:46
mordredthere's that rebased22:46
*** dizquierdo has joined #openstack-infra22:48
fungiare we good to approve it at this stage?22:48
mordredyah22:49
fungidone22:49
*** jkilpatr_ has joined #openstack-infra22:49
mordredso - as soon as that lands, v3 is 're-enabled'22:49
fungiit's actually enqueued in infra-check, so not a syntax error at least22:49
mordred\o/22:49
fungiand now it's in infra-gate22:50
fungiwhich should hopefully no longer exist as soon as it merges22:50
fungimildly mind-bending22:50
mordredright?22:50
*** markvoelker has joined #openstack-infra22:50
clarkband this one only works because we don't look at its future state22:51
clarkb?22:51
mordredyup22:51
* mordred watching to see the pipeline go away22:51
jeblairi think the mouseover time estimates aren't working; something to fix after mnaser's status page refresh22:52
openstackgerritMerged openstack-infra/project-config master: Remove infra specific pipelines and enable pipelines  https://review.openstack.org/51089222:53
pabelangerodd, I didn't see ^ on zuulv3.o.o status page22:53
* mordred did22:53
clarkbheh that queued things to infra-post22:53
clarkbthen removed it22:54
jeblairand it's gone :)22:54
pabelangerjust did CTRL-f5 to be sure22:54
mordred\o/22:54
jeblairpabelanger: it was in infra-gate22:54
jeblairrip22:54
pabelangerah22:54
pabelanger++22:54
jeblairi will enqueue the one change from zuulv2 now22:54
mordredjeblair: do you think we can handle the load?22:54
jeblairwe will find out22:55
mordredneutron change just hit the gate22:55
jeblairthat's the one22:55
*** markvoelker has quit IRC22:55
pabelangeryah, zuul-scheduler is playing it nice and cool right now :D22:55
jeblairpabelanger: can you +3 https://review.openstack.org/512120 please?22:55
pabelangerlooking22:55
jeblairunfortunately, after that lands we're going to need a full zuulv3 restart to get stats.  sorry.22:56
pabelangerjeblair: +322:56
mordredwe should always do zuul v3 rollouts on sundays22:56
fungiit's been long enough, i've gone ahead and set the by-hand redirect on zuul.o.o so consumers of the old status.json are getting the new one22:57
*** dizquierdo has quit IRC22:57
mordrednice22:57
jeblairfungi: thx.  i approved 51106722:57
pabelangerfungi: ++22:57
fungier, whoops, typo on my redirect though, fixing22:58
jeblairwhen that lands, i think we're done?22:58
mordredjeblair: I agre with you22:58
jeblairit really did take us 1 hour to flip that really big switch22:58
fungiokay, fixed now22:58
fungisilly $ backreferences instead of \22:58
fungiwho wants to send the celebratory announcement to the ml?22:59
*** markvoelker has joined #openstack-infra22:59
mordredthe vertical space under infra-check is interesting22:59
pabelangerand POST_FAILURES look to be mitigated with our ignore_errors patch to logstash task23:00
fungimordred: i think it's trying to grid the layout or something there. it's doing similar vertical whitespace under experimental too for me23:00
jeblairpabelanger: have there been more errors?23:00
pabelangerjeblair: I haven't see any yet23:01
jeblairpabelanger: oh, how do you know it's mitigated?23:01
*** slaweq has quit IRC23:01
mordredjeblair: neutron 511960,1 is in both check and gate - is that expected right now?23:01
pabelangerjeblair: yah, maybe not the correct word. I maybe fixed but ignored?23:02
jeblairmordred: probably was in check when v2 was running (it's been in there for > 1 hour)23:02
fungimordred: it's expected if they rechecked it23:02
mordredjeblair: ah - yes - good point23:02
fungimordred: though that does not appear to be the case23:02
mordrednice to see legacy-grenade-dsvm-neutron-dvr-multinode be success :)23:03
fungioh, yep23:03
fungiright, so it made it into the gate in v2 while it was still getting check resources in v323:03
jeblairpabelanger: well, that would be great, but again, it'd be nice if there were evidence of that.  I don't see "No connected" in the executor logs, so i don't know that we have any evidence that the problem has happened again.23:03
fungiand then got enqueued into the gate in v323:03
jeblairpabelanger: and we *need* it to happen again so we can fix it.23:04
pabelangeragree23:04
jeblairpabelanger: so if we've masked it too much, we'll have to unmask it.23:04
pabelangeryah23:04
jeblairbut i think we can give it a day or so and see if it shows up in the log23:04
*** jkilpatr_ has quit IRC23:05
pabelangerkk, see a job issue networking-cisco23:06
mordredpabelanger: I feel like I +2d a networking-cisco job recently23:06
pabelangermordred: yah, checking zuulv3-issues now23:06
pabelangerhttp://logs.openstack.org/54/505654/5/check/openstack-tox-compare-cover/bd44cbb/ara/result/15ca4ff6-30da-4370-b77f-a1c94f25aca5/23:06
mordredpabelanger: sounds like our friend-neighborhood "neutron missing" issue23:07
*** slaweq has joined #openstack-infra23:08
pabelangeryar23:08
mordredpabelanger: easy fix I think - networking-cisco seems to be the only one using that job :)23:08
pabelangermordred: yup, just checking to see of we have an existing project-template some place23:09
pabelangerbut will push up a patch in a minute23:09
mnaserso everyone is taking the day off tomorrow to celebrate zuulv3 v2? hah.23:09
openstackgerritMerged openstack-infra/puppet-zuul master: Move statsd to correct zuul.conf file  https://review.openstack.org/51212023:09
mordred++23:09
mordredmnaser: flying to copenhagen is the weirdest way I've ever celebrated :)23:10
mnaseroh fun23:10
jeblairmordred: raise a glass of swamp water!23:10
*** baoli has joined #openstack-infra23:11
jeblairfungi, mordred: i didn't see anyone jump on the followup 'all-done' email... are either of you inspired?  if not, i can, but i'll probably just say "done".  :)23:11
clarkbfungi: did we get npm sorted out for the status page?23:11
fungijeblair: i'm happy to do it23:12
fungimordred probably needs to pack23:12
fungiclarkb: good question, lemme see if i dropped the ball on that patch23:12
jeblairfungi: all yours!23:12
clarkbfungi: iirc it had problems with trusty?23:13
clarkbbasically fixed for one platform but not the other23:13
fungihttps://review.openstack.org/50856423:13
fungistill needs work apparently23:13
*** claudiub has quit IRC23:13
clarkbfungi: could also be puppet loading up path details before the package gets installed?23:14
clarkbI know the pip package module did/does that23:14
fungipossibly23:14
jeblairmordred: +3 512027 ?23:14
*** baoli has quit IRC23:15
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Add neutron as require-project for networking-cisco jobs  https://review.openstack.org/51212223:15
jeblairclarkb: +3 508906 ?23:16
clarkbpabelanger: I think its required on all those jobs (pep8, py27, etc) isn't it?23:16
openstackgerritMerged openstack-infra/system-config master: Revert "Revert "Add redirect from status.o.o/zuul to zuulv3.openstack.org""  https://review.openstack.org/51106723:16
mordredjeblair: dne23:17
jeblairand that's our last rollout change merged!23:17
pabelangerclarkb: shouldn't be, there are templates for that I think23:17
*** dingyichen has joined #openstack-infra23:17
clarkbpabelanger: except its overridden there to require neutronclient and each job is specified by hand?23:17
pabelangerclarkb: openstack-python-jobs-neutron and openstack-python-jobs-neutron23:17
clarkbpabelanger: ya but they aren't using those23:17
pabelangerclarkb: doesn't it append?23:17
pabelangerhttp://logs.openstack.org/54/505654/5/check/openstack-tox-pep8/20c7a7f/job-output.txt.gz23:18
clarkbpabelanger: oh wait they use both the template and spcify the jobs23:18
pabelangerthat is from working networking-cisco job23:18
clarkbthat seems like a bug23:18
mordrednah -it's on purpose23:18
pabelangerclarkb: yah, not sure23:18
mordredthe template defines neutron23:18
mordredthe local define adds neutronclient23:18
pabelangerYah, that's was the first time I seen that23:18
clarkbthat is probably worthy of a comment at least23:18
pabelangerkinda clever23:19
clarkbjeblair: kind of confused by your comment on 8906. Seems like you are saying the files restriction should be removed but +2'd the change?23:20
jeblairclarkb: sorry, i mean that the files restriction is correct because it will still run the job on .zuul.yaml changes23:21
pabelangerfungi: clarkb: think I see issue with trusty job on 50856423:21
pabelangerI can push up patch23:21
fungipabelanger: thanks!!!23:21
jeblairclarkb: oh, i think i misread andreaf's comment23:21
jeblairclarkb: your confusion makes more sense if i read that one right.  :)23:21
jeblairclarkb: yes.23:22
jeblairclarkb: tell you what, i'll re-comment, and then make a followup change to drop it23:22
clarkbjeblair: thanks23:22
openstackgerritMerged openstack-infra/project-config master: Add neutron as require-project for networking-cisco jobs  https://review.openstack.org/51212223:22
jeblairclarkb: that way, if you are otherwise okay with it, we can keep the check result and go ahead and merge 90623:22
openstackgerritMerged openstack-infra/project-config master: Add 90s connect timeout for submit log jobs  https://review.openstack.org/51210523:23
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Limit multinode-integration, update base-integration  https://review.openstack.org/51210023:24
openstackgerritMerged openstack-infra/project-config master: Add OpenStack client check to python-rsdclient  https://review.openstack.org/51165023:24
jeblairclarkb: remote:   https://review.openstack.org/512123 Zuul: remove file matcher for devstack job23:25
pabelangerfungi: Oh, maybe not...23:26
pabelangerwe should be installing npm from nodejs puppet manifest I think23:26
*** abelur_ has quit IRC23:26
jeblairi'm going to put some nails in some things, but will check back in a bit23:26
fungiinfra-root: draft announcement is https://etherpad.openstack.org/p/2FSNrxRQGZ23:27
mordredfungi: lgtm23:27
ianwjeblair: https://review.openstack.org/#/c/508906/ ... that's the "devstack" job right?23:28
pabelangerfungi: ++23:28
clarkbjeblair: andreaf I've approved the devstack job addition and +2'd the followup to run it more23:29
fungii'll go ahead and send what i've got there23:30
clarkbjeblair: when do you want to get the statsd pickup restart in?23:31
clarkb(that is going to be important for groups like neutron so earlier the better I think)23:31
pabelangerYah, forgot about that23:32
dmsimardcloudnull, evrardjp, logan-, odyssey4me: looks like OSA jobs aren't working on v3. http://logs.openstack.org/15/512115/1/check/legacy-openstack-ansible-openstack-ansible-ceph-centos-7/43c8bd9/job-output.txt.gz#_2017-10-15_21_00_54_60185223:33
dmsimardI don't see any ongoing patches to add the required projects for the OSA jobs, is there any WIP ?23:33
fungiclarkb: which statsd changes? jeblair restarted the executors a little while ago to pick some up23:34
* mordred needs to step away for a bit - it seems like things are in good shape though23:34
*** hongbin_ has joined #openstack-infra23:35
*** hongbin has quit IRC23:35
clarkbfungi: https://review.openstack.org/512120 jeblair said we will need a full zuul restart to pick that up23:35
clarkbwe should double check puppet has applied it too23:36
mnaseris it possible the `use-cached-repos` role is broken? :X23:36
dmsimardwow what is this repo, I've never seen it before o_O https://github.com/openstack/openstack23:36
mnaser ubuntu-xenial | error: Could not remove config section 'remote.origin'23:36
clarkbdmsimard: its a repo that absues gerrit behavior around sub modules to track all the merges to openstack (though I don't know that we've properly updated it since the big tent)23:37
pabelangermnaser: got log?23:37
clarkbmnaser: have a link? origin isn't expected to exist so I don't think that is entirely unexpected23:37
mnaserthe job is currently being retried23:38
mnaserlet me see23:38
mnaser(cause it happened/failed in pre)23:38
mnaserhttp://zuulv3.openstack.org/static/stream.html?uuid=0767a19b1ec8490987b6398ccee613ef&logfile=console.log23:38
mnaserwatch this if you manage to catch it, it should fail23:38
clarkbmnaser: seemd to work ?23:39
mnaseryeah it just went through23:39
clarkbat least use-cached-repos went by successfully23:39
mnaseri could swear its failed before, let me find a log before that it failed23:39
mnaseri saw it once and recheck'd23:39
mnaserhttp://logs.openstack.org/02/509202/3/check/puppet-openstack-integration-4-scenario002-tempest-ubuntu-xenial/61735de/job-output.txt.gz#_2017-10-15_08_43_53_89187823:39
mnaserthis is from today23:39
clarkbianw: I think dtroyer mentioned that he had seen hardlink errors recently. Do you know if the simplification for hardlink checking merged?23:40
*** abelur_ has joined #openstack-infra23:40
ianwclarkb: not in my queue, so must have23:41
ianwis it in the etherpad (looking...)23:41
*** abelur_ has joined #openstack-infra23:41
*** slaweq has quit IRC23:41
pabelangermnaser: looking23:41
mnaseroh wait a second23:41
mnaserit runs use-cached-repos twice?23:41
mnaserand the second run fails23:42
pabelangeryah23:42
pabelangerI see that23:42
*** tosky has quit IRC23:42
*** bobh has joined #openstack-infra23:42
pabelangerwow23:42
pabelangeryour playbooks order is amazing23:42
pabelangerhttp://logs.openstack.org/02/509202/3/check/puppet-openstack-integration-4-scenario002-tempest-ubuntu-xenial/61735de/ara/23:43
pabelangerwhat is going on there23:43
ianwclarkb: "not permitted" ... that seems different, and i wonder if it's the user stuff i pointed out it was doing23:43
mnaseris it possible my pre's are messing things up?23:43
clarkbianw: oh with the perms?23:43
pabelangermnaser: yes23:43
ianwclarkb: yeah, just above it -> http://logs.openstack.org/15/509015/2/check/osc-functional-devstack/26b5228/job-output.txt.gz#_2017-10-14_02_23_08_39901423:43
clarkbianw: could be since hardlinking needs perms on the source side right?23:43
pabelangeryou seem to have created a new base?23:43
mnaseri have prepare-node-common.yaml which is the base of all puppet jobs, then prepare-node-integration which is the base of integration jobs23:44
clarkbreading that role it blindly deletes origin23:44
mnaserit..worked okay before23:44
clarkbso if origin isn't there because it already deleted it once then ya it should fail23:44
mnaserbut pre.yaml is running twice23:44
clarkbI think what we want t o do is make that git remote rm origin only happen if the clone task above it ran23:45
* clarkb makes a patch23:45
mnaserclarkb isnt this a bug though?23:45
pabelangerI'm trying to understand your playbooks now23:45
mnaserif you look at the ARA output, pre.yaml ran twice23:45
mnaseronce before the entire job, once before the actual job was going to start23:46
pabelangerbecause you are some how running playbooks before base/pre.yaml23:46
pabelangerwhich, I do not know how is possible23:46
clarkbmnaser: yes it is a bug that it is running twice23:46
pabelangerin fact, base/pre.yaml is run twice23:46
openstackgerritMerged openstack-infra/puppet-openstack_health master: Don't set npm_package_ensure  https://review.openstack.org/50856423:46
mnaseri think whats happening is that its running before the entire "tree" of jobs start, then it runs again after all the pre's of the parents have ran23:47
mnaser oh you know what23:48
mnaserpabelanger do you think that because `puppet-openstack-integration-4` has no pre: it default to the base one?23:48
clarkbya maybe the extra belt and suspenders not necessary since this is only happening due to a different bug that should be fixed23:48
mnaserhttps://github.com/openstack/puppet-openstack-integration/blob/master/.zuul.yaml#L301-L304 has no pre, so maybe pre gets set to the $world default pre, which runs twice23:49
pabelangermnaser: no, it should be like an onion23:49
ianwclarkb / dtroyer : possible quick fix, use $USER https://review.openstack.org/512125 ... but i can all be rewritten i think23:49
mnaserpabelanger oh an interesting note is that this is running in a stable branch23:50
mnaseri am not sure if thats related, but i dont think ive seen a similar failure in master23:50
* mnaser looks23:50
*** stakeda has joined #openstack-infra23:50
*** slaweq has joined #openstack-infra23:51
pabelangermnaser: do you have .zuul.yaml on stable branches?23:51
clarkbmnaser: pabelanger the zuul_info invetnroy iirc should log the inheritance path23:51
mnaserpabelanger that was the patch to get .zuul.yaml into stable23:51
mnaserclarkb ok ill have a look23:51
pabelangerclarkb: it does and looks correct23:51
mnaserhttp://logs.openstack.org/04/509204/1/check/puppet-openstack-lint/0767a19/ <-- officially the most "recent" failure23:52
pabelangerclarkb: I think this might be a jeblair or mordred issue23:52
pabelangerbut, IIRC, we should always be running base/pre.yaml before anything else23:52
pabelangerand have no idea why twice23:52
*** bobh has quit IRC23:53
pabelangermnaser: and no failure on master?23:53
mnaserpabelanger nope, trying to find a stable branch change but none logged by zuul23:53
mnaseroh23:53
mnaserhttps://review.openstack.org/#/c/511862/23:53
mnasera stable change that didnt fail23:54
mnaser(or not the same failure anyways)23:54
pabelangermnaser: so, first question. Why adding .zuul.yaml to stable branches? and not just manage it from master?23:54
pabelangerbecause, they look like same jobs, yes?23:55
mnaserpabelanger there was actual changes that we had to do to some parts of the code, ex: changes in `functions` wrt ZUUL_URL no longer being set23:55
mnaserand some stuff relating to remotes being removed from zuul23:55
mnaser*zuulv323:55
mnaserso job would pretty much fail without those changes23:55
pabelangermnaser: right, I you'd be able to make those changes, but keep .zuul.yaml in master branch only I think23:56
pabelangermnaser: I admit, I have looked at your jobs or patches23:56
*** felipemonteiro has joined #openstack-infra23:56
pabelangerbut, I'm going to defer to jeblair on this one, I am not sure what is happening atm23:57
clarkbdoes it consistently fail?23:57
clarkbor if you recheck on a single change sometimes it works?23:57
mnaserpabelanger at the time, fungi also mentioned it would be a good idea for jobs to have their own zuul.yaml files so that changes don't break old jobs anymore without knowing23:57
openstackgerritIan Wienand proposed openstack-infra/openstack-zuul-jobs master: Remove bashate legacy jobs  https://review.openstack.org/51211723:57
mnaserclarkb yes, consistently, fi you look at zuul's status ,you can see them all in retry_failure across all stable branches23:57
mnaserthe weird thing is this used to work23:58
*** felipemonteiro_ has joined #openstack-infra23:58
mnaseror maybe it didnt, i see RETRY_FAILURE way back, then normal fails and RETRY_FAILURE again23:58
mnaseron october 13th, it did fail "properly" - http://logs.openstack.org/02/509202/3/check/puppet-openstack-integration-4-scenario001-tempest-centos-7/ab9f239/ara/23:59
pabelangeryah, that one looks right23:59
*** felipemonteiro_ has quit IRC23:59
*** felipemonteiro_ has joined #openstack-infra23:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!