Thursday, 2017-08-10

*** thorst has joined #openstack-infra00:04
openstackgerritMerged openstack-infra/system-config master: Also add buildlogs.cdn.centos.org  https://review.openstack.org/49225600:09
*** dingyichen has joined #openstack-infra00:12
*** sflanigan has joined #openstack-infra00:16
*** sflanigan has joined #openstack-infra00:16
*** thingee_ has quit IRC00:19
openstackgerritJeremy Stanley proposed openstack-infra/system-config master: Limit PTL rolls to foundation members  https://review.openstack.org/49232900:21
*** kornicameister has quit IRC00:25
*** kornicameister has joined #openstack-infra00:26
ianwgeez i hate the total lack of formatting available in launchpad00:29
ianwclarkb: the plot thickens on the pypi error mismatches -> https://bugs.launchpad.net/openstack-gate/+bug/1708707/comments/100:29
openstackLaunchpad bug 1708707 in OpenStack-Gate "Pip finds hash mismatch for package during installation" [Undecided,New]00:29
*** thorst has quit IRC00:31
*** gildub has joined #openstack-infra00:34
ianwit seems like it thought it got it from the mirror, but didn't actually00:36
*** yamamoto_ has quit IRC00:37
*** yamamoto has joined #openstack-infra00:39
*** kornicameister has quit IRC00:40
*** Apoorva_ has quit IRC00:43
*** sbezverk has quit IRC00:44
*** kornicameister has joined #openstack-infra00:45
*** xinliang has joined #openstack-infra00:45
*** xinliang has quit IRC00:45
*** xinliang has joined #openstack-infra00:45
*** markvoelker has joined #openstack-infra00:45
*** liujiong has joined #openstack-infra00:51
pabelangercool: http://mirror.regionone.infracloud-vanilla.openstack.org:8080/buildlogs.cdn.centos/centos/7/cloud/x86_64/openstack-pike/00:52
pabelangerlooks to be caching things now00:52
*** _ryan_ has quit IRC00:53
pabelangerwe likely can update configure-mirrors.sh variables and hit it directly to avoid the redirect00:55
pabelangerhttp://paste.openstack.org/show/617998/00:55
*** EricGonczer_ has quit IRC00:57
*** eharney has joined #openstack-infra00:59
mnaserfungi great, ill be able to bump it up once raw images are uploaed01:00
*** bobh has joined #openstack-infra01:02
*** rhallisey has quit IRC01:02
*** jkilpatr has quit IRC01:04
fungipost jobs all have nodes now, so we seem to finally be caught up for the day01:05
openstackgerritPaul Belanger proposed openstack-infra/system-config master: Replace buildlogs.centos with buildlogs.cdn.centos  https://review.openstack.org/49233601:06
*** dave-mccowan has joined #openstack-infra01:07
openstackgerritMohammed Naser proposed openstack-infra/project-config master: Fix Grafana graphs for VEXXHOST  https://review.openstack.org/49233801:08
*** thorst has joined #openstack-infra01:08
*** thorst has quit IRC01:08
*** thorst has joined #openstack-infra01:09
pabelangerfungi: I'm hoping in the next 2 hours change pipeline for tripleo will also look much better01:12
*** gildub has quit IRC01:12
*** xarses_ has quit IRC01:12
pabelangerI think that was part of the reason infracloud was seeing networking issues, every request to mirror was 302 redirected to internet01:13
*** rwsu_ has quit IRC01:13
*** thorst has quit IRC01:13
pabelangerhopefully this will conserve some bandwidth once cached01:14
*** mwarad has joined #openstack-infra01:16
fungimakes sense01:17
*** tuanluong has joined #openstack-infra01:19
openstackgerritPaul Belanger proposed openstack-infra/tripleo-ci master: Stop trying to build networking-bagpipe with DLRN  https://review.openstack.org/49233901:24
*** thorst has joined #openstack-infra01:25
*** thorst has quit IRC01:27
*** kornicameister has quit IRC01:30
*** mwarad has quit IRC01:30
*** bobh has quit IRC01:35
*** kornicameister has joined #openstack-infra01:35
*** ramishra has quit IRC01:39
*** rwsu has joined #openstack-infra01:41
openstackgerritPaul Belanger proposed openstack-infra/system-config master: Add registry.npmjs.org reverse proxy cache  https://review.openstack.org/45772001:42
pabelangerfungi: clarkb: ianw: ^next one up, npm reverse proxy cache, just seen a job in infracloud timeout fetching npms01:43
*** cuongnv has joined #openstack-infra01:44
mnaserim not seeing any data in graphite for jobs that run on our cloud and searching logstash for "node_provider:vexxhost-ca-ymq-1" yields no results01:44
mnaseri know jobs are running.. but is there a reason why they might not be reporting ?01:45
mnaserlogstash job queue doesnt look too bad01:45
*** gongysh has joined #openstack-infra01:45
pabelangermnaser: node_provider:"ymq-1"01:49
pabelangerhttp://logs.openstack.org/33/492133/1/check/gate-tempest-dsvm-neutron-full-ubuntu-xenial/6505200/console.html timedout on vexxhost01:50
mnaserpabelanger oh i see, no idea why, but good to know01:50
pabelangerpossible that job was warming up the afs cache01:51
mnasercould be, it probably hasnt been touched for a long time and seems like its the only timeout.. ill keep an eye01:53
*** zhaozhenlong has joined #openstack-infra01:57
*** cshastri has joined #openstack-infra01:58
openstackgerritPaul Belanger proposed openstack-infra/elastic-recheck master: Add query for bug 1709744  https://review.openstack.org/49234202:01
openstackbug 1709744 in OpenStack-Gate "Gem fetch networking errors" [Undecided,New] https://launchpad.net/bugs/170974402:01
*** aeng has quit IRC02:03
*** _ryan_ has joined #openstack-infra02:04
dmsimardianw: so platform:redhat is only RHEL and CentOS ?02:07
dmsimardI don't seem to see a different classifier but there is platform:fedora02:08
pabelangermnaser: I think we might need to turn off vexxhost for a bit, I don't think hostname is setup on our nodes02:15
pabelangerhttp://logs.openstack.org/57/475457/4/gate/gate-tripleo-ci-centos-7-containers-multinode/54b3deb/console.html ran on vexxhost, but hostnamectl was empty02:15
*** spligak has quit IRC02:16
mnaserpabelanger is that something that we have to do on our side?02:16
pabelangermnaser: we setup hostname with glean IIRC02:17
openstackgerritYAMAMOTO Takashi proposed openstack-infra/project-config master: networking-midonet: Add runtime graphs to grafana dashboard  https://review.openstack.org/49235002:17
pabelangerso because it is not running, its likely possible we don't get a valid setting02:18
mnaserit is very much possible because i remember --ssh and --hostname was only being called when it found the configdrive02:18
pabelangermnaser: I propose we stop vexxhost for tonight and work on getting a new release of glean tomorrow02:18
mnaserpabelanger no problem02:19
mnaseri rather not disturb builds in a critical time02:19
pabelangermnaser: mind propose a patch for max-server 0?02:19
mnasersure 1 second02:19
pabelangermnaser: agree02:19
* fungi is on hand to review such02:22
openstackgerritMohammed Naser proposed openstack-infra/project-config master: Set max-servers down to 0 for VEXXHOST  https://review.openstack.org/49235102:23
mnaserpabelanger fungi ^02:23
*** Sukhdev_ has quit IRC02:23
pabelanger+2 thanks02:24
fungidone and done02:25
fungithanks again mnaser!02:25
*** yamahata has quit IRC02:25
*** eharney has quit IRC02:26
*** dave-mccowan has quit IRC02:27
*** jamesmcarthur has joined #openstack-infra02:27
mnaserfungi np, im writing a small glean patch now02:27
*** iyamahat has quit IRC02:28
*** thorst has joined #openstack-infra02:28
*** hongbin has joined #openstack-infra02:28
*** thorst has quit IRC02:29
*** kornicameister has quit IRC02:29
*** Guest93343 has quit IRC02:29
*** ramishra has joined #openstack-infra02:30
*** jamesmcarthur has quit IRC02:32
*** Hal has joined #openstack-infra02:32
*** Hal is now known as Guest5475902:32
*** mwarad has joined #openstack-infra02:34
openstackgerritMohammed Naser proposed openstack-infra/glean master: Add checks for uppercase config drive label  https://review.openstack.org/49235302:35
mnaserpabelanger ^ if you wanna go over that tomorrow :)02:35
mnaseri tested the detection logic locally but not all of glean to be fully honest02:35
*** kornicameister has joined #openstack-infra02:35
*** esberglu has quit IRC02:40
*** LindaWang has joined #openstack-infra02:42
*** EricGonczer_ has joined #openstack-infra02:46
*** hongbin has quit IRC02:48
*** _mwarad_ has joined #openstack-infra02:49
*** hongbin has joined #openstack-infra02:49
*** dklyle has quit IRC02:49
*** zhurong has joined #openstack-infra02:51
*** mwarad has quit IRC02:52
*** hongbin has quit IRC02:52
*** hongbin has joined #openstack-infra02:53
*** zhaozhenlong has left #openstack-infra02:54
*** bunnyKun has joined #openstack-infra02:54
*** dave-mccowan has joined #openstack-infra02:54
*** yamahata has joined #openstack-infra02:55
openstackgerritMerged openstack-infra/project-config master: Set max-servers down to 0 for VEXXHOST  https://review.openstack.org/49235102:56
*** markvoelker has quit IRC03:02
*** markvoelker has joined #openstack-infra03:03
ianwdmsimard: sorry, i'm not following?03:14
*** david-lyle has joined #openstack-infra03:22
*** ramishra has quit IRC03:28
*** ramishra has joined #openstack-infra03:29
*** nicolasbock has joined #openstack-infra03:29
*** cshastri has quit IRC03:32
*** yamamoto has quit IRC03:36
*** bunnyKun has quit IRC03:37
*** bunnyKun has joined #openstack-infra03:41
*** EricGonczer_ has quit IRC03:41
*** ramishra has quit IRC03:42
*** hongbin has quit IRC03:43
*** cshastri has joined #openstack-infra03:45
*** _mwarad_ has quit IRC03:47
*** yamamoto has joined #openstack-infra03:47
*** dave-mccowan has quit IRC03:48
*** yamamoto has quit IRC03:48
*** yamamoto has joined #openstack-infra03:49
*** ramishra has joined #openstack-infra03:50
*** kornicameister has quit IRC03:53
*** kornicameister has joined #openstack-infra03:54
*** luzC has quit IRC03:57
*** markvoelker has quit IRC04:01
*** markvoelker has joined #openstack-infra04:01
*** bunnyKun has quit IRC04:03
*** cshastri has quit IRC04:03
*** thorst has joined #openstack-infra04:04
*** rmcall has quit IRC04:07
*** thorst has quit IRC04:09
*** spligak has joined #openstack-infra04:12
*** ykarel|away has joined #openstack-infra04:18
*** Sukhdev has joined #openstack-infra04:20
openstackgerritEric Kao proposed openstack-infra/project-config master: Increasing job timeout  https://review.openstack.org/49237304:25
*** thorst has joined #openstack-infra04:26
johnsomAny thoughts on these instant timeouts?  http://logs.openstack.org/43/491643/2/gate/gate-octavia-python35/aa4c23b/console.html04:26
johnsomI guess the ansible is failing04:27
johnsomhttp://logs.openstack.org/43/491643/2/gate/gate-octavia-python35/aa4c23b/_zuul_ansible/ansible_log.txt04:28
*** cshastri has joined #openstack-infra04:28
*** dhajare has joined #openstack-infra04:29
johnsomThe error was: OSError: [Errno 2] No such file or directory: '/home/jenkins/workspace/gate-octavia-python35'04:29
*** gongysh has quit IRC04:30
*** thorst has quit IRC04:30
*** gouthamr has joined #openstack-infra04:31
*** sbezverk has joined #openstack-infra04:35
*** adisky__ has joined #openstack-infra04:37
*** hareesh has joined #openstack-infra04:38
*** sbezverk has quit IRC04:40
*** rmcall has joined #openstack-infra04:40
*** claudiub has joined #openstack-infra04:40
*** squid has joined #openstack-infra04:41
openstackgerritMerged openstack-infra/project-config master: [magnum] Move -nv test to experimental  https://review.openstack.org/49217704:42
*** calebb has quit IRC04:43
*** squid is now known as calebb04:43
*** david-lyle has quit IRC04:44
*** rmcall has quit IRC04:45
*** dklyle has joined #openstack-infra04:45
*** cshastri has quit IRC04:46
mnaserwould anyone be able to review this? https://review.openstack.org/#/c/491800/04:50
*** calebb has quit IRC04:52
*** squid has joined #openstack-infra04:53
*** squid is now known as calebb04:53
*** rwsu has quit IRC04:58
openstackgerritArtur Basiak proposed openstack-infra/project-config master: Provide unified gate configuration  https://review.openstack.org/49079004:58
*** gongysh has joined #openstack-infra05:09
*** sree has joined #openstack-infra05:11
openstackgerritMerged openstack-infra/system-config master: Add Fedora Atomic mirrors  https://review.openstack.org/49180005:13
mnaserthank you ianw05:14
*** gouthamr has quit IRC05:16
openstackgerritArtur Basiak proposed openstack-infra/project-config master: Change service name  https://review.openstack.org/49237905:29
*** Sukhdev has quit IRC05:30
*** luzC has joined #openstack-infra05:31
*** jamesdenton has quit IRC05:36
*** sshnaidm|off has quit IRC05:37
*** luzC has quit IRC05:37
*** liujiong has quit IRC05:44
*** liujiong has joined #openstack-infra05:45
*** ccamacho has left #openstack-infra05:47
*** ccamacho has quit IRC05:47
*** armax has joined #openstack-infra05:48
*** e0ne has joined #openstack-infra05:48
*** armax has quit IRC05:48
*** _ryan_ has quit IRC05:51
*** psachin has joined #openstack-infra05:52
*** luzC has joined #openstack-infra05:54
*** ykarel_ has joined #openstack-infra05:54
*** luzC has quit IRC05:57
*** jamesdenton has joined #openstack-infra05:57
*** ykarel|away has quit IRC05:57
openstackgerritArtur Basiak proposed openstack-infra/project-config master: Provide unified gate configuration  https://review.openstack.org/49079005:59
*** yamamoto has quit IRC06:01
*** e0ne has quit IRC06:03
*** dhajare has quit IRC06:04
*** rcernin has joined #openstack-infra06:05
*** slaweq has quit IRC06:05
*** yamamoto has joined #openstack-infra06:05
*** dhajare has joined #openstack-infra06:06
*** rwsu has joined #openstack-infra06:07
*** pgadiya has joined #openstack-infra06:07
*** kjackal_ has joined #openstack-infra06:08
*** junbo has quit IRC06:12
*** cshastri has joined #openstack-infra06:14
*** junbo has joined #openstack-infra06:15
*** jamesdenton has quit IRC06:16
*** martinkopec has joined #openstack-infra06:19
*** dhajare has quit IRC06:20
*** dhajare has joined #openstack-infra06:23
*** thorst has joined #openstack-infra06:26
*** jamesmcarthur has joined #openstack-infra06:28
*** tnovacik has quit IRC06:29
*** abelur_ has quit IRC06:29
*** slaweq has joined #openstack-infra06:29
*** thorst has quit IRC06:31
*** jamesmcarthur has quit IRC06:32
openstackgerritArtur Basiak proposed openstack-infra/project-config master: Change service name  https://review.openstack.org/49237906:32
*** eranrom has joined #openstack-infra06:33
*** jamesdenton has joined #openstack-infra06:34
*** dhajare has quit IRC06:37
kklimondawhat's responsible for displaying test result table on gerrit review pages? is it a plugin, or a config knob?06:38
*** jamesdenton has quit IRC06:39
openstackgerritIan Wienand proposed openstack/diskimage-builder master: [DNM] lvm testing of spanned vg  https://review.openstack.org/49198606:39
*** jamesdenton has joined #openstack-infra06:48
*** ykarel_ is now known as ykarel06:48
*** ccamacho has joined #openstack-infra06:57
*** shardy has joined #openstack-infra07:09
*** luzC has joined #openstack-infra07:10
*** psachin has quit IRC07:12
*** gcb has joined #openstack-infra07:13
*** tesseract has joined #openstack-infra07:13
*** luzC has quit IRC07:14
*** luzC has joined #openstack-infra07:14
*** luzC has quit IRC07:18
*** markus_z has joined #openstack-infra07:23
*** aarefiev_afk is now known as aarefiev07:29
*** dizquierdo has joined #openstack-infra07:36
*** luzC has joined #openstack-infra07:38
openstackgerritMerged openstack-infra/project-config master: networking-midonet: Add runtime graphs to grafana dashboard  https://review.openstack.org/49235007:39
*** martinkopec has quit IRC07:39
openstackgerritMerged openstack-infra/project-config master: Reduce infracloud by 50%  https://review.openstack.org/49220707:40
openstackgerritMerged openstack-infra/project-config master: Fix Grafana graphs for VEXXHOST  https://review.openstack.org/49233807:40
*** eranrom has quit IRC07:40
*** cshastri has quit IRC07:41
*** LindaWang has quit IRC07:42
*** LindaWang has joined #openstack-infra07:43
*** LindaWang has quit IRC07:45
*** LindaWang has joined #openstack-infra07:45
*** xinliang has quit IRC07:48
*** ralonsoh has joined #openstack-infra07:48
openstackgerritDima Kuznetsov proposed openstack-infra/project-config master: Dragonflow: increase timeout for fullstack jobs  https://review.openstack.org/49241607:48
*** alexchadin has joined #openstack-infra07:48
yuvalHello infra! Tons of jobs seem to be queued (some for over 8 hours)07:51
yuvalmaybe nodepool is stuck? (again :\ )07:52
yuvalAJaeger_: yolanda: ?07:52
openstackgerritBogdan Dobrelya proposed openstack-infra/tripleo-ci master: Rework the getthelogs helper script for wget recursive  https://review.openstack.org/49217807:52
yuvalianw: fungi: ?07:52
*** martinkopec has joined #openstack-infra07:53
openstackgerritMerged openstack-infra/project-config master: Add release permission for neutron-vpnaas and dashboard  https://review.openstack.org/49167007:54
*** sflanigan has quit IRC07:54
*** makowals has quit IRC07:56
*** makowals has joined #openstack-infra08:00
*** cshastri has joined #openstack-infra08:00
*** xinliang has joined #openstack-infra08:00
*** xinliang has quit IRC08:00
*** xinliang has joined #openstack-infra08:00
openstackgerritYAMAMOTO Takashi proposed openstack-infra/project-config master: networking-midonet: Remove v2 jobs from grafana dashboard  https://review.openstack.org/49242408:01
*** martinkopec has quit IRC08:03
*** markus_z has quit IRC08:03
*** martinkopec has joined #openstack-infra08:03
*** markus_z has joined #openstack-infra08:04
*** jpena|off has quit IRC08:12
openstackgerritMerged openstack-infra/project-config master: Update footer to follow docs.o.o  https://review.openstack.org/49098908:16
*** mriedem has quit IRC08:16
*** electrofelix has joined #openstack-infra08:19
*** xinliang has quit IRC08:23
*** lucas-afk is now known as lucasagomes08:25
*** derekh has joined #openstack-infra08:25
openstackgerritBogdan Dobrelya proposed openstack-infra/tripleo-ci master: Rework the getthelogs helper script for wget recursive  https://review.openstack.org/49217808:25
*** jcjo has joined #openstack-infra08:26
*** thorst has joined #openstack-infra08:27
*** yamamoto has quit IRC08:27
*** yamamoto has joined #openstack-infra08:28
*** psachin has joined #openstack-infra08:29
*** jcjo has quit IRC08:30
*** dingyichen has quit IRC08:31
*** yamamoto has quit IRC08:31
*** yamamoto has joined #openstack-infra08:32
*** thorst has quit IRC08:32
*** jaosorior has quit IRC08:37
*** dizquierdo has quit IRC08:37
*** xinliang has joined #openstack-infra08:39
*** xinliang has quit IRC08:39
*** xinliang has joined #openstack-infra08:39
*** jaosorior has joined #openstack-infra08:41
*** ociuhandu has quit IRC08:41
ianwyuval: hmm ... looking08:45
yuvalianw: some are stuck for over 20 hours08:45
*** ykarel is now known as ykarel|lunch08:46
*** jtomasek has joined #openstack-infra08:47
ianwit looks like jobs have timed out, but zuul has not cleaned them up08:47
tobiashkklimonda: it's partly gerrit config (commentlinks) and partly gerrit css config08:48
tobiashkklimonda: you should find both somewhere in the system-config repo08:48
kklimondathanks, I'll take a look08:48
openstackgerritIlya Shakhat proposed openstack-infra/project-config master: Add gerritbot notications for osprofiler into #openstack-performance  https://review.openstack.org/49243708:50
*** eroux has joined #openstack-infra08:51
tobiashkklimonda: that should be the css part: https://github.com/openstack-infra/system-config/blob/master/modules/openstack_project/files/gerrit/GerritSite.css#L12308:51
*** xinliang has quit IRC08:52
*** jaosorior has quit IRC08:54
*** makowals has quit IRC08:54
ianwno, that's not right, it is moving along.  i think we're just down a couple of providers08:57
tobiashkklimonda: in zuul you also need to configure job_name_in_report: https://github.com/openstack-infra/puppet-zuul/blob/master/templates/zuul.conf.erb#L2508:58
*** makowals has joined #openstack-infra08:58
*** pgadiya has quit IRC08:58
kklimondagreat, thanks a lot - I'll pass it along08:59
tobiashkklimonda: and this I found for the commentlink condfig in gerrit: https://review.openstack.org/#/c/42495/209:00
*** eroux has quit IRC09:00
*** e0ne has joined #openstack-infra09:01
dalvarezhi guys, just a quick question... we saw there was a new tag in ubuntu kernel on Aug 31 and we're running that kernel right away in the gate. How does it work? We always run latest tagged kernel? Thanks!!09:02
*** xinliang has joined #openstack-infra09:06
*** xinliang has quit IRC09:06
*** xinliang has joined #openstack-infra09:06
*** ralonsoh has quit IRC09:06
ianwdalvarez: the trusty/xenial images are built daily, and will basically be running what was apt-get'ed that day09:07
*** ralonsoh has joined #openstack-infra09:07
*** eroux has joined #openstack-infra09:07
ianwbut, if you really mean aug 31, well we're not quite at the point we can time travel :)09:07
*** pgadiya has joined #openstack-infra09:12
openstackgerritDaniel Lublin proposed openstack-infra/git-review master: Actually output the warning  https://review.openstack.org/44347409:12
openstackgerritIsaku Yamahata proposed openstack-infra/project-config master: manasca: rename monitoring-log-api to monitoring-logging  https://review.openstack.org/49244409:12
openstackgerritDaniel Lublin proposed openstack-infra/git-review master: Allow choosing which field to use as author when naming branch  https://review.openstack.org/44457409:12
*** jcjo has joined #openstack-infra09:13
*** Guest54759 has quit IRC09:13
*** jcjo has quit IRC09:17
*** ramishra has quit IRC09:18
*** gongysh has quit IRC09:19
openstackgerritAlexander Chadin proposed openstack-infra/project-config master: Register watcher-tempest-plugin jobs  https://review.openstack.org/49040009:19
*** ramishra has joined #openstack-infra09:20
dalvarezianw, lol july 31 :)09:21
dalvarezianw, thanks for the info... and is there an "easy" way to revert to the previous image? there's been a regression in ubuntu kernel affecting neutron functional tests... not sure if other projects are affected too09:22
*** luzC has quit IRC09:22
ianwdalvarez: there is not an easy way to revert to something several days ago, we only keep a couple of days images09:23
*** luzC has joined #openstack-infra09:23
ianwpinning a kernel is going to be ... tricky09:23
yamahatahello, now project-config gate-project-config-jenkins-project is broken. The fix is https://review.openstack.org/#/c/492444/ or https://review.openstack.org/#/c/492379/09:25
yamahatacan you please review it?09:25
dalvarezianw, ack pinning would work until the regression's fixed but gotcha09:25
dalvarezthanks a lot :)09:25
*** jamesmcarthur has joined #openstack-infra09:28
yamahatathanks for quick review.09:29
openstackgerritFlávio Ramalho proposed openstack-infra/project-config master: zuul: layout: osa-os_sahara: Add nv centos job  https://review.openstack.org/49245009:30
*** xinliang has quit IRC09:31
*** jamesmcarthur has quit IRC09:32
*** sree has quit IRC09:34
*** sree has joined #openstack-infra09:34
*** ykarel|lunch is now known as ykarel09:34
*** sambetts|afk is now known as sambetts09:36
openstackgerritMerged openstack-infra/project-config master: Change service name  https://review.openstack.org/49237909:37
*** gongysh has joined #openstack-infra09:37
*** luzC has quit IRC09:39
*** sree has quit IRC09:39
*** sree has joined #openstack-infra09:40
*** markvoelker has quit IRC09:42
*** xinliang has joined #openstack-infra09:44
*** xinliang has quit IRC09:44
*** xinliang has joined #openstack-infra09:44
*** dizquierdo has joined #openstack-infra09:44
*** luzC has joined #openstack-infra09:47
*** udesale has joined #openstack-infra09:50
*** sdague has joined #openstack-infra09:50
*** jaosorior has joined #openstack-infra09:54
*** cuongnv has quit IRC09:56
yamamotoinfra team, can you add me to new groups, neutron-vpnaas-release and neutron-vpnaas-dashboard-release?09:56
ianwyamamoto: done10:00
*** Hal has joined #openstack-infra10:02
*** Hal is now known as Guest3494910:03
*** gcb has quit IRC10:04
*** lihi has joined #openstack-infra10:05
*** shardy has quit IRC10:08
yamamotoianw: thank you10:09
*** shardy has joined #openstack-infra10:11
*** cshastri has quit IRC10:14
Diabelkoweird, my zuul works fine, however when I do zuul-server -t /etc/zuul/layout.yaml it keeps saying FAILURE: Job X not defined10:14
Diabelkoshould I worry?10:14
*** katkapilatova has joined #openstack-infra10:17
*** yamamoto has quit IRC10:21
openstackgerritBogdan Dobrelya proposed openstack-infra/elastic-recheck master: Fix Generic job timeout bug match  https://review.openstack.org/49246310:22
*** liujiong has quit IRC10:22
*** katkapilatova has quit IRC10:22
*** yamamoto has joined #openstack-infra10:25
*** LindaWang1 has joined #openstack-infra10:27
*** yamamoto has quit IRC10:28
*** yamamoto has joined #openstack-infra10:28
*** thorst has joined #openstack-infra10:28
*** LindaWang has quit IRC10:30
*** LindaWang1 is now known as LindaWang10:30
*** thorst has quit IRC10:33
*** dtantsur|afk is now known as dtantsur10:35
*** sree has quit IRC10:35
*** sree has joined #openstack-infra10:35
*** katkapilatova has joined #openstack-infra10:35
*** udesale has quit IRC10:36
*** zhurong has quit IRC10:39
*** sree has quit IRC10:40
*** yamamoto has quit IRC10:41
*** ykarel_ has joined #openstack-infra10:41
*** ykarel__ has joined #openstack-infra10:42
*** ykarel_ has quit IRC10:43
*** ykarel has quit IRC10:43
*** iyamahat has joined #openstack-infra10:45
*** iyamahat has quit IRC10:45
*** iyamahat has joined #openstack-infra10:45
*** yamahata has quit IRC10:48
*** iyamahat has quit IRC10:50
*** rwsu has quit IRC10:53
sdaguefyi, citycloud timing out on pep8 install - http://logs.openstack.org/04/490304/1/gate/gate-nova-pep8-ubuntu-xenial/6a64b56/10:57
openstackgerritRui Chen proposed openstack-infra/shade master: Support to get resource by id  https://review.openstack.org/49208010:57
*** ykarel__ is now known as ykarel10:58
*** aeng has joined #openstack-infra10:58
*** jkilpatr has joined #openstack-infra10:59
*** gongysh has quit IRC11:01
*** psachin has quit IRC11:03
*** katkapilatova has quit IRC11:07
*** yamamoto has joined #openstack-infra11:09
ianwsdague: interesting ... that makes me suspect the mirror11:09
sdagueianw: yeh, that would make sense11:10
ianwlogging in it's stuck at  "debug1: pledge: network"11:10
ianwwhich stack overflow tells me might be related to systemd / dbus11:10
ianwremember the days when you could log in without message busses getting in the way!11:11
*** katkapilatova has joined #openstack-infra11:12
sdagueheh11:13
ianw"System information disabled due to load higher than 2.0"11:13
ianwif i could get a prompt ...11:13
sdaguehonestly, I noticed other citycloud long timeouts the other day11:13
ianwi am guessing this host is very unhappy11:13
sdagueI wonder if the move from "sometimes" to "always" is overloading those cloud hosts11:13
sdaguegiven that we've not had much idle cycles in nodepool this week11:14
ianwload average: 100.7411:14
sdagueianw: where are you getting that?11:15
ianwsdague: i'm logged into the mirror for the region of the job you just mentioned11:16
ianwthere are a lot of htclean processes i think11:16
sdagueah11:16
ianwhtcacheclean, all stuck on D wait channel11:16
sdagueI wonder if they are iops limitting it at the hypervisor level11:16
sdaguethat would definitely explain some of these job timeouts if everything is handing on the mirror11:17
ianwianw@mirror:~$ ps -aef | grep htcacheclean | wc -l11:17
ianw17211:17
ianwthat would be absolutely killing I/O11:18
ianwok ... and exim going mad.  that seems bad11:19
*** dhajare has joined #openstack-infra11:20
*** yamamoto_ has joined #openstack-infra11:20
*** ykarel_ has joined #openstack-infra11:21
openstackgerritPetr Benas proposed openstack/python-jenkins master: Allow specifying a non-standard port  https://review.openstack.org/49247811:23
ianwahh, i think exim was going bananas trying to send out messages from cron when i killed the htcacheclean11:23
*** yamamoto has quit IRC11:24
*** ykarel has quit IRC11:25
openstackgerritIan Wienand proposed openstack-infra/system-config master: Run htcacheclean under lock  https://review.openstack.org/49248111:34
*** sshnaidm|off has joined #openstack-infra11:34
ianwsdague / infra-root: ^ I think something like this should be considered without too much delay11:35
*** ldnunes has joined #openstack-infra11:35
sdagueianw: good call11:35
ianwalso, rebooting mirror.lon1.citycloud.openstack.org is probably not a bad idea.  it's coming back to life, but the logs are full of kernel oops's from stuck tasks.  i think probably best to start fresh11:36
sdagueianw: no objections here11:36
sdagueit seems to be failing jobs anyway11:36
sdagueianw: I'd say do it now before the US starts hammering the gate11:37
ianwsdague : hmm, ok, but if it doesn't come back i'm blaming you ;)11:38
*** sshnaidm|off has quit IRC11:39
*** sshnaidm|off has joined #openstack-infra11:39
*** dhajare has quit IRC11:39
*** dhajare has joined #openstack-infra11:39
*** coolno1 has joined #openstack-infra11:40
coolno1hello11:40
coolno1Need some inputs on whether openstack CI infra can be used by some third party projects11:41
coolno1Including gerrit, launchpad, gating ?11:41
*** ldnunes has quit IRC11:42
*** rhallisey has joined #openstack-infra11:42
*** dave-mccowan has joined #openstack-infra11:42
*** lucasagomes is now known as lucas-hungry11:42
*** alexchadin has quit IRC11:43
*** alexchadin has joined #openstack-infra11:44
*** alexchadin has quit IRC11:44
*** alexchadin has joined #openstack-infra11:45
*** aeng has quit IRC11:45
*** alexchadin has quit IRC11:45
ianwsigh ... i think something must be up with it's networking11:45
*** alexchadin has joined #openstack-infra11:45
ianwit's booting but really slow, and i'm guessing it's dns related11:45
*** alexchadin has quit IRC11:46
*** alexchadin has joined #openstack-infra11:46
*** alexchadin has quit IRC11:51
*** rhallisey has quit IRC11:52
*** katkapilatova has left #openstack-infra11:53
coolno1Hello Looking for some guidance can someone provide pointers11:53
*** katkapilatova has joined #openstack-infra11:53
sdaguecoolno1: the general statement has been that it's fine, however being specific about what you'd like to use it for would be helpful11:53
*** ldnunes has joined #openstack-infra11:54
sdagueianw: yeh, I kind of wonder if that is more indicative of a bigger issue on that cloud11:54
coolno1sdague, the gerrit review system, OpenStack gates, launchpad11:54
sdaguecoolno1: right... but for what kind of project11:54
sdaguelaunchpad isn't openstack specific11:54
coolno1sdague, it is a Cloud Foundry related project capable of deploying on OpenStack11:55
openstackgerritwes hayutin proposed openstack-infra/tripleo-ci master: WIP: containers periodic test  https://review.openstack.org/47574711:57
sdaguecoolno1: that would probably be fine11:57
coolno1sdague, wow11:57
sdagueianw: I wonder if it's worth disabling lon1 for now11:57
sdaguejust to not have jobs fail because of that mirror11:58
coolno1sdague, can you please provide me some pointers on what it takes to get this done11:58
coolno1sdague, some reference documentation11:58
sdaguehttps://docs.openstack.org/infra/manual/creators.html11:58
*** esberglu has joined #openstack-infra11:59
coolno1sdague, Thanks a lot for this. I will come back to irc as well as mailing list for more information11:59
ianwsdague: yeah, i think we have to.  i've tried disconnecting & reconnecting the interface and it still won't get past cloud init11:59
sdagueyeh11:59
sdagueit's only a 50 node drop, so not the end of the world11:59
ianwit gets an address from dhcp, but then no bueno11:59
*** ykarel__ has joined #openstack-infra11:59
sdaguethe aggregate throughput will go up if we aren't rando failing there12:00
*** ykarel_ has quit IRC12:00
*** ykarel__ is now known as ykarel12:00
sdagueianw: can you hot patch disable that?12:01
sdagueit will be 4+ hours for the commit to go through12:01
sdagueI'll propose the commit12:02
*** thorst has joined #openstack-infra12:02
coolno1sdague, Few questions. I hope the project can be hosted on github.com?12:03
coolno1sdague, Secondly I hope it can be non-python project. It is developed in nodejs12:03
*** esberglu has quit IRC12:04
sdaguecoolno1: we mirror some stuff to github, gut the hosting would be in gerrit if you use that12:04
*** alexchadin has joined #openstack-infra12:04
*** rlandy has joined #openstack-infra12:04
*** jtomasek has quit IRC12:05
*** tuanluong has quit IRC12:06
*** dhajare has quit IRC12:07
coolno1sdague, and the review system will be "https://review.openstack.org"12:07
openstackgerritSean Dague proposed openstack-infra/project-config master: Disable citycloud lon1  https://review.openstack.org/49249312:07
sdaguecoolno1: yeh, it's probably worth reading through all the project team docs, they are pretty extensive12:08
coolno1sdague, Sure I am just trying to understand if it will be misleading as it is a CF related project12:09
*** alexchadin has quit IRC12:09
*** sree has joined #openstack-infra12:09
sdaguecoolno1: that's fine, but I feel like it will be easier to ask about specific mismatches (if they exist) once you prime yourself with that documentation.12:10
*** dhajare has joined #openstack-infra12:10
coolno1sdague, yeah got it12:10
*** jpena has joined #openstack-infra12:10
ianw#status log nodepool in emergency file and citycloud-lon1 region commented out while we investigate issues with mirror12:13
openstackstatusianw: finished logging12:13
sdagueianw: thanks!12:13
*** sree has quit IRC12:13
*** ykarel_ has joined #openstack-infra12:15
*** trown|outtypewww is now known as trown12:16
*** Goneri has joined #openstack-infra12:17
*** ykarel has quit IRC12:18
*** rwsu has joined #openstack-infra12:19
*** sshnaidm|off is now known as sshnaidm12:19
*** xinliang has quit IRC12:20
*** sree has joined #openstack-infra12:20
openstackgerritJames Page proposed openstack-infra/project-config master: Add Gnocchi charm and associated interfaces  https://review.openstack.org/48994612:22
*** rwsu has quit IRC12:23
trownis there anyway to remove https://review.openstack.org/#/c/485689/10 from gate? it looks hung there. I did recheck, but I am worried it wont re-enter gate queue after rerunning check12:23
*** shardy has quit IRC12:23
*** dhajare has quit IRC12:24
*** shardy has joined #openstack-infra12:25
*** pgadiya has quit IRC12:25
ianw#status log mirror.lon1.citycloud.openstack.org migrated to a new compute node by Kim from citycloud.  appears up.  nodepool conf restored & nodepool.o.o taken out of emergency file12:27
openstackstatusianw: finished logging12:27
*** markmcd has quit IRC12:31
*** slaweq has quit IRC12:31
*** slaweq has joined #openstack-infra12:32
*** lucas-hungry is now known as lucasagomes12:32
*** xinliang has joined #openstack-infra12:33
*** pgadiya has joined #openstack-infra12:35
*** jamesdenton has quit IRC12:35
ianwsdague / infra-root : i believe things with lon1 are back to status quo.  i wrote http://lists.openstack.org/pipermail/openstack-infra/2017-August/005546.html12:36
*** dhajare has joined #openstack-infra12:36
*** slaweq has quit IRC12:36
*** jamesdenton has joined #openstack-infra12:36
ianweither fungi is awake or the robot he uses to do reviews is on autopilot.  either way, i feel ok going to bed :)  have a good day americans!12:38
*** sshnaidm is now known as sshnaidm|afk12:38
fungigo to bed ianw. still waking up so on silent running for now, but around (mostly)12:39
fungiand thanks!!!12:39
*** alexchadin has joined #openstack-infra12:44
*** mriedem has joined #openstack-infra12:45
*** EricGonczer_ has joined #openstack-infra12:47
*** sbezverk has joined #openstack-infra12:48
*** slaweq has joined #openstack-infra12:51
*** vhosakot has joined #openstack-infra12:52
*** markmcd has joined #openstack-infra12:56
*** jrist has joined #openstack-infra12:58
*** bh526r has joined #openstack-infra12:58
*** jpena is now known as jpena|mtg12:59
*** esberglu has joined #openstack-infra13:00
*** dhajare has quit IRC13:00
mhaydeni'm still seeing CI jobs come online with old versions of project-config, unfortunately13:00
mhaydenmordred was looking into it yesterday13:00
*** EricGonczer_ has quit IRC13:02
*** LindaWang has quit IRC13:02
*** Goneri has quit IRC13:04
openstackgerritMerged openstack-infra/project-config master: Remove opensuse-422 from jobs  https://review.openstack.org/49218113:13
*** baoli has joined #openstack-infra13:15
*** katkapilatova_ has joined #openstack-infra13:15
*** katkapilatova has quit IRC13:16
*** katkapilatova_ is now known as katkapilatova13:16
*** ramishra has quit IRC13:16
*** jcoufal has joined #openstack-infra13:17
*** katkapilatova1 has joined #openstack-infra13:17
*** katkapilatova1 has quit IRC13:18
*** ociuhandu has joined #openstack-infra13:18
*** ramishra has joined #openstack-infra13:18
*** rhallisey has joined #openstack-infra13:19
*** katkapilatova1 has joined #openstack-infra13:20
*** vhosakot has quit IRC13:21
*** vhosakot has joined #openstack-infra13:21
*** katkapilatova1 has quit IRC13:23
*** bobh has joined #openstack-infra13:24
*** LindaWang has joined #openstack-infra13:26
*** katkapilatova1 has joined #openstack-infra13:29
*** jaypipes has joined #openstack-infra13:30
*** Julien-zte has quit IRC13:34
andreafianw, fungi, EmilienM: I see gate-puppet-openstack-integration-4-scenarioNNN jobs failing rather consistently with timeout13:34
*** Julien-zte has joined #openstack-infra13:34
*** coolno1 has quit IRC13:34
andreafdid either the timeout or the test change somehow?13:35
andreafor am I hitting slow nodes perhaps13:35
andreafI see timeout after 1h (according to console.log) - that sounds too short13:37
openstackgerritMerged openstack-infra/project-config master: networking-odl: retire boron task  https://review.openstack.org/49195113:39
*** rwsu has joined #openstack-infra13:39
*** jamesmcarthur has joined #openstack-infra13:51
*** rwsu has quit IRC13:52
*** marst_ has joined #openstack-infra13:54
mnaserandreaf we've been seeing those timeouts happen often unfortunately13:55
*** alexchadin has quit IRC13:55
*** rwsu has joined #openstack-infra13:55
mnaserthe issue seems to be slow nodes, i said i'd put some time to try and catch the timeouts and get some logs but need some time to work on that13:55
*** camunoz has joined #openstack-infra13:56
mnaseri usually notice very long puppet runs on failed jobs (~40 minutes first run, 10 minutes second run) ... doesnt leave much time for tempest13:56
andreafmnaser: if you looks at failures from pupper jobs in https://review.openstack.org/#/c/492190/ you have a lot of timeouts13:56
*** vhosakot has quit IRC13:56
andreafmnaser: but 1h timeout sounds too short anyways - does the timeout include the time for the node to boot? I would not think so13:57
*** vhosakot has joined #openstack-infra13:57
mnaserandreaf i think the timeout starts the second the job starts13:57
andreafmnaser: ok yeah that's what I thought13:57
mnaserandreaf in that case the first puppet run seems to have taken ~30 minutes and the second ~10 minutes13:58
mnaserinstance took 8 minutes to setup too so lets round that up to 1013:58
mnaserthat means tempest has 10 minutes to run13:58
mnaserandreaf pabelanger noticed that some of the centos mirrors are not working properly so he was working on a solution for the caching https://review.openstack.org/#/c/492333/ -- the job runtimes in there seem promising13:59
*** gouthamr has joined #openstack-infra14:00
*** katkapilatova1 has quit IRC14:00
*** katkapilatova has left #openstack-infra14:00
*** davidsha has joined #openstack-infra14:01
*** gongysh has joined #openstack-infra14:02
*** gongysh has quit IRC14:03
openstackgerritAndrea Frittoli proposed openstack-infra/project-config master: Increase puppet integration job timeout to 90m  https://review.openstack.org/49254414:04
*** Goneri has joined #openstack-infra14:04
andreafmnaser, EmilienM: what about this ^^ until the problem is sorted?14:04
andreafpabelanger: ^^14:05
mnaserandreaf id let EmilienM or mwhahaha make the call on that, they know the CI much better, but i think its a good idea cause we're always timing out on successful installs14:05
mwhahahaI think it's related to the bandwidth problems14:06
mwhahahai guess we could increase it but i'm not sure that's going to really help14:06
mwhahahawe'll probably see it take 90mins and still timeout14:07
*** rwsu has quit IRC14:07
andreafmwhahaha: I see jobs usually timeout in the middle of a Tempest run, so I was hoping the extra time would help14:07
mwhahahayea we can go for it for now14:08
mwhahahai'm trying to get some tripleo-ci fixes merged so we stop reseting the gate queue14:08
mwhahahawhich is also not helping14:08
andreafmwhahaha: on Tempest patches I'm blind now wrt puppet jobs, so I don't to risk breaking something14:08
mwhahahasure14:09
*** rlandy has quit IRC14:09
*** spzala has joined #openstack-infra14:10
*** ramishra has quit IRC14:11
*** ramishra has joined #openstack-infra14:12
mnaserwould any infra-root be able to help out with this?  https://review.openstack.org/#/c/491800 merged last night but mirrors still have not appeared anywhere yet.  would anyone be able to investigate?  we've been blocked in magnum for quite a long time because of this - http://mirror.regionone.infracloud-vanilla.openstack.org/fedora/ shows nothing even though timestamp is updated14:13
*** jpena|mtg is now known as jpena|off14:14
*** rbrndt has joined #openstack-infra14:14
*** sree has quit IRC14:18
*** slaweq has quit IRC14:19
*** jkilpatr has quit IRC14:20
*** EricGonczer_ has joined #openstack-infra14:20
*** jkilpatr has joined #openstack-infra14:20
*** EricGonc_ has joined #openstack-infra14:24
*** EricGonczer_ has quit IRC14:25
jeblairmnaser: rsync: failed to set permissions on "/afs/.openstack.org/mirror/fedora/atomic/.": Permission denied (13)14:26
mnaserjeblair: did i miss or forget to do something in my patch?14:26
jeblairmnaser: not sure yet14:26
mnaserjeblair i pretty much tried to replicate what was being done for fedora in the same file, but at least we know it's failing14:27
*** lbragstad has quit IRC14:28
jeblairmnaser: btw, how big is the atomic mirror?14:29
mnaserjeblair: every qcow2 file is ~650M and i filtered it so that it only gets the qcow2 files, there are 10 images right now so rougly 6.5GB?14:30
fungioh, yep, i didn't think to check how much space that was going to add. sounded like only a few files, but... it's images so could actually be huge i guess14:30
jeblairshould be okay, but the partition is at 90% so it's probably time to make  new volume and move stuff around14:31
mnaserthere was a lot more content originally but i stripped it down with excludes14:31
mnaseri left a --dry-run output in the changeset which shows total size 6936506222 to be fully accurate (6.93gb)14:32
dimakHey, zuul shows there's a lot of queued jobs, some for 20 hours now. Is there an infra issue?14:33
mnaserdimak a few providers that donate infra have been disabled because of some issues so we're heavily capped by # of instances available to run tests14:36
mnaseryou can see the number of test nodes just capped in a perfect line :p14:37
*** spzala has quit IRC14:37
*** priteau has joined #openstack-infra14:39
jeblairchmod(".", 02755)                       = -1 EACCES (Permission denied)14:41
jeblairit's trying to setgid on the directory14:42
fungiahh, we likely need to not attempt to preserve permissions with rsync?14:42
jeblairor at least not that permission14:42
jeblairbut yeah, i would think the default perms would be fine14:43
jeblairmnaser: so maybe just drop "p" from -rlptDvz ?14:43
jeblairi'll try it real quick to make sure14:43
fungii'm wondering to what extent any perms from the source need to be kept14:44
mnaserjeblair okay cool, i can propose a fix if it's working14:44
jeblairfungi: yeah, probably none14:44
dimakmnaser, I see, thanks for the answer! :)14:44
jeblairmnaser: yeah it's happy now14:44
fungiagreed, i'm not coming up with any scenarios. maybe corner cases where some consuming application expects a regular file to have executable permission on the mirror? but even that seems like a pathological/broken behavior14:45
jeblairmnaser: it has actually completed the rsync; that was the last thing it was trying to do.  so once the fix lands, we shouldn't need to wait long for it to release the volume.14:45
jeblairdu says 6.5G14:46
openstackgerritMonty Taylor proposed openstack-infra/zuul feature/zuulv3: Add sphinx-autodoc-typehits sphinx extension  https://review.openstack.org/49255714:46
*** yamahata has joined #openstack-infra14:46
davidshaHi would this be the place to ask questions about tempest tests?14:48
jrichOdd behavior:  from a fresh checkout of project-config - I can init gerrit just fine (git review -s).   When trying the same on openstack-dev/sandbox - I get errors: "Could not connect to gerrit".  Had to manually add a remote.  This normal?14:48
openstackgerritMohammed Naser proposed openstack-infra/system-config master: Stop rsync from managing setgid permissions for Fedora Atomic mirror  https://review.openstack.org/49255814:48
mnaserjeblair fungi ^ thank you for your help/investigation14:48
jeblairmnaser: you're welcome!14:49
fungidavidsha: #openstack-qa is probably the channel you're looking for14:50
*** PsionTheory has joined #openstack-infra14:50
fungijrich: definitely not normal. you may need to make sure you don't have any local or config changes to the sandbox repo14:51
jeblairjrich: it looks like the .gitreview file in the sandbox repo is incorrect14:51
jeblairor... at least... unusual14:51
fungiwow, that may be one side effect of giving approval rights to that repo we didn't consider!14:51
openstackgerritsebastian marcet proposed openstack-infra/openstackid-resources master: External Calendar Sync  https://review.openstack.org/48768314:51
jeblairit has the ipv4 address rather than hostname14:51
jrichI noticed it had an IP address instead of the resolvable name.  I changed that first, but still had the issue.14:52
*** tmorin has joined #openstack-infra14:52
davidshafungi: Thanks!14:52
jrichcorrection: after addressing the IP address, I had a different error =)14:52
fungijrich: git review -s sets some remote lines in the local .git/config based on what it finds in the .gitreview file, so fixing that file doesn't necessarily correct the behavior afterward14:53
*** armax has joined #openstack-infra14:53
tmorininfraroot ?  would someone be around to help on a gate job that (apparently) keeps restarting  ( https://review.openstack.org/492142 )14:53
tmorin(sorry not a gate job, but the full set of gate jobs for a given change)14:54
jeblairfungi, jrich: https://review.openstack.org/49256014:54
*** sbezverk has quit IRC14:55
tmorinwhat I observe is that: all jobs restarted, at a point where most jobs for this change had passed, except two which were in progress14:55
*** quite has quit IRC14:55
jeblairfungi, jrich: fwiw, i did not have a git remote set up and used git-review to submit that successfully.  so based on what fungi said, it may be worth completely deleting your local sandbox repo and trying again.14:55
*** LindaWang has quit IRC14:55
tmorinthis has been occurring a few times in the past two hours14:56
jrichjeblair: Hah, I checked it out 10 mins ago -=)14:56
*** LindaWang has joined #openstack-infra14:56
fungitmorin: that's usually a sign that something in your change completely crashes or brings down network connectivity for the job node, so zuul thinks something unrelated happened to the node (ECLOUDS) and starts the job over for you14:56
jrichjeblair: but I'll do it anyway since I manually added the remote to fix my issue.  I want to try it fresh again to see if I can reproduce.14:56
jeblairjrich: yes, let us know if it still happens14:56
fungitmorin: have you streamed the console log to see what point in the job it stops? is it roughly the same point each time?14:57
tmorinfungi: nope, I'm currently behind a fw making this painful :-/14:57
*** makowals_ has joined #openstack-infra14:57
jeblairfungi, tmorin: if *all* jobs restarted, that's because zuul reset the gate queue14:57
fungiahh, yes, i'm switching computers so i can more easily see whether the change in question is in the gate pipeline or in check14:58
*** Swami has joined #openstack-infra14:58
fungii was (perhaps wrongly) assuming the latter14:58
tmorinjeblair: I'm not sure I know what 'reset the gate queue' means14:58
jeblairit's in gate with about 7 changes ahead, one failing14:58
*** makowals has quit IRC14:58
fungiyeah, so changes failing ahead of it causing it to be retested14:58
tmorinyes, the issue is in 'gate' queue14:59
tmorinfungi: ah, ok14:59
fungicompletely normal behavior, just didn't expect anyone to be surprised by that so my mind jumped to jobs getting requeued14:59
jeblairtmorin: zuul takes all of the changes that have been approved and tests each one with the ones ahead of it14:59
*** lbragstad has joined #openstack-infra14:59
jeblairtmorin: if one of the changes ahead fails, it pulls that change out of the line, reorders the list, and starts jobs again14:59
*** sree has joined #openstack-infra15:00
jeblairtmorin: if you look at the status page at http://status.openstack.org/zuul/ you'll see that your change is behind 492120 which failed15:00
*** jamesmcarthur has quit IRC15:00
jeblairtmorin: you'll also note that change is disconnected from the others in that queue.  so the most recent time all the jobs restarted on your change was when that one failed and zuul pulled it out15:00
*** felipemonteiro has joined #openstack-infra15:00
tmorinfungi, jeblair: ok, I had read about that a while ago, but somehow could accept this to be the reality today, given how likely it seems tobe the past few days that jobs fail15:01
jeblairtmorin: you can read more about this here: https://docs.openstack.org/infra/zuul/gating.html15:01
jrichBloody strange.  Exception: Could not connect to gerrit at ssh://jrich@review.openstack.org:29418/openstack-dev/sandbox.git15:01
jrichnc review.openstack.org 2941815:01
jrichSSH-2.0-GerritCodeReview_2.11.4-22-ge0c0f29 (SSHD-CORE-0.14.0)15:01
jeblairjrich: do you get any more information if you add the '-v' option?15:02
jrichjeblair: will try that next.  Good idea.15:02
fungijrich: git-review is attempting to do a test push, and the sandbox repo has contributor license agreement enforcement turned on, so it may be confused by the rejection gerrit is giving it if you haven't agreed to the icla15:02
tmorinjeblair, fungi: one general question on the overall CI load/slowness : can we expect to this this improve when some things get fixed / improved/ scaled, or rather that we should leave with that for a while as a result of OSIC decomissioning ?15:03
jrichfungi: that might be it.  However, that would imply the project-config repo is missing a setting.  (I was able to set that one up without a problem)15:04
fungitmorin: right now we're down not just osic, but also all of ovh, one region in citycloud and running infra-cloud at half-capacity due to network issues. we hope once some of those are addressed it will pick back up15:05
fungijrich: i don't believe we enforce a cla on project-config but i'm checking now15:05
jeblairtmorin: unfortunately, all of those are out of our control :(15:05
*** jaypipes has quit IRC15:05
*** LindaWang has quit IRC15:05
*** jaypipes has joined #openstack-infra15:05
jrichfungi: winner winner chicken dinner.  I had not accepted the icla on this account.15:05
tmorinjeblair, fungi: :-(15:05
tmorinfungi: thanks for the answer15:06
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Set basepython to python3  https://review.openstack.org/49159415:06
*** derekh has quit IRC15:06
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Setup ANSIBLE_ROLES_PATH in tox.ini  https://review.openstack.org/49159515:06
*** yamamoto_ has quit IRC15:08
*** dhajare has joined #openstack-infra15:08
openstackgerritKevin Carter (cloudnull) proposed openstack-infra/project-config master: Up the quota within RAX  https://review.openstack.org/49256615:09
* clarkb tries to catch up on all the fun15:09
cloudnullinfra-core ^15:09
clarkbpabelanger: fungi is glean's config drive label thing still a problem?15:09
cloudnullif you all can get that through I've spoken to pub cloud folks over here and they say we're good to up the limits for now15:10
clarkbcloudnull: I've +2'd thank you!15:10
openstackgerritSlawek Kaplonski proposed openstack-infra/project-config master: Enable missing "qos" extension driver for Neutron ML2 plugin  https://review.openstack.org/49256715:10
*** yamamoto has joined #openstack-infra15:10
jrichIf I may ask a process question:   I've been reading up on all the processes to create a 3rd party neutron plugin for my company.  I've been reading the creators page over and over and find myself a bit confused as some of it implies you might already have a repo in the openstack tree, and others imply you might use launchpad to host your code before it is accepted to the openstack tree.  At what point would I ask infra to create an empy repo in the15:10
jrich openstack tree?  (I've got launchpad all setup, working on CI related matters now)15:10
*** slaweq has joined #openstack-infra15:10
*** jamesmcarthur has joined #openstack-infra15:10
*** kjackal_ has quit IRC15:11
*** pgadiya has quit IRC15:11
*** yamamoto has quit IRC15:11
*** hareesh has quit IRC15:12
cloudnullclarkb: any way that could get in without having to go through the 5+ hours of gating ?15:12
clarkbcloudnull: we can bypass the gate it deemed necessary. I was mostly afk yesterday and catching up now so I will let others more in the know make that judgement call15:12
cloudnull++15:12
mnaserclarkb i proposed a fix yesterday https://review.openstack.org/#/c/492353/15:13
mnaseri didnt want to recheck as i wasnt sure the suse thing was a real issue or a timeout15:13
mnaserbut i guess i can always throw a recheck15:13
*** jamesmcarthur has quit IRC15:15
openstackgerritMerged openstack-infra/system-config master: Run htcacheclean under lock  https://review.openstack.org/49248115:16
clarkbmnaser: I've gone ahead and approved it too15:16
*** ramishra has quit IRC15:16
fungiclarkb: i think we still need a glean release if that label change has merged15:17
tmorininfraroot: I think I heard someone talk about a script that retrieves all logs for a test job run, is there something available somewhere doing that ?15:17
clarkbfungi: ya we will need one once the change merges (if it merges...)15:17
*** quite has joined #openstack-infra15:19
*** quite has quit IRC15:19
*** quite has joined #openstack-infra15:19
*** jamesmcarthur has joined #openstack-infra15:20
clarkbtmorin: the tox test runner knows to grab common tox command related logs (things like subunit and such for unittests) and devstack-gate knows how to get openstack related logs (openstack services, libvirt, mysql, and so on)15:21
*** annegentle has joined #openstack-infra15:21
pabelangermorning15:22
*** rlandy has joined #openstack-infra15:22
fungiclarkb: tmorin: the question may be more about how to retrieve a bundle of logs from the logs site for a completed test run?15:22
clarkbianw: thanks for digging into the mirror related items. Re the rackspace mirror being mostly odd on its own ya I think restarting may be a good place to start then we can apply mpm changes if problem persists15:22
pabelangerclarkb: fungi: do you mind looking at https://review.openstack.org/492336 help avoid a 302 redirect from yum client on every request to buildlogs15:22
tmorinclarkb: yes exactly, sorry if I wasn't clear15:22
clarkbto retrieve a bundle of logs from the logs site I just use wget15:23
*** spzala has joined #openstack-infra15:23
fungii used to do something for an ol ftp archive i ran, where requesting a directory name with .tar appended would recursively bundle it up and serve that to you, but no idea if that would be generally useful for logs.o.o or a possible nuisance for us15:24
fungiand agreed, wget can be set to recursively mirror content for you as an alternative anyway15:25
openstackgerritsebastian marcet proposed openstack-infra/openstackid-resources master: External Calendar Sync  https://review.openstack.org/48768315:25
openstackgerritMerged openstack-infra/elastic-recheck master: Add query for bug 1709744  https://review.openstack.org/49234215:25
openstackbug 1709744 in OpenStack-Gate "Gem fetch networking errors" [Undecided,New] https://launchpad.net/bugs/170974415:25
*** annegentle has quit IRC15:26
fungipabelanger: for some reason i thought i had already reviewed that one, but i guess not15:26
pabelangerfungi: I think you did review adding buildlogs.cdn.centos patch, this once is new a collapses it back to a single entry15:27
*** sdake_ is now known as sdake15:27
fungioh, so it does15:28
*** annegentle has joined #openstack-infra15:28
clarkbjeblair: fungi http://status.openstack.org/elastic-recheck/#1686542 shows there may be a recent drop off in job timeouts. Do we think that is likely related to reducing max-servers in infracloud?15:31
clarkboh I guess there was a change to pause image uploads there too15:31
*** thingee_ has joined #openstack-infra15:31
fungiclarkb: yeah, i think the combination of the two impacted that15:31
pabelangerclarkb: and fixes to buildlogs.cdn.centos has helped alot, we're now caching RPMs properly15:32
pabelangerruby gems / npm is the next ones to do I think15:32
pabelangerhttps://review.openstack.org/457720 for npm15:33
tmorinfungi, clarkb: yes, ok, recursive wget will do! thanks15:35
*** LindaWang has joined #openstack-infra15:35
openstackgerritSean Handley proposed openstack-infra/project-config master: Add Public Cloud WG project.  https://review.openstack.org/48954815:35
*** LindaWang has quit IRC15:36
openstackgerritMerged openstack-infra/system-config master: Replace buildlogs.centos with buildlogs.cdn.centos  https://review.openstack.org/49233615:36
*** e0ne has quit IRC15:39
seanhandleyAJaeger_: I rebased with master ^ Was that what your last comment was suggesting? I notice there's a "Conflicts with" on the right of the UI (never seen that before).15:39
openstackgerritPetr Benas proposed openstack/python-jenkins master: Allow specifying a non-standard port  https://review.openstack.org/49247815:42
*** jamesmcarthur has quit IRC15:42
*** jamesmcarthur has joined #openstack-infra15:43
openstackgerritsebastian marcet proposed openstack-infra/openstackid-resources master: External Calendar Sync  https://review.openstack.org/48768315:44
clarkbinfra-root with more udnerstadning of load issues do we want to direct enqueue the glean fix for config drive labels and the rax quota bump changes into the gate so they merge more quickly?15:44
jeblairclarkb: ++15:45
clarkbok I'll go ahead and do that15:46
*** jamesmcarthur has quit IRC15:46
pabelangerOh, we have more quota in RAX, nice15:46
*** jamesmcarthur has joined #openstack-infra15:46
fungii'm on board with that suggestion15:47
clarkbthe enqueue command is not returning as quickly as I would expect. I will attempt to practice patience :)15:48
fungicloudnull: somehow i missed your change in the scrollback until pabelanger just mentioned it. thanks!!!15:48
pabelangerclarkb: fungi: maybe we should consider brining citycloud-sto2 online and monitor mirrors.sto2 this time.  I think htcache could explain some networking issues we see in that region15:48
clarkbpabelanger: does that mirror have the problem that lon1 has/had?15:49
fungipabelanger: it wouldn't have explained our dns resolution failures there, but worth trying i guess15:49
pabelangerfungi: clarkb: and with buildlogs.cdn.centos changes, I think that will also help no more mirror issues we seen15:49
pabelangerfungi: Oh, right. DNS15:50
pabelangerI forgot about that15:50
*** jpena|off is now known as jpena15:50
fungimore just that it's been a few weeks since we brought it to their attention, they mentioned some things they were looking into... then nothing15:50
fungiso maybe they've fixed it already?15:50
clarkbok both changes are being enqueued, will hopefully merge in the near future15:50
pabelangermaybe? or maybe we should try live chat like ianw did15:51
clarkbmay be worthwhile15:51
mordredpabelanger: https://review.openstack.org/#/c/492567 <-- got a sec for a trivial +A?15:52
pabelangerfungi: do you mind also looking at 45772015:52
pabelangerdone15:52
mordredthanks15:53
*** Apoorva has joined #openstack-infra15:56
*** vhosakot has quit IRC15:58
*** camunoz has quit IRC15:59
*** jamesmcarthur has quit IRC15:59
*** camunoz has joined #openstack-infra15:59
*** jamesmcarthur has joined #openstack-infra15:59
*** aarefiev is now known as aarefiev_afk16:00
*** skelso has joined #openstack-infra16:01
fungipabelanger: i don't know about live chat for something complex like intermittent network connectivity for all instances in a region. plus they escalated it to one of the engineers already who was e-mailing us about it16:01
*** annegentle has quit IRC16:01
*** dklyle is now known as david-lyle16:02
openstackgerritsebastian marcet proposed openstack-infra/openstackid-resources master: External Calendar Sync  https://review.openstack.org/48768316:03
*** jamesmcarthur has quit IRC16:05
openstackgerritMerged openstack-infra/project-config master: Up the quota within RAX  https://review.openstack.org/49256616:06
openstackgerritWitold Bedyk proposed openstack-infra/project-config master: Add documentation jobs for monasca-api  https://review.openstack.org/49056916:07
*** jamesmcarthur has joined #openstack-infra16:07
openstackgerritMerged openstack-infra/system-config master: Add registry.npmjs.org reverse proxy cache  https://review.openstack.org/45772016:09
pabelangerclarkb: fungi: just seen a job fail to download from mirror.iad.rax.openstack.org, it had load of about 6 and also multiple htcacheclean process, I've killed them for now and load is back down to under 1.016:10
pabelangerwe likely should audit all of the mirrors16:10
pabelangergoing to do that now16:10
*** yamamoto has joined #openstack-infra16:12
clarkbok thanks16:12
dmsimardclarkb, pabelanger, fungi, mordred: let me know if there are any low hanging fruits I can help with ? reviews or stuff I can fix. I'm not root or core but would like to help if I can.16:13
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Bind secrets to their playbooks  https://review.openstack.org/49230716:14
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Expose final job attribute  https://review.openstack.org/47938216:14
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Remove 'auth' dict from jobs  https://review.openstack.org/49230916:14
pabelangerdmsimard: for now, I've just been keeping a close eye on jobs that are failing in gate (with help of elastic-recheck).16:14
fungithanks pabelanger, good idea16:14
pabelangermirror.iad same issue, load 6+, back under 1 once htcleancache killed16:14
pabelangerdfw* sorry16:15
fungidmsimard: yeah, helping us sort out what's slow or failing and why, trying to identify commonalities and bucket them together, is useful16:15
*** yamamoto has quit IRC16:17
dmsimardmhayden and I did notice some strange unexplained behavior about some jobs that seemed to be running with outdated images or project-config, I don't know if we have been able to get to the bottom of it but maybe there is a correllation16:17
*** eroux has quit IRC16:17
dmsimardmhayden: did someone figure that one out ?16:17
fungidmsimard: that would be great to dig into. it doesn't sound like a problem i'm aware of yet (other than the very brief mention in scrollback earlier)16:18
fungidmsimard: do you have any details?16:18
mhaydennot yet16:18
*** Swami has quit IRC16:19
mhaydeni supplied a few examples to mordred yesterday -- un momento16:19
*** mriedem is now known as mriedem_away16:20
dmsimardfungi: to make a long story short, we had jobs that were randomly not able to install a pyopenssl package, even long after the supposed fix had landed16:20
dmsimardI can surely come up with a logstash query, hang on.16:20
*** martinkopec has quit IRC16:21
mhaydenfungi: https://gist.github.com/major/4a6760f1f90303625061b40d16b7937416:21
fungidmsimard: mhayden: a link to the expected fix would also be helpful16:21
mhaydenfungi: well, the fix is already merged into project-config, and has been for 5-6days16:21
mhaydenhowever, some nodes come up with old versions of project-config16:22
mhaydenprior to the fix16:22
*** eroux has joined #openstack-infra16:22
fungimhayden: right, that's why i'd like to know which one it was16:22
mhaydenoh, i see what you mean16:22
mhayden:)16:22
funginodes don't themselves necessarily rely on project-config, which is part of my confusion16:22
*** ccamacho has quit IRC16:22
fungiso trying to start from the bottom and work my way up16:22
mhaydenfungi: https://github.com/openstack-infra/project-config/commit/c6cc5abe77ebcab2f55fffbc8ec1ee1c27c1307416:23
dmsimardmhayden, fungi: http://logstash.openstack.org/#dashboard/file/logstash.json?query=message%3A%5C%22Package%20pyOpenSSL-0.15.1-1.el7.noarch%20is%20obsoleted%20by%20python2-pyOpenSSL-16.2.0-3.el7.noarch%20which%20is%20already%20installed%5C%2216:23
dmsimardmessage:"Package pyOpenSSL-0.15.1-1.el7.noarch is obsoleted by python2-pyOpenSSL-16.2.0-3.el7.noarch which is already installed"16:23
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Rename allow-secrets to allow-untrusted-secrets  https://review.openstack.org/49261416:23
mhaydenfungi: well, that last one is missing, as is https://github.com/openstack-infra/project-config/commit/5fd623e1c30cf73faf01c355b1f769a68794aa4916:23
*** tmorin has quit IRC16:23
pabelangerfungi: clarkb: I think I got all the mirrors, each had multiple processes of htcacheclean running. RAX improved the most16:24
pabelangerbut did notice: htcacheclean -n -d120 -i -p/var/cache/apache2/mod_cache_disk -l300M16:24
pabelangernot sure where that is getting called from16:24
*** pcaruana has quit IRC16:25
*** psachin has joined #openstack-infra16:26
dmsimardmhayden, fungi: I'll actually go ahead and create a bug for that and create an elastic recheck query.16:28
dmsimardsince it's quite easy to track down in logstash16:29
mhaydendmsimard: teach me the ways when you're done ;)16:29
*** lucasagomes is now known as lucas-afk16:32
*** hongbin has joined #openstack-infra16:33
openstackgerritGabriele Cerami proposed openstack-infra/tripleo-ci master: WIP: containers periodic test  https://review.openstack.org/47574716:33
*** rcernin has quit IRC16:33
openstackgerritDavid Moreau Simard proposed openstack-infra/elastic-recheck master: Add query for PyOpenSSL installation failures  https://review.openstack.org/49261616:34
dmsimardmhayden: ^16:34
mhaydenthanks16:34
dmsimardmhayden: however ironically it might take a long time to land due to the state of the gate :(16:35
openstackgerritMerged openstack-infra/glean master: Add checks for uppercase config drive label  https://review.openstack.org/49235316:36
pabelangermnaser: clarkb: Yay^16:38
oansonHi. I am trying to add neutron-dynamic-routing tempest tests to Dragonflow's tempest gate. I've added the (^neutron_dynamic_routing...) regex, but I am not sure how to tell it to load neutron-dynamic-routing's tempest plugin. Could someone please assist?16:38
clarkbmhayden: how does redirecting stdout fix openssl?16:38
fungimhayden: dmsimard: okay, shifted gears for a sec so i can look into this... looks like the error from that query has appeared in logs going back at least a week according to logstash16:39
mhaydenclarkb: that was a secondary patch16:39
clarkboanson: for tempest plugin help #openstack-qa is probably better location to ask16:39
oansonclarkb, Sure. Will cross post. Thanks!16:39
dmsimardclarkb: yeah mhayden confused everyone with his devnull patch, you have to look for the one before that :p16:39
mhaydenclarkb: this was the main patch -> https://github.com/openstack-infra/project-config/commit/5fd623e1c30cf73faf01c355b1f769a68794aa4916:39
mhaydenthe CentOS image has *tons* of repos enabled16:40
mhaydenso we disable all except the basics in our gates, to more closely simulate a production environment16:40
dmsimardyeah, it installs centos-release-openstack-ocata which bundles about 4 repos I think16:40
dmsimardrdo, virt, virt-common, ceph16:40
dmsimard(probably shouldn't do that btw)16:40
mhaydeni was tempted to propose a patch to stop doing that, but i don't know how the RDO/triple-o folks feel about such things ;)16:40
*** ykarel_ has quit IRC16:41
mhaydeni would hope that their gate jobs would ensure that repo is present16:41
pabelangermhayden: dmsimard: yes, I want to remove that too. I wanted to discuss it with ianw first however16:41
dmsimardmhayden: maybe pabelanger or ianw would know why those are there. I know that OOO jobs get rid of all the repos too16:41
fungiso for starters, the suspected fix was to the {pipeline}-{name}-ansible-{scenario}-{ostype}{suffix} template16:41
fungimost recent hit i pulled from logstash was in a job called gate-openstack-ansible-os_aodh-ansible-func-centos-716:42
mhaydensounds reasonable16:42
pabelangerdmsimard: mhayden: we add rdo-ocata repo in images right now to install things like openvswitch. Need to see when that package gets installed16:42
mhaydenpabelanger: happy to join that conversation as i seem to be the RPM flag carrier in the land of openstack-ansible ;)16:42
fungiopenstack-ansible-os_aodh is a project i guess?16:42
mhaydenfungi: yes, it deploys aodh16:42
fungiyeah, just confirmed from the project column in logstash16:43
fungiokay, trying to make sure we were at least looking at a job built from this template16:43
mhaydenpabelanger: ah okay -- our expectation is that we would need to ensure the proper repos are present when our gate job scripts run (the ones from the OSA repos)16:43
fungiit's a shame the shell block there isn't set -x so we can see the commands16:44
pabelangermhayden: I am thinking we'd do the same thing we did for EPEL, only enable it for the package we need, but leave it disabled / removed16:44
mhaydenwe don't expect the nodepool image to contain those repos, if that makes sense -- we expect to configure them ourselves since our prod envs likely wouldn't have them installed16:44
mhaydenpabelanger: 100% agreed16:44
pabelangermhayden: but like I said, need to confirm with ianw16:44
odyssey4meI think ianw works in a EU timezone, so we'd likely only get feedback tomorrow.16:44
dmsimardfungi: FWIW you'll find some of those errors in ARA gate too. Which is one of the reasons I turned those jobs non-voting .. but I suspect they run less than the jobs in the OSA gate and thus occur less often.16:44
dmsimardodyssey4me: AU timezone :)16:45
fungioh, i guess it does inherit a -x16:45
fungizuul does that by default i guess16:45
odyssey4meah, I stand corrected - he's always on in my morning - so that makes sense :)16:45
fungiand yes looks like `sudo yum-config-manager --enable epel` gets run but not the others that patch should have added16:45
openstackgerritMonty Taylor proposed openstack-dev/pbr master: Put test-requirements into an extra named 'test'  https://review.openstack.org/49261916:46
openstackgerritMonty Taylor proposed openstack-dev/pbr master: Add support for a docs extra aligned with RTD locations  https://review.openstack.org/49262016:46
fungiso this probably suggests we're behind in updating configuration for one of the launchers16:46
openstackgerritGabriele Cerami proposed openstack-infra/tripleo-ci master: WIP: containers periodic test  https://review.openstack.org/47574716:46
clarkbiirc those repos are there for ovs16:46
pabelangerodyssey4me: dmsimard: right now, it looks like python-pip / python2-setuptools are getting installed from centos-openstack-ocata: http://nb04.openstack.org/dib.centos-7.log16:46
fungi"Launched by zl02" on this particular example16:46
clarkbif rhel would be so kind to realize people use ovs without openstack this would bea non issue :)16:46
dmsimardpabelanger: fair for python-pip, python2-setuptools is also in base OS so not required16:46
mhaydenfungi: your assessment sounds spot-on16:47
fungithe "build_master" column in logstash gets us that16:47
fungiand they _all_ look to be zl0216:47
dmsimardfungi: that seems like a winner16:47
fungiso i think this means zl02 isn't getting config updates for some reason16:47
odyssey4mepabelanger clarkb given that centos isn't used for devstack (or does it?) - why would OVS need to be there for jobs?16:47
pabelangerdmsimard: right, we likely need to build a centos-7 DIB with centos-openstack-ocata and see what breaks in the build.  Then find which projects (devstack) are installing packages for it, like OVS16:47
fungii'll check out that launcher16:47
dmsimardodyssey4me: they are trying to make it run devstack, though.16:48
odyssey4medmsimard ah, I see16:48
*** dhajare has quit IRC16:48
pabelangerodyssey4me: we use it for multinode jobs, but need to check I think they are still non-voting16:48
clarkbodyssey4me: it is used for devstack and multinode testing16:48
odyssey4mesurely whatever uses it should have the tolling to add the repo it needs instead of building the repo config into the image?16:49
clarkbpabelanger: devstack-gate is what will break bte16:49
odyssey4me*tooling16:49
fungithe /etc/project-config/jenkins/jobs/ansible-role-jobs.yaml file on zl02 looks current (has the fix) to project-config itself is getting updates there, suggesting the launcher daemon isn't loading them successfully. now to see why16:49
clarkbyes there are lots of options16:49
clarkbiirc we decided to do it this way so that the images would be roughly equivalent16:50
*** Swami has joined #openstack-infra16:50
clarkbrather than needing to special case every silly distro derp16:50
pabelangerclarkb: thanks16:51
*** bhavik1 has joined #openstack-infra16:51
odyssey4meclarkb yeah, it's hard to balance the cost/benefit for all needs16:52
jeblairlauncher-debug.log.14.gz:2017-07-27 06:54:22,262 DEBUG zuul.CommandSocket: Received reconfigure from socket16:52
jeblairfungi: last command received ^16:52
*** jamesmcarthur has quit IRC16:52
fungijeblair: thanks, i was grepping through them for reconfig but hadn't gotten that far back in time yet16:52
fungiprobably need to look at puppet logs instead in that case16:53
*** slaweq has quit IRC16:53
*** caphrim007 has joined #openstack-infra16:53
fungiAug 10 09:39:47 zl02 puppet-user[7913]: (/Stage[main]/Zuul::Launcher/Exec[zuul-launcher-reload]) Triggered 'refresh' from 1 events16:54
*** marst_ has quit IRC16:54
*** marst_ has joined #openstack-infra16:54
fungino errors there16:54
jeblairroot@zl02:~# zuul-launcher --help16:55
jeblairzuul-launcher: command not found16:55
fungiit's new-style reconfigure so not signal handler based... maybe it's stopped watching its fifo for that?16:55
*** jamesmcarthur has joined #openstack-infra16:55
fungioh, weird16:55
mordredfungi, clarkb, odyssey4me: I've been thinking recently that it would be 'nice' to be able to list alternate repos that are needed for bindep16:56
fungicannot access /usr/local/bin/zuul-launcher: No such file or directory16:56
pabelangereep16:56
dmsimardmordred: I think I created a story or a bug somewhere wit ianw to support a syntax like [platform:epel] or something16:56
* dmsimard searches16:56
fungipip list does't show zuul installed16:56
mordredran in to a similar thing the other day where there was a tool that need a repo added but there was no way to communicate that other than readme, which meant putting the tool depend into bindep was a non-starter16:56
*** Guest34949 has quit IRC16:56
pabelangerfungi: pip3?16:57
pabelangerfungi: possible it was installed as python3?16:57
fungipabelanger: doesn't look like it either, no16:57
*** yamahata has quit IRC16:57
pabelangerodd indeed16:57
mordreddmsimard: yah- something like that - although the specific case I had was like [platform:ubuntu:repo:https://example.com/ubutu:gpg-id:223452]16:57
jeblairfungi, pabelanger: syslog has aged out far enough for us to lose relevant info i think16:58
fungiso best guess, zuul somehow got uninstalled late last month on zl02, but we don't have syslog going back that far16:58
fungiright, that16:58
mordredwhich is obviously a terrible ui for that - but was the thing I wanted to be able to express and couldn't16:58
mordredfungi, jeblair: that's both very strange and also I'm sad there is no more log :(16:59
jeblairfungi, pabelanger: shall i "pip install /opt/zuul" ?16:59
fungijeblair: though next question, why isn't puppet installing it?16:59
jeblairdo we want a -U in there?16:59
odyssey4memordred we've always wanted to be able to point to a remote bindep file to reduce the requirement to synchronise things all over the place, but that's a whole different story16:59
jeblairfungi: i think we only install on git repo updates16:59
odyssey4mes/always/also/16:59
jeblairfungi: oh, the date of the last commit on that git repo is july 2717:00
mordredodyssey4me: nod. that would also be nice17:00
*** dtantsur is now known as dtantsur|afk17:00
jeblairfungi: it seems likely that pip *uninstalled* it for us then17:00
jeblairer puppet17:00
fungijeblair: okay, so that's more than coincidence at least17:00
fungijeblair: how about this: roll back the on-disk repo to HEAD^1 and see what puppet does next?17:00
odyssey4meheh, the puppetmaster has stolen its ghost17:00
jeblairpossibly it uninstalled it, then hit an error, then since it only fires on repo updates, never got around to fixing it.17:00
mordredoh - yah - the uninstall/reinstall dance maybe died in the middle17:01
mordredsince that's how upgrades work17:01
fungiworried that a manual run of pip won't turn up the same issue as puppet's attempt for heisenbug reasons17:01
*** baoli has quit IRC17:01
*** ralonsoh has quit IRC17:01
mnaserpabelanger awesome glad the glean patch merged17:01
fungibut yeah, could have been something as innocuous as a network error hitting pypi17:02
mordredmaybe we should also put in a puppet resource that will run the pip install if /usr/local/bin/zuul-launcher doesn't exist?17:02
jeblairfungi: i'd wager the bug was transient, but if you want to test it with HEAD^1, i think that would work and get us a little more data.  have at it (i'll stand down)17:02
openstackgerritJohn L. Villalovos proposed openstack/gertty master: Change usage of exit() to sys.exit()  https://review.openstack.org/49262217:02
fungimordred: seems like good belt-and-braces engineering to me17:02
mordred(assuming any of us know how to express such a thing in puppet)17:02
jeblairmordred: ++17:02
clarkbpabelanger: odyssey4me also the reason we use ovs instead of linux bridges has to do with vxlan support which may not be a problem anymore for linux bridge? We can't GRE in all clouds as it is its own ip protocol and you can't reliably enable it via neutron17:03
mnaseri'd kindly ask if it is possible to request the following to be promoted - https://review.openstack.org/#/c/492558/ -- we've been blocked in magnum for quite sometime due to our dependency on fedorapeople.org and it being really slow, almost all jobs are timing out (wasting resources)17:03
clarkbpabelanger: odyssey4me all that to say we could potentially go back to linux bridge with vxlan instead of gre assuming linux bridge vxlan support has grown enough to support it17:03
mnaserwith that patch, we will get those mirrored and we'll be able to churn less failed tests :>17:03
odyssey4meclarkb yep, LXB has had VXLAN since trusty17:03
clarkband just bypass ovs repo trouble entirely17:03
clarkbodyssey4me: ya but iirc it can't do braodcasts or something without multicast whereas ovs has hacks for that?17:04
fungijeblair: okay, because of merge commits i guess "Your branch is behind 'origin/master' by 2 commits" now that i've reset it to HEAD^117:04
clarkbodyssey4me: its easy enough to test by pushing a change to run with linux bridge and seeing if it works though17:04
*** baoli has joined #openstack-infra17:04
openstackgerritMerged openstack-infra/project-config master: Add periodic-stable jobs to oslo projecst that assert stable:follows-policy  https://review.openstack.org/49198017:04
fungii'm tailing syslog on zl02 now filtering for puppet17:04
odyssey4meclarkb since the inception of OSA in its previous life as os-ansible-deployment (icehouse) we were using vxlan... so I think those days are long, long gone17:04
openstackgerritMerged openstack-infra/project-config master: Add release notes jobs for python-swiftclient  https://review.openstack.org/49194017:04
clarkbodyssey4me: with multicast though or without?17:05
odyssey4methat I couldn't answer - perhaps mhayden or cloudnull are aware of such things?17:05
*** ociuhandu has quit IRC17:06
clarkbbut ya simple enough to switch it over in devstack gate and just see if anything breaks :)17:06
openstackgerritMerged openstack-infra/project-config master: Add periodic python jobs to kolla  https://review.openstack.org/49113317:06
*** bh526r has quit IRC17:06
clarkbif I find time today I may push that up17:06
pabelangerclarkb: I'm going to try: yum-config-manager --disable centos-openstack-ocata; yum -y install --enablerepo=centos-openstack-ocata openvswitch. That should also work17:07
pabelangerwe do the same thing with haveged and EPEL17:07
clarkbpabelanger: in devstack-gate?17:07
pabelangerclarkb: ya, see what breaks17:07
clarkbnote that we use devstack's install routines iirc17:07
pabelangerok17:08
odyssey4meclarkb please add me to the review, I plan on doing a bunch of work with OSA to use devstack-gate in Queens so I need to get more familiar with it17:08
*** electrofelix has quit IRC17:08
odyssey4mewe'd like to converge tooling for host prep where possible to stop reinventing wheels17:08
*** sree has quit IRC17:08
openstackgerritMerged openstack-infra/project-config master: Revert "Change job type for 3nodes job to move to oooq runner"  https://review.openstack.org/49147717:09
pabelangerwe're also about to start the push to ansiblify devstack-gate for zuulv3 too17:09
odyssey4meoh, good show17:09
odyssey4meI would be very happy to review anything along that line. In fact, let me add that repo to my review dashboard.17:10
fungiyeah, the eventual goal there is to decompose a lot of the reusable bits of devstack-gate as ansible roles so projects can mix-n-match them separate of the devstack-gate framework as a whole17:10
openstackgerritMonty Taylor proposed openstack-infra/puppet-zuul master: Ensure zuul gets reinstalled if it's missing  https://review.openstack.org/49262417:11
mordredclarkb, fungi, jeblair, pabelanger: ^^ how does that look?17:12
*** iyamahat has joined #openstack-infra17:12
odyssey4mecertainly works as a goal - which will hopefully bring the general projects using ansible together for test/infra bits17:13
*** marst_ has quit IRC17:13
jeblairmordred: i wonder if install_zuul can just have its own creates?  or does refreshonly prohibit that?17:13
clarkbrefresh only won't prohibit it but would undermine it as sicne those conditions will be ANDed iirc17:14
odyssey4meand once PR's are enabled as as a patch submission tool for openstack-infra, it might make it possible to share with non-openstack projects too17:14
*** annegentle has joined #openstack-infra17:14
*** Apoorva_ has joined #openstack-infra17:14
jeblairodyssey4me: can you elaborate on "once PR's are enabled as as a patch submission tool for openstack-infra" ?17:14
odyssey4mewell, it's my understanding that PR's via github may become a possible way to submit patches to openstack repositories17:15
*** iyamahat has quit IRC17:15
mordredjeblair: I think refreshonly breaks that17:15
*** iyamahat has joined #openstack-infra17:15
mordredor, yeah, what clarkb said17:15
*** iyamahat has quit IRC17:16
*** iyamahat has joined #openstack-infra17:16
*** slaweq has joined #openstack-infra17:16
jeblairodyssey4me: is there a spec about that?17:16
fungiodyssey4me: we'd need some tool to ingest those, and also to finish the work necessary to drop the icla in favor of the dco17:16
clarkbodyssey4me: ah ok looks like you can set bridge fdb rules to forward all l2 traffic to the null address to all interfaces. This way you don't have to run l2pop17:17
clarkbodyssey4me: that is probably workable for this setup since we are talking small numbers of hosts (also I think it is equivalent to how ovs hacks around this problem)17:17
*** Apoorva has quit IRC17:18
odyssey4mejeblair everything I just said was from hearsay - don't believe a word of it ;)17:18
odyssey4meI'm not sure how far along that is from idea to reality, or if it will ever be.17:18
jeblairodyssey4me: heh, ok.  yeah i think it's more at the idea stage.  could be reality eventually, but i don't think there's any current work planned.17:19
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: WIP: Add upload-pypi job  https://review.openstack.org/49192617:19
openstackgerritPaul Belanger proposed openstack-infra/openstack-zuul-jobs master: Create publish-openstack-python(-branch)-tarball jobs  https://review.openstack.org/49109317:19
odyssey4meI like any idea that broadens the community that can contribute and use what we work on.17:19
fungiodyssey4me: it's been discussed as a hypothetical once we can rely on the dco (update the pull request closer to instead leave comments if signed-off-by is missing, git-am any with a signed-off-by and close with a link to the gerrit review) but we're far from being able to implement that for logistical reasons for now17:19
*** bhavik1 has quit IRC17:20
fungiodyssey4me: the main catch right now is that we need contributors to official deliverables to have agreed to the individual contributor license agreement first17:20
odyssey4meNot that contributing to openstack is hard, or consuming what we build is hard... certainly with the CLA requirement the barrier is lower now.17:20
odyssey4meby that I mean http://lists.openstack.org/pipermail/openstack-dev/2017-August/120771.html17:21
Diabelkohow can I push job result to Gerrit so it will get picked up by test_result_table? can't find that part anywhere17:21
fungipuppet is starting to update zl02 now17:22
mordredodyssey4me: also, someone will need to write and maintain a thing that can do sync/forward between PRs and gerrit changes, which would increase the surface area of 'important things that are bad when they break' realted to github17:22
odyssey4memordred yes, there is that17:22
odyssey4meand with the shrinking resource base of people people both capable and able to do such a thing, I guess it's more likely to remain an idea17:22
fungimordred: i expect the first version would just close ingested pull requests with a link to the corresponding change in gerrit and some blurb about how to update it via further pushes directly to gerrit17:23
mordredodyssey4me: yah - but discussing doing that will be legally possible with the switch to dco17:23
mordredfungi: yup17:23
fungithe icla->dco switch is something i want to start focusing on soon, personally17:24
fungijust trying to get myself a little un-buried first17:24
fungihopefully not being a ptl will help ;)17:24
mnaseri would like to ask if it's possible to get https://review.openstack.org/#/c/492558/ bumped up the queue?  magnum is completely blocked because the image we download is failing and it's churning job timeouts for no reason17:24
*** rbrndt has quit IRC17:24
caphrim007mordred: are you aware of any update to pbr breaking the shade install?17:26
*** baoli has quit IRC17:26
fungiinfra-root: zm02 now has zuul (2.5.3.dev2) installed again, so whatever caused it to get left uninstalled back on the 27th was transient as suspected17:27
*** baoli has joined #openstack-infra17:27
*** baoli has quit IRC17:27
odyssey4memhayden ^17:27
odyssey4medmsimard &17:28
odyssey4methanks y'all for getting that figured out17:28
mhaydenfungi: thanks so much for digging into that17:28
fungithanks to everyone who reported it. this was an unusual situation, to be sure17:28
odyssey4meit'd still be interesting to understand how it happened - likely a failed install17:28
dmsimardsweet17:28
mordredcaphrim007: I am not aware of such a thing - are you seeing such an issue?17:28
dmsimardfungi: thanks for tracking that down, hopefully it will help the gate to some extent17:28
caphrim007mordred: i saw this today https://gist.github.com/caphrim007/21f12e899c212ae076e9360b8ac5828717:29
*** jamesmcarthur has quit IRC17:29
openstackgerritOmer Anson proposed openstack-infra/project-config master: Add nuetron-dynamic-routing to Dragonflow's tempest's local.conf  https://review.openstack.org/49262617:29
openstackgerritJohn L. Villalovos proposed openstack/gertty master: Change usage of exit() to sys.exit()  https://review.openstack.org/49262217:29
*** jamesmcarthur has joined #openstack-infra17:29
caphrim007mordred: pbr==3.0.0 didnt show a similar error17:29
fungimnaser: i went ahead and enqueued 492558,1 into the gate, so hopefully should land shortly17:30
mnaserfungi thank you so much!17:30
mordredcaphrim007: interesting. I'm ont sure why pbr version would affect that - but lemme look real quick17:30
*** jpena is now known as jpena|mtg17:30
odyssey4mevery happy to see the nodes in-use count climb after the quota increase merge17:30
*** markus_z has quit IRC17:30
fungimnaser: the concern we were using a lot of additional test resources on continually failing magnum changes seemed reasonable17:31
*** yamahata has joined #openstack-infra17:31
* fungi thinks things may have finally quieted down enough he can finish catching up on this morning's e-mail17:32
mnaserfungi: yeah, i figured just as much.  unfortunately most of the cores are on EU so i would have asked them to abandon/restore to stop the jobs till we land this + patch for mirrors17:32
mnaserbut cant have it all D:17:32
caphrim007mordred: erm...now i cant reproduce it. i guess nevermind?17:32
openstackgerritDavid Shrewsbury proposed openstack-infra/project-config master: Disable py2 dsvm on nodepool feature/zuulv3 branch  https://review.openstack.org/49262917:32
Shrewspabelanger: ^^^ i think this needs to be done for nodepool too?17:33
mordredcaphrim007: maybe you got lucky! definitely ping if you see it again, it looks like an issue in a transitive depend so it would be weird for a pbr version to be involved17:33
Shrewspabelanger: i'm not sure about the coverage job17:33
pabelangerShrews: left question17:35
openstackgerritSean McGinnis proposed openstack-infra/project-config master: Skip bandit and functional tests for doc changes  https://review.openstack.org/49263017:35
*** jamesmcarthur has quit IRC17:36
*** baoli has joined #openstack-infra17:37
Shrewspabelanger: where are the gate jobs defined?17:38
*** shardy has quit IRC17:38
*** krtaylor has quit IRC17:38
pabelangerShrews: same place, just add a gate: key17:39
Shrewspabelanger: what controls the gate jobs when not explicitly listed?17:40
Shrewsoh, probably in the templates17:41
pabelangerShrews: the templates above, they have gate jobs listed also17:41
openstackgerritMerged openstack-infra/system-config master: Stop rsync from managing setgid permissions for Fedora Atomic mirror  https://review.openstack.org/49255817:43
*** annegentle has quit IRC17:44
*** slaweq_ has joined #openstack-infra17:46
*** SumitNaiksatam has joined #openstack-infra17:48
*** mriedem_away is now known as mriedem17:48
*** slaweq has quit IRC17:50
*** ociuhandu has joined #openstack-infra17:50
*** sekelso has joined #openstack-infra17:50
*** gouthamr_ has joined #openstack-infra17:51
*** gouthamr has quit IRC17:51
*** xarses_ has joined #openstack-infra17:52
openstackgerritDavid Shrewsbury proposed openstack-infra/project-config master: Disable py2 dsvm on nodepool feature/zuulv3 branch  https://review.openstack.org/49262917:52
*** skelso has quit IRC17:53
*** spzala has quit IRC17:55
openstackgerritDavid Shrewsbury proposed openstack-infra/project-config master: Disable py2 dsvm on nodepool feature/zuulv3 branch  https://review.openstack.org/49262917:58
*** dizquierdo has quit IRC18:00
*** iyamahat has quit IRC18:00
*** iyamahat has joined #openstack-infra18:01
*** trown is now known as trown|lunch18:01
openstackgerritMatt Riedemann proposed openstack-infra/elastic-recheck master: Add query for same-host cellsv1 fail bug 1709946  https://review.openstack.org/49263618:05
openstackbug 1709946 in OpenStack Compute (nova) "ServersAdminTestJSON.test_create_server_with_scheduling_hint randomly fails SameHostFilter in cells v1 job" [Low,Confirmed] https://launchpad.net/bugs/170994618:05
*** Apoorva_ has quit IRC18:08
*** Apoorva has joined #openstack-infra18:09
*** davidsha has quit IRC18:10
*** apetrich has quit IRC18:10
*** annegentle has joined #openstack-infra18:11
*** apetrich has joined #openstack-infra18:13
*** Apoorva_ has joined #openstack-infra18:15
clarkbmordred: pabelanger fungi mnaser so do we want to make a glean release?18:16
fungiclarkb: i believe so18:16
fungihave you looked at the history since the last tag yet?18:17
clarkbI have not. Currently getting babysitter settled in then I can dig through and see what we'd be adding (my guess is not much)18:17
mnaserclarkb i guess we could, then once we get new images, i made some tweaks/changes to the flavors so ill probably have to submit a change which will be able to give 8 cores (shared) instead of 2 (dedicated).  looks like 100% cpu usage is not much of an issue18:17
fungiclarkb: no problem, just trying not to duplicate efforts. i'm looking through it now18:18
*** SumitNaiksatam has quit IRC18:18
clarkbkk18:18
mnaserfungi my little hack - https://github.com/openstack-infra/glean/compare/1.9.1...master18:18
mnaser:p18:18
*** Apoorva has quit IRC18:19
fungigiven that 1.9.1 was a revert of a feature from 1.9.0 which is being reintroduced, we probably need to tag 1.10.0 for this18:19
fungiTheJulia: ^ that's yours, do you concur with that version numbering choice?18:19
openstackgerritSam Yaple proposed openstack-infra/project-config master: Add additional LOCI repos  https://review.openstack.org/49263718:23
*** rbrndt has joined #openstack-infra18:27
*** ldnunes has quit IRC18:31
*** krtaylor has joined #openstack-infra18:32
*** rcernin has joined #openstack-infra18:35
clarkbfungi: that sounds right to me. Its just the systemd detection right?18:38
fungii suppose? i mean that's what ci and code review said18:38
clarkbya arguably that is a bugfix? so bug fix of bug fix could be 1.9.2?18:39
clarkbI don't think it mattes too much18:39
fungipabelanger and mordred seemed to think it was an appropriate reimplementation, along with prometheanfire, wznoinsk and sambetts18:40
clarkbya it looks fine I'm just trying to reason about whether or not it is a feature deserving the .1018:40
clarkbor just a .9.218:40
clarkbI think either way work since it straddles the line18:41
fungithe reason i say 1.10.0 is that 1.9.0 introduced "enable network.service with systemd" which 1.9.1 reverted and now we're reintroducing18:41
clarkbah18:41
fungibut it's a fuzzy argument, i agree18:42
clarkbfungi: I'm not seeing that in the diff mnaser linked (but I may just be blind)18:42
*** spzala has joined #openstack-infra18:42
fungihrm, yeah revising my analysis here18:43
*** Apoorva_ has quit IRC18:43
*** ldnunes has joined #openstack-infra18:43
*** spzala has quit IRC18:43
openstackgerritMerged openstack-infra/project-config master: Enable missing "qos" extension driver for Neutron ML2 plugin  https://review.openstack.org/49256718:43
fungioh! "Enable network.service with systemd" merged before we tagged 1.5.018:44
*** spzala has joined #openstack-infra18:44
*** Apoorva has joined #openstack-infra18:44
fungiwow that revert was a long time coming18:44
*** Sukhdev has joined #openstack-infra18:44
fungii would have thought the revert itself warranted a 1.10.0 in that case18:44
openstackgerritMerged openstack-infra/project-config master: Register watcher-tempest-plugin jobs  https://review.openstack.org/49040018:44
fungiand looking at diffs, "Revise systemd determination to verify systemctl presence" is not a reintroduction of that feature18:45
clarkbya I don't think the feature has been added in this delta18:45
fungiso now i'm thinking 1.9.1 probably _should_ have been 1.10.0, but water under the bridge now18:45
fungiso i'm cool with 1.9.2 as those changes do look like trivial fixes18:46
clarkbah ok now I grok18:46
AJaeger_seanhandley: I did not request you do anything. I left a comment with an explanation that wasn't clear, sorry. Recheck means: Run the tests again. We automatically rebase before running tests, so the new test run would have worked since it automatically rebases ...18:46
clarkbfungi: ya I think its purely bug fixes in this one18:46
fungiwfm. 1.9.2 it is18:46
fungii should have looked closer at those commit diffs18:46
*** eranrom has joined #openstack-infra18:47
kklimondais openstack-infra/system-config a good starting point to deploy a copy of OS CI for a different project?18:50
fungikklimonda: there is an openstack-infra/puppet-openstackci module which we consider to be the main entrypoint for most of that18:50
fungisystem-config drags in a ton of other stuff like our wiki, etherpad, ethercalc, listserv, codesearch...18:51
*** psachin has quit IRC18:52
fungikklimonda: what you likely want to take a look at is https://docs.openstack.org/infra/openstackci/ and then separately set up gerrit (assuming you mean a full ci system including your own gerrit code review server) separately with the puppet-gerrit module, maybe using the openstack_project::gerrit class from system-config as an example18:53
*** kjackal_ has joined #openstack-infra18:53
fungiyou might also consider other options for installing/managing gerrit, as our puppet-gerrit module isn't so great at hands-off bootstrapping a new gerrit deployment from scratch18:54
fungithe puppet-gerrit module depends on some separate manual steps anyway18:54
kklimondathat's what I'm worried about, i've seen it often enough that the puppet (or config management) drifts off and is unsuitable for deploying from scratch18:54
fungiwell, in the gerrit case it was never completely suitable for deploying from scratch18:55
fungigerrit's a fairly complex java app running in a jvm which needs interactive setup like initial account creation and pushing in at least minimal configuration to grant permission to your automation18:56
kklimondamhm18:56
kklimondawe can handle small things like that for sure18:57
fungimy recommendation would be to first experiment with installing gerrit by hand following their instructions, and then look at our system-config documentation about how we configure and operate ours: https://docs.openstack.org/infra/system-config/gerrit.html18:57
kklimondabtw, I was going through your system-config (and docs) and you have puppetmaster, but you use ansible for running puppet on nodes.18:57
fungithe puppet-gerrit module is not bad for maintaining a running and configured gerrit deployment, but there are some chicken-and-egg/catch-22 issues trying to automate a from-scratch gerrit deployment18:58
kklimondahave you repurposed puppetmaster node for some other things (like a central place to run ansible from) or is there something I'm missing18:58
kklimondawe've installed gerrit a couple of times18:58
kklimondaand we already have gerrit, our current CI is based on an old system-config fork I think, at least partially18:58
fungikklimonda: yes, our puppetmaster server is no longer a puppet master, it's just a place where we centrally host and manage our secrets, and where the cron job that calls ansible lives18:59
*** rcernin has quit IRC18:59
*** sekelso has quit IRC19:00
fungiansible in turn copies puppet manifests and secrets (in the form of hiera trees) onto individual servers in the inventory and then calls puppet apply locally on them19:00
fungikklimonda: so anyway, if all you want is a ci system then system-config contains a bunch of extra stuff you don't need and also a lot of settings that are very specific to the openstack community's needs which would likely need removing/adjusting to suit your community's19:01
fungiwe consider system-config be the entrypoint for our entire community infrastructure, not just our ci system19:02
*** adisky__ has quit IRC19:02
kklimondamhm, it also contains a lot of battle stories that we'd love to leverage :)19:02
fungino doubt19:02
kklimonda(for example going with ansible for running puppet probably had a reason etc.)19:02
*** portdirect is now known as eteppete19:02
fungii mean, that's probably the primary reason we make all of this available publicly under a free software license, after all19:02
*** sekelso has joined #openstack-infra19:02
*** jamesdenton has quit IRC19:03
fungiwe want other communities to be able to learn from our mistakes and not need to duplicate effort on problems we've already solved, after all19:03
*** openstackgerrit has quit IRC19:03
*** jamesdenton has joined #openstack-infra19:03
kklimondabut I see your point - I'll look into puppet-openstackci module and then start thinking how to leverage it along with system-config to get something of our own :)19:03
*** eteppete is now known as portdirect19:03
fungiso i guess my point was, since you initially mentioned wanting to duplicate our ci system, system-config is a much larger proposition19:04
fungibut if you really want to duplicate other bits of our community infrastructure besides just our ci system, it is indeed a good example19:04
kklimondanot for now, you are probably handling mailing lists, irc channels etc.19:05
fungiyep19:05
kklimondaso your suggestion for the entry point makes more sense19:05
fungiwe have puppet wrapper classes and modules for doing mailing lists and irc channels, for sure19:05
*** eranrom has quit IRC19:06
fungibut ultimately, we've organized it the way we have because we need a team of about half a dozen root sysadmins to be able to look after the infrastructure needs of a community of many thousands of developers, operators, users, et cetera19:06
fungiand be able to solicit help doing so from random interested members of our community and also from other communities with shared interests19:07
fungikklimonda: we also have some slide decks linked from https://docs.openstack.org/infra/publications/ which may provide some interesting insights (though they're in varying states of up-to-dateness so definitely don't assume all the technical specifics are current)19:08
*** openstackgerrit has joined #openstack-infra19:09
openstackgerritClark Boylan proposed openstack-infra/devstack-gate master: Hack in linux bridge vxlan support  https://review.openstack.org/49265419:09
clarkbpabelanger: odyssey4me ^ thats a quick first pass at using linux bridge. I expect that there will be bugs and something won't work19:09
*** sree has joined #openstack-infra19:09
clarkbwe can also refer back to the old gre code to see what might be wrong if thing sbreak19:09
*** trown|lunch is now known as trown19:10
clarkbpabelanger: odyssey4me feel free to push fixes if you like19:10
*** eranrom has joined #openstack-infra19:11
seanhandleyAJaeger_: Aha ok, that makes sense. Thanks!19:12
*** baoli has quit IRC19:13
*** sree has quit IRC19:13
*** portdirect has quit IRC19:17
*** portdirect has joined #openstack-infra19:17
*** eranrom has quit IRC19:21
*** nicolasbock has quit IRC19:21
openstackgerritMatthew Treinish proposed openstack-infra/subunit2sql master: Add graph for aggregate run time grouped by run metadata  https://review.openstack.org/49265519:24
mtreinishfungi: ^^^ if you were curious that's what I was using to generate the graphs before19:24
fungimtreinish: oh, neat!19:25
openstackgerritMatthew Treinish proposed openstack-infra/subunit2sql master: Add api func to get list of unique values for run_metadata key  https://review.openstack.org/49265619:25
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Bind secrets to their playbooks  https://review.openstack.org/49230719:26
*** tnovacik has joined #openstack-infra19:27
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Remove 'auth' dict from jobs  https://review.openstack.org/49230919:28
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Expose final job attribute  https://review.openstack.org/47938219:28
*** slaweq has joined #openstack-infra19:33
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Create publish-openstack-python(-branch)-tarball jobs  https://review.openstack.org/49109319:33
openstackgerritRamamani Yeleswarapu proposed openstack-infra/devstack-gate master: [TESTING][DO NOT MERGE] Testing TLS in Ironic jobs  https://review.openstack.org/49266119:35
clarkbfungi: will you be pushing the glean tag?19:35
*** xyang1 has joined #openstack-infra19:35
*** eranrom has joined #openstack-infra19:35
fungiclarkb: i can, just a sec lemme make sure there's nothing special like reno or release management going on with it19:36
*** slaweq_ has quit IRC19:36
*** sekelso has quit IRC19:36
TheJuliafungi:  I do19:36
fungiclarkb: looks like you pushed the last tag for it, so seems like our normal process. doing now19:36
clarkbya pretty sure it isn't using relmgmt tooling19:37
fungithanks for checking in, TheJulia! turns out i was mistaking your systemd systemctl detection fix for the other systemd-related thing we'd reverted in 1.9.119:37
*** iyamahat has quit IRC19:38
*** iyamahat has joined #openstack-infra19:38
fungiclarkb: easy enough to tell, tags get pushed/signed by our infra release key instead of individuals if release management automation is doing it19:38
*** sbezverk has joined #openstack-infra19:39
fungiclarkb: wow, you were very detailed with your tag description on 1.9.119:40
fungii think i'm just going to let the changelog speak for me with 1.9.219:40
*** baoli has joined #openstack-infra19:40
clarkb:)19:40
*** baoli has quit IRC19:41
fungiweird. i can't seem to sign tags on my workstation suddenly. gimme a check to troubleshoot19:41
fungi"error: gpg failed to sign the data"19:41
fungioh, i bet it's that DISPLAY isn't set and it's trying to use x11 for the askpass19:42
funginope, that's not it19:42
fungiaha19:44
openstackgerritMatt Riedemann proposed openstack-infra/elastic-recheck master: Add query for rebuild timeout cellsv1 bug 1709985  https://review.openstack.org/49266519:44
openstackbug 1709985 in OpenStack Compute (nova) "test_rebuild_server_in_error_state randomly times out waiting for rebuilding instance to be active" [Undecided,New] https://launchpad.net/bugs/170998519:45
clarkbmriedem: so whats the story on the live migration issues?19:45
mnaserif someone has an extra minute (somehow?!) on infra-root, could you check and see if the rsync script is working ok or not?  we merged the fix to get the mirrors working again but i dont see anything here - http://mirror.regionone.infracloud-vanilla.openstack.org/fedora19:45
*** baoli has joined #openstack-infra19:45
openstackgerritMerged openstack-infra/elastic-recheck master: Add query for same-host cellsv1 fail bug 1709946  https://review.openstack.org/49263619:45
openstackbug 1709946 in OpenStack Compute (nova) "ServersAdminTestJSON.test_create_server_with_scheduling_hint randomly fails SameHostFilter in cells v1 job" [Low,Confirmed] https://launchpad.net/bugs/170994619:45
*** sambetts is now known as sambetts|afk19:45
*** tnovacik has quit IRC19:45
*** mhickey has joined #openstack-infra19:46
clarkbmriedem: are the two libvirt talking to each other too slowly?19:46
mriedemclarkb: which one?19:46
*** jrist has quit IRC19:48
*** rlandy has quit IRC19:48
clarkbmriedem: the citycloud one19:49
clarkbwonder what run outlasted interval by 1.03 sec means19:49
clarkb*I wonder19:49
fungi#status log glean 1.9.2 released to properly support vfat configdrive labels19:50
openstackstatusfungi: finished logging19:50
*** e0ne has joined #openstack-infra19:50
mriedemclarkb: it's a periodic checkin with the service group api19:50
mriedemevery 10 seconds by default i think19:50
mriedemif the service doesn't check in within that time, it's considered down19:50
mriedemand you can't schedule to it19:50
mriedemwhich blows up any move operation test like live migration19:51
fungiturns out, if you're like me and you keep a short expiration on your signing key but periodically extend the expiration date on it, then other systems where you use that key to sign things need an occasional `gpg --refresh-keys` or they're refuse to keep using it thinking it expired19:51
clarkbmriedem: because you need >1 computes to do moves right?19:51
mriedemyes19:51
mriedemunless you resize to same host19:51
mriedembut this isn't that19:51
openstackgerritSam Yaple proposed openstack-infra/project-config master: Add additional LOCI repos  https://review.openstack.org/49263719:51
clarkbmriedem: and that happens over rabbit?19:51
clarkbthe check in19:52
mriedemyeah19:53
mriedemwell,19:53
mriedemi'd have to dig,19:53
clarkblooking at http://logs.openstack.org/12/491012/12/check/gate-tempest-dsvm-py35-ubuntu-xenial/2dfbf13/logs/screen-n-cpu.txt it doesn't just happen once either19:54
*** apetrich has quit IRC19:54
clarkbmriedem: if we can confirm the channel over which that happens we should be able to do a bit more profiling of that specifically within citycloud19:55
fungihttps://pypi.python.org/pypi/glean has 1.9.2 now, so next image updates should get it. do we want to trigger some now?19:55
*** apetrich has joined #openstack-infra19:55
clarkbfungi: probably a good idea so that if there are any problems we can delete new images and that won't happen overnight19:55
mriedemit's a thread group timer thing here https://github.com/openstack/nova/blob/master/nova/servicegroup/drivers/db.py#L5319:55
mriedembut i'd have to dig into it more - this is always confusing19:56
fungiclarkb: easiest way to do that? just nodepool dib-image-delete the oldest ones?19:56
mriedemhttps://github.com/openstack/nova/blob/master/nova/service.py#L18619:57
clarkbfungi: I think you want nodepool image-build19:57
clarkband then if there is a problem you dib-image-delete the newer ones19:57
*** e0ne has quit IRC19:57
mriedemjaypipes: sdague: ^ speaking of things to document at some point, the servicegroup api and how it monitors19:58
mriedemi always have to re-learn this19:58
sdaguemriedem: yeh19:58
*** tnovacik has joined #openstack-infra19:58
jaypipesmriedem: ack19:59
fungilooks like we have three ubuntu-xenial images ready according to nodepool dib-image-list, presumably due to adding raw?19:59
mriedemi think once the service starts, it reports into the thread group timer, which runs _report_state every 10 seconds19:59
mriedemhttps://github.com/openstack/nova/blob/master/nova/servicegroup/drivers/db.py#L9019:59
mriedemsets a counter and if the save() fails, or times out, we consider it down19:59
sdaguemriedem: you mean like this - https://github.com/openstack/nova/blob/2d2bf2a26bb49a3a8db9a3dddfef9097aea5739b/doc/source/admin/service-groups.rst#L219:59
clarkbmnaser: I still see -p in the rsync command for fedora. To make sure I'm looking at the right thing can you point me at the change that was supposed to fix that?19:59
mriedemsdague: heh yeah20:00
clarkbfungi: the third one may not be uploaded everywhere yes so we keep the last two20:00
clarkbfungi: once third is uploaded everywhere we should delete the oldtest one20:00
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Revert "Remove NPM mirror settings"  https://review.openstack.org/49266620:00
mnaserclarkb https://review.openstack.org/#/c/492558/20:00
pabelangerclarkb: fungi: ^would be nice to land our NPM mirror revert. Just tested in rackspace and npm install worked as expected20:01
clarkbmriedem: sdague ah ok so its doing a database table update20:01
fungihrm, we have opensuse-42.2, opensuse-422 and opensuse-423 images still. we merged changes to drop 422 and presumably 42.2 is cruft as well, should i just dib-image-delete those?20:01
clarkband we know the database can get pretty well loaded up here? also I thought computes couldn't talk to the db...20:02
pabelangerya, opensuse-422 can be removed20:02
pabelangerI haven't don't that in nodepool yet20:02
mriedemclarkb: it goes through conductor20:02
pabelangerfungi: want me to propose a patch for nodepool.yaml?20:02
fungi#status log Image builds manually queued for centos-7, debian-jessie, fedora-25, fedora-26, opensuse-423, ubuntu-trusty and ubuntu-xenial to use latest glean (1.9.2)20:03
openstackstatusfungi: finished logging20:03
fungipabelanger: i thought we already had approved one20:03
pabelangerfungi: no, just removal from JJB20:03
fungipabelanger: yep, i concur. nodepool.yaml on the server still have 42220:04
mriedemclarkb: so it's not the compute service that does this really, i mean it is, kind of20:04
fungithough 42.2 probably needs manual cleanup?20:04
mriedemit's this on a timer20:04
mriedemhttps://github.com/openstack/nova/blob/master/nova/servicegroup/drivers/db.py#L9020:04
mriedemwhich triggers this https://github.com/openstack/nova/blob/master/nova/db/sqlalchemy/api.py#L61420:04
mriedemwhere we update the last_seen_up value20:04
mriedemwhich is checked here https://github.com/openstack/nova/blob/master/nova/servicegroup/drivers/db.py#L6020:05
mriedemto tell if it's up or not20:05
mriedembased on some window20:05
clarkbmnaser: oh its just the one rsync of many we updated /me checks the file again20:05
fungiclarkb: https://review.openstack.org/492558 Stop rsync from managing setgid permissions for Fedora Atomic mirror20:05
mriedemi guess service_down_time is 60 seconds by default20:05
mriedemprobably to match the default rpc timeout20:06
fungioh, mnaser already linked it and i missed that20:06
clarkbmnaser: cron runs every two hours, change merged at 1800something UTC so its only just now rerunning at 2000UTC with new code20:06
mnaserahh that explains it20:06
fungicron to run rsync, specifically20:06
mnaserhttp://mirror.regionone.infracloud-vanilla.openstack.org/fedora/20:06
clarkbmnaser: so it is running now, you should see things after rsync runs and afs publishes20:06
mnaseri see it :D20:06
clarkbmriedem: and so that log message means it checked it at 61 seconds?20:07
mnaseroh this is awesome, this will help builds so much in magnum20:07
mnaserand take a big pressure off network20:07
fungimnaser: now if only heat will review your patch!20:07
*** baoli has quit IRC20:07
fungii'm surprised they were still going out to the internet for all that20:08
mriedemclarkb: yeah i think so20:08
mnaserfungi and given that its fedora 24 which was in there, must have been happening for quite sometime20:08
fungieek20:08
clarkbslowly whittling down the list of reliability fixes :)20:09
mriedemclarkb: the report interval is 10 seconds by default,20:09
clarkbmriedem: gotcha so it has ~6 chances to report in before being marked as bad20:09
mriedemwhich is what i think is the _report_state periodic20:09
mriedemyar20:09
clarkbthat seems reasonable and the wall time also seems more than sufficient20:10
clarkbI wonder if packet loss is part of hte problem there20:10
clarkb(since I cna't imagine it takes more than a minute to update the database)20:10
*** jkilpatr has quit IRC20:11
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Set opensuse-422 min-ready to -1  https://review.openstack.org/49266720:11
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Remove opensuse-422 from nodepool  https://review.openstack.org/49266820:11
pabelangerclarkb: fungi: removal of opensuse-422^20:11
*** tnovacik has quit IRC20:11
pabelangercannot remember if we need -1 on min-ready or if 0 is enough now20:11
clarkbmriedem: so nextstep is lets hold a test env and do some testing directly. What job should I be looking at holding?20:12
clarkb020:12
pabelangermnaser: cool, let me know if magnum has issue. Would be good to get another project on to AFS mirrors20:13
clarkbpabelanger: pretty sure its 0 across the board now then we just remove the images themselves to stop building and delete them (or pause)20:13
openstackgerritMerged openstack-infra/elastic-recheck master: Add query for rebuild timeout cellsv1 bug 1709985  https://review.openstack.org/49266520:13
openstackbug 1709985 in OpenStack Compute (nova) "test_rebuild_server_in_error_state randomly times out waiting for rebuilding instance to be active" [Low,Confirmed] https://launchpad.net/bugs/170998520:13
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Set opensuse-422 min-ready to 0  https://review.openstack.org/49266720:14
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Remove opensuse-422 from nodepool  https://review.openstack.org/49266820:14
pabelangerclarkb: updated20:14
*** mhickey has quit IRC20:14
*** funzo has quit IRC20:14
fungithanks pabelanger20:14
pabelangerI check with ianw tonight about doing the same with fedora-2520:15
pabelangerand we just started running jobs for post pipeline20:16
pabelangerextra 100 nodes does help :)20:16
mriedemclarkb: i saw it in gate-tempest-dsvm-multinode-live-migration-ubuntu-xenial when i reported the bug20:17
clarkbpabelanger: fungi did we need dirk to chime in on suse things or should I just go for it?20:17
clarkbmriedem: thanks will try to catch one of those on citycloud and hold it20:17
pabelangerclarkb: fungi: I confirmed with dirk before removing JJB jobs, so I think we are good to proceed20:18
clarkbpabelanger: ok first change is proceeding. Second one has +2's just needs approval when ready20:18
fungiyeah, no need to keep old images which no longer have any jobs20:18
pabelanger++20:19
*** Hunner has quit IRC20:19
*** jaypipes has quit IRC20:20
*** bmjen has quit IRC20:20
*** Sukhdev has quit IRC20:21
openstackgerritSam Yaple proposed openstack-infra/project-config master: Add additional LOCI repos  https://review.openstack.org/49263720:22
openstackgerritClark Boylan proposed openstack-infra/devstack-gate master: Hack in linux bridge vxlan support  https://review.openstack.org/49265420:24
*** eranrom has quit IRC20:24
dirkclarkb: pabelanger : I am okay with it.20:24
*** hamzy has quit IRC20:24
*** amotoki has quit IRC20:24
dirkI am just unsure whether the mirroring can be removed as well, I believe dib somehow depends on it20:24
fungii don't think our dib elements actually rely on our mirroring20:25
dirkThere was recently a switch to set a mirror during testing so that it doesn't pull from the internet20:25
fungioh, you mean changes to dib might have been getting tested on 422?20:25
clarkbmriedem: also that isn't a 100% failure in citycloud-lon1 is it?20:26
mriedemnot 100% failure no20:26
dirkfungi: https://review.openstack.org/47844320:26
mriedemwhen the live migration job fails, and it's with that type of warning, it's 90+% in that node provider though20:27
clarkbmriedem: ya I do see that specific region is far more common20:27
sdagueclarkb: also, it took 56 minutes to clean the node at the end20:27
mnaser2017-08-10 20:24:19.406877 | MAGNUM_GUEST_IMAGE_URL='\''http://mirror.mtl01.internap.openstack.org/fedora/atomic/stable/Fedora-Atomic-26-20170723.0/CloudImages/x86_64/images/Fedora-Atomic-26-20170723.0.x86_64.qcow2'\''20:27
mnaserthanks fungi / pabelanger and everyone else :D20:27
sdaguehttp://logs.openstack.org/54/487954/13/check/gate-tempest-dsvm-neutron-full-ubuntu-xenial/49d7ac4/console.html#_2017-08-10_19_08_56_20600720:27
sdaguesorry, 58 minutes20:27
sdaguewhich is indicative of really bad io20:28
clarkbhuh I wonder if those nodes are on the same host as the mirror was20:28
clarkb(maybe this is the downside to using online chat support, they will fix the one problem without doing the rest of it?)20:28
sdagueyeh, who knows20:28
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Bindmount /etc/lsb-release into bubblewrap  https://review.openstack.org/49020020:28
sdagueclarkb: yeh, any chance of getting a rep to irc lurk?20:29
sdagueor is that out of scope20:29
clarkbnot sure20:29
*** jcoufal has quit IRC20:29
clarkbpabelanger did send them email about possible bad hypervisor and sent instance uuids20:30
clarkbpabelanger: we haven't heard back on that irght (I don't see a response at least)20:30
fungiwe're also still waiting to hear back from them about the random network issues in sto220:31
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Add SSH private key for static.o.o  https://review.openstack.org/49267120:33
pabelangerjeblair: mordred: ^ our first secret for project-config20:33
pabelangerstart the painting20:33
jeblairpabelanger: that is much larger than i expected!  i'm guessing that's the pem-encoded version.20:34
jeblairpabelanger: can we encode the secret as binary and then pem-encode it on the other side?20:34
pabelangerjeblair: I believe so? that was the output using ./tools/encrypt_secret.py20:34
fungiyeah, no strict need to double-encode it20:34
jeblairpabelanger: sorry, i mean the *input* was pem encoded20:35
pabelangerOh, yes20:35
jeblairwhich is the normal thing for an ssh key20:35
openstackgerritClark Boylan proposed openstack-infra/devstack-gate master: Hack in linux bridge vxlan support  https://review.openstack.org/49265420:35
jeblairso we'll need to figure out how to translate that to binary, and then how to get ansible to write a pem-encoded version of the binary data (maybe a module)?20:35
pabelangerclarkb: I did send the email, I cc both you and fungi20:36
jeblairi expect that to be two pxkcs1 blocks if we do that20:36
fungithanks pabelanger!20:36
clarkbpabelanger: ya I see the one you sent just wondering if there was a reply that I may not have gotten (because they hit reply instead of reply all for example)20:36
clarkbguessing not20:36
fungii'm still catching up on e-mail from this morning, regrettably20:36
*** sdake is now known as k2so20:37
pabelangerjeblair: I need to step away for next 45mins, feel free to itterate on 492671 if needed20:37
openstackgerritMerged openstack-infra/project-config master: Set opensuse-422 min-ready to 0  https://review.openstack.org/49266720:37
clarkbfungi: and no word from ovh either I take it?20:37
jeblairpabelanger: i'm just going to leave some notes; i'm in the middle of other stuff20:37
fungiclarkb: not a few hours ago when i last checked, but will let you know shortly20:37
pabelangerjeblair: ack20:38
fungiit's not entirely clear to me whether tickets i open through the ovh dashboard will get me e-mail replies so i'm trying to watch for the dashboard updating, potential e-mail to infra-root@ and also jean-daniel replying to me directly from my earlier request20:39
*** jamesdenton has quit IRC20:39
*** jamesmcarthur has joined #openstack-infra20:40
jeblair(i also did send an email, and cc'd pierre who was on earlier threads, but have not gotten a personal reply)20:41
*** jamesdenton has joined #openstack-infra20:42
openstackgerritClark Boylan proposed openstack-infra/devstack-gate master: Hack in linux bridge vxlan support  https://review.openstack.org/49265420:43
fungithanks jeblair!20:44
*** sekelso has joined #openstack-infra20:45
*** annegentle has quit IRC20:45
*** esberglu has quit IRC20:47
*** skelso has joined #openstack-infra20:48
mordredmriedem: anything change recently (last couple of days) wrt console logs from nova in devstack/devstack-gate?20:48
mordredmriedem: we just failed a functional test herE: http://logs.openstack.org/48/491248/1/check/gate-shade-functional/d3fe2ef/console.html#_2017-08-10_20_28_34_958225 on an unrelated thing and figured I'd check to see if there's anything you know of off the top of your head before I dig further20:49
fungiheads up, http://blog.recurity-labs.com/2017-08-10/scm-vulns (git vulnerable to shell command injection via malicious ssh:// urls)20:50
*** gouthamr_ has quit IRC20:50
*** sekelso has quit IRC20:50
fungii don't think we need to worry about anything in our infrastructure being impacted (and we'll have updated git installed shortly anyway if it isn't already) but be mindful of your local dev environments20:51
*** jamesmcarthur has quit IRC20:53
*** jamesmcarthur has joined #openstack-infra20:53
ianwpabelanger: https://review.openstack.org/#/c/490331/ is the last outstanding issue for devstack + fedora26 ... any non-devstack f25 jobs happy to switch20:55
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Bindmount /etc/lsb-release into bubblewrap  https://review.openstack.org/49020020:55
mriedemmordred: what's the actual failure?20:56
mriedemi don't see any problems in the n-api or n-cpu logs for req-d90afce6-9ef9-4059-830c-2338ad18448320:56
clarkbpabelanger: odyssey4me latest patchset for using linuxbrdige instead of ovs actually looks pretty good. I can ping all the nodes from all the nodes on the dvr ha job setup20:57
clarkbpabelanger: odyssey4me still to be seen if there are any problems with neutron etc running on top of it but that could be a good option20:57
clarkbwe'll also want to test on centos since its kernel is older iirc20:57
*** vhosakot has joined #openstack-infra20:57
mordredmriedem: I think it's a poorly written test and is subject to there just not being any actual console log content20:58
openstackgerritMonty Taylor proposed openstack-infra/shade master: Make get_server_console tests more resilient  https://review.openstack.org/49268320:58
openstackgerritMonty Taylor proposed openstack-infra/shade master: Remove keystoneclient and ironicclient as direct depends  https://review.openstack.org/49268420:58
*** jamesmcarthur has quit IRC20:58
mordredmriedem: thanks for looking - I just pushed up  arework of the test that should test that shade calls nova correctly and gets a response and doesn't try to test that the guest in nova has produced console log output, which we have no real control over20:58
mriedemmordred: http://logs.openstack.org/48/491248/1/check/gate-shade-functional/d3fe2ef/logs/screen-n-cpu.txt.gz#_Aug_10_20_28_24_170095 says there is no console lot21:00
mriedem*log21:00
*** jkilpatr has joined #openstack-infra21:01
*** esberglu has joined #openstack-infra21:01
mordredmriedem: cool - so in this case there is just legitmately no console log - and nova returned {'console': ''} appropriately21:01
*** iyamahat has quit IRC21:01
*** trown is now known as trown|outtypewww21:02
*** iyamahat has joined #openstack-infra21:02
*** esberglu_ has joined #openstack-infra21:02
*** jpena|mtg is now known as jpena|off21:02
*** esberglu has quit IRC21:02
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Borrow some internap quota for Zuulv3  https://review.openstack.org/49268521:03
*** rhallisey has quit IRC21:04
ianwpabelanger / dmsimard : on the repos ... i'm more than willing to take advice.  primarily we've been focused on the devstack case but obviously all input is welcome.21:04
pabelangerclarkb: fungi: 492668 can land, no more opensuse-422 nodes running21:05
clarkbfungi: did the leaked nodes in vanilla get cleaned up?21:05
clarkbpabelanger: and the job against glean etc for opensuse 422 is removed?21:05
ianwissues we've had include openvswitch coming from rdo, a couple of python packages, but i'd have to check the logs, and making sure we're using the rhev forward-port for kvm21:05
fungiclarkb: there were no leaked nodes in vanilla as it turns out21:05
clarkbianw: https://review.openstack.org/#/c/492654/ is an effort to stop needing ovs in devstack-gate21:06
fungiclarkb: eventually the differential reached two entries, both of which nova knew about (one is the mirror and the other is a test instance pabelanger appears to have left there)21:06
clarkbianw: if ^ ends up working we could remove that repo at least from d-eg21:06
clarkb*d-g21:06
pabelangerclarkb: ya, we should have no jobs using opensuse-42221:06
clarkbfungi: gotcha21:06
pabelangerfungi: oh, my instance can likely be deleted if you want21:06
clarkbpabelanger: ok I say approve at will then you should have the +2's you need21:06
dmsimardianw: oh btw re: making f26 voting. There was two issues. The first was bindep to get the right python-devel package and the second (still haven't dug into this one) is that it doesn't seem like f26 has a "python3.5" interpreter21:07
fungiclarkb: the only other anomaly was the undeletable instance that nova thought was there but virsh on the compute node said didn't actually exist. not sure how to clean that up though21:07
pabelangerokay, now I have to step away for 45mins21:07
clarkbfungi: fun, I think we hav eone of those in chocolate too21:07
clarkband ya no good ideas on how to clean that up other than manual database munging21:07
fungii like to call it, "openstack mitaka"21:08
fungithose instances will go away when we redeploy21:08
clarkb+21:08
clarkb+21:08
*** krtaylor has quit IRC21:08
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Zuulv3: add the gate pipeline  https://review.openstack.org/49268721:09
*** baoli has joined #openstack-infra21:10
mordredjeblair: \o/ ... btw, on the etherpad, when you do the check/gate swap - perhaps we should include zuul-jobs/openstack-zuul-jobs/zuul-sphinx in that too?21:10
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Add Zuul to gate pipeline  https://review.openstack.org/49268921:10
mnaserfungi clarkb wouldnt a nova delete just do a noop delete on those undeletable instances?21:11
*** aeng has joined #openstack-infra21:12
jeblairmordred: seems reasonable21:12
*** Goneri has quit IRC21:13
clarkbmnaser: I think nodepool is trying to nova delete them in a loop every few minutes21:13
clarkbmnaser: so pretty sure that it isn't working21:13
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Add zuul-jobs to gate pipeline  https://review.openstack.org/49269121:14
*** mriedem has left #openstack-infra21:14
openstackgerritMonty Taylor proposed openstack-infra/shade master: Make get_server_console tests more resilient  https://review.openstack.org/49268321:14
*** mriedem has joined #openstack-infra21:14
fungimnaser: clarkb: also i manually tried to nova delete and the node remained in "active" state21:14
funginot even error21:15
fungieven as an admin user21:15
*** funzo has joined #openstack-infra21:15
mnaseris it possible that the hypervisor of that vm is no longer alive?21:15
fungii did not look at the nova service logs however21:15
*** ldnunes has quit IRC21:15
fungimnaser: the compute node nova show (as admin) claimed it belonged to is the one where i ran teh virsh show --all or whatever21:16
openstackgerritJames E. Blair proposed openstack-infra/openstack-zuul-jobs master: Add openstack-zuul-jobs to gate pipeline  https://review.openstack.org/49269221:16
mnasercause that could happen if n-cpu is crashed and not responding to rabbitmq21:16
mnaserso stuff like that gets queue'd up and never executed by it21:16
fungioh, maybe, except that compute node was booting and deleting other instances21:16
fungiso seemed to be working otherwise21:17
openstackgerritMerged openstack-infra/project-config master: Remove opensuse-422 from nodepool  https://review.openstack.org/49266821:17
mnaserwaits21:17
jeblairmordred: maybe not zuul-sphinx; it's not in zuulv3 at all yet21:17
mnaserso it doesnt even go in deleting state?21:17
mordredjeblair: ah. well, yeah - we certainly shouldn't enable it21:18
mriedemclarkb: fyi, it's not just live migration jobs21:19
mriedemgate-tempest-dsvm-neutron-full-ubuntu-xenial: https://bugs.launchpad.net/bugs/170950621:19
openstackLaunchpad bug 1709506 in OpenStack-Gate "Random live migration failures due to ComputeServiceUnavailable in citycloud-lon1 nodes" [Undecided,Confirmed]21:19
mriedeme-r just commented on that21:19
*** funzo has quit IRC21:20
openstackgerritMerged openstack-infra/tripleo-ci master: Replace references to deprecated controllerExtraConfig  https://review.openstack.org/48039521:20
clarkbmriedem: ok, maybe you want to update the bug title as that job won't run live migrations? something like "Nova compute heartbeats are slow and nova marks computes as offline"?21:21
openstackgerritRyan proposed openstack-infra/bindep master: Add ability to list all deps  https://review.openstack.org/49269321:21
clarkbmriedem: also that implies it isn't a networking problem because single node tests won't heartbeat over network21:21
clarkbmriedem: which puts weight behind the poor disk io on a hypervisor theory21:21
mriedemdone21:22
*** rybridges has joined #openstack-infra21:23
*** hrubi has quit IRC21:23
rybridgesHello! I have a review up here for bindep. Please take a  look at your convenience -> https://review.openstack.org/#/c/492693/21:23
rybridgesThanks!21:23
*** gouthamr has joined #openstack-infra21:24
clarkbmriedem: thinking about that we could add an fio run to pull general performance data on io maybe put that in devstacks world dump? though that only happens if devstack fails21:24
*** hrubi has joined #openstack-infra21:24
*** thorst has quit IRC21:24
sdagueclarkb: is there enough date in dstat?21:25
sdagueif io is bad, there should be a lot of wait time, right?21:25
clarkbsdague: ya, though I half expect we'd see lots of wait time in general. But we can look at the data we have and see21:25
sdaguehttp://logs.openstack.org/54/487954/13/check/gate-tempest-dsvm-neutron-full-ubuntu-xenial/49d7ac4/logs/screen-dstat.txt.gz21:26
sdagueI'm seeing a 19.2 load at one poin21:26
*** thorst has joined #openstack-infra21:27
clarkbrybridges: any reason you couldn't just bindep | sed or awk to do that?21:27
sdagueyeh, it's regularly going 50 - 60% wait21:27
*** Sukhdev has joined #openstack-infra21:27
fungiclarkb: i think the need is to have bindep output even not-missing dependencies21:27
fungie.g. list all dependencies whether they're installed or not21:28
sdague94% wait at one point21:28
clarkbfungi: oh I thoguht it did that if you left off the -b21:28
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Zuulv3: Remove check/gate jobs from zuul and friends  https://review.openstack.org/49269721:28
sdaguereight before the load 19.221:28
rybridgesPerhaps my knowledge of bindep is what is holding me back. From what I could tell, there is no way to get bindep to print all of the required packages regardless of whether or not they are already installed on the system21:28
sdagueso io is definitely not good21:28
rybridgesIf there was some way to pull out the list of all required packages, then sure i could sed/awk it into the format that I want21:28
sdaguethose are far outside our norms21:29
rybridgesbut whenever i run bindep, it only shows missing packages, which is not useful for me when I am trying to build rpms with concrete dependency lists21:29
sdagueload average of 10 was kind of our danger mark, and why we stopped doing scenario tests in parallel21:29
*** annegentle has joined #openstack-infra21:29
sdaguealso why its cpucount / 221:29
rybridgesIf you leave off the -b it just prints the missing packages in a slightly different format than it does with the -b flag on21:30
fungiclarkb: rybridges: right, even without -b you still only get the list of missing packages currently (just in a more verbose form), so it seems like a reasonable feature addition21:30
clarkbsdague: ya it tended to stay under 8 when I was doing testing to see if 3/4 was sane21:30
clarkbsdague: so 19 is quite high21:30
clarkbfungi: gotcha21:30
clarkbfungi: rybridges in that case I would list them like the other outputs list them (line by line iirc) and not assume formating21:30
clarkbthen it is easy to add in formatting if necessary using sed or whatever21:31
*** bmjen has joined #openstack-infra21:31
*** thorst has quit IRC21:31
rybridgesclarkb: I am printing it as a csv which is a very standard format that people can parse themselves however they like just as easily. and it plays nice with my particular use case which is putting the list of dependencies directly into an RPM .spec file. Furthermore, we are all python people here. We have a native csv library for python which makes dealing with the output very easy and flexible21:33
*** Hunner has joined #openstack-infra21:33
*** Hunner has quit IRC21:33
*** Hunner has joined #openstack-infra21:33
mordredclarkb, fungi, rybridges: maybe two features - a "list all" flag, and a "output in csv format" flag21:33
fungiyeah, one-line-per-package would make sense if building dependency lists for an rpm specfile or debian/control file is the primary use case21:33
mordredthat way we have the new feature "I want all packages" in a form consumable byanyone who is using the current output - and we can also give the comma-separated option to people who are using the '-b' option too, should they desire it21:34
*** Hypercube32 has joined #openstack-infra21:34
mnaserrybridges is it.. normal to download an image at 700Kb/s from a region mirror?21:36
clarkbya  Ithink providing both would be fine. I just worry that we'd have multiple output formats that differ between what output you want21:36
clarkbwhere the format should be separate from what output you want21:36
mnasersorry about that extra highlight rybridges -- getting tired, no idea why you were in my chatbox :x21:36
clarkbmnaser: if caches are stale yes21:36
mnaserahh that might explain it then21:36
rybridgesheh no worries mnaser21:36
mnaserthis was the first job to use it21:37
clarkbmnaser: afs in particular invalidates its entire cache whenever afs volumes are published21:37
jeblairfungi: would you mind okaying 429685?21:37
*** iyamahat has quit IRC21:37
clarkband if there is large geogrpahic distance between the mirror and the afs volume itslef that is painful21:37
*** spzala has quit IRC21:37
clarkbmnaser: as for reverse apache proxy I'd imagine thats what the download speed is for whatever backend item is being pulled and isn't cached21:37
*** spzala has joined #openstack-infra21:38
mnaser17 minutes to download the image, job took 1h10m .. so the k8s job should be around 50 minutes which is a huge improvement over the 1h40 or so it took before21:38
*** baoli has quit IRC21:38
mnaserexciting21:38
fungijeblair: looking21:38
*** spzala has quit IRC21:38
*** baoli has joined #openstack-infra21:39
jeblairclarkb, mnaser: as long as the image isn't changed, it should be faster after the cache warms up, even with subsequent volume releases (in that case, it only needs to do a roundtrip to stat the file)21:39
*** spzala has joined #openstack-infra21:39
clarkbjeblair: oh right its the metadata, with small pypi packages that ends up being a large chunk of time but I imagine for big ataomic fedora images it isn't21:39
jeblairclarkb: yeah, that's what i'd expect21:40
mnaserjeblair oh we'll be using the same image and bumping it only when we need, so a stat is nothing compared to the pain we'd deal with21:40
fungijeblair: that's a openstack/openstack-ansible-openstack_openrc change, is that the number you meant?21:40
mnaserthis reminds me of back when i had to run a glusterfs cluster21:40
jeblairfungi: nope21:40
mnaserit did a stat on every access across the entire cluster21:40
jeblairfungi: how about 492685 ? :)21:40
fungilet's see21:40
fungijeblair: done, seems prudent to keep work on that flowing21:41
jeblairmnaser: afs's forward cache invalidation is nice there -- we only have to do the stat once after a volume release (which we're doing no more often than every 2 hours)21:41
openstackgerritRyan proposed openstack-infra/bindep master: Add ability to list all deps  https://review.openstack.org/49269321:42
mnaserjeblair seems much more reasonable for low write patterns (which is really the case of mirrors)21:42
*** jtomasek has joined #openstack-infra21:43
*** spzala has quit IRC21:44
jeblairmnaser: yeah, i think the biggest disappointment is the pypi case where we release very frequently and also have tons of small files.  other uses still seem to be holding up well.21:44
mnaserjeblair depending on how nice pypi is with cache-control it could be a more suitable proxy caching case21:44
fungidistro package mirrors, or original use case, seem to continue to be a good fit21:44
fungis/or original/our original/21:45
clarkbmnaser: it doens't end up working so well for that at least not out of the box with a naive implementation21:45
clarkbmnaser: we tried it for a few days and pretty quiickly ran into "lib released we want it now its not there because cache"21:45
mnaserclarkb i guess it would need a bit more investment of time .. such as no caching on indexes but cache the .tar.gz files or something only21:46
mnaserim sure its a lot more complicated than that21:46
jeblairfungi, clarkb: more fun can be had with https://review.openstack.org/49268721:46
*** bobh has quit IRC21:47
clarkbmnaser: ya indexes is the biggest thing, but then you ar emore susceptible to failures because you ar egrabbing every index21:47
clarkbmnaser: definitelyworth fiddling more with after the release21:47
*** jtomasek has quit IRC21:47
mnaserclarkb eek true21:48
*** bobh has joined #openstack-infra21:48
clarkbianw: were you wanting to restart hte rax-ord mirror today?21:49
*** krtaylor has joined #openstack-infra21:49
*** thorst has joined #openstack-infra21:49
ianwclarkb: i can, if we agree it's worth a try21:49
ianwthe pypi mismatch errors do seem to be largely isolated to it21:50
clarkbjeblair: left a comment but +2'd (did not approve in case that is something yo uwant to change)21:50
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Zuulv3: Add project-config to check pipeline  https://review.openstack.org/49270021:50
clarkbianw: ya I think it likely is owrth a short21:50
clarkbwow I can't type21:50
openstackgerritMerged openstack-infra/project-config master: Borrow some internap quota for Zuulv3  https://review.openstack.org/49268521:51
jeblairclarkb: yes, it would compete, but there's a stack of changes to pull zuul and 2 other repos out of zuulv2 so only zuulv3 will gate them.  so we'll still only have one of the zuuls gating each repo.21:51
jeblairclarkb: (child of that change and its dependencies)21:52
*** bobh has quit IRC21:53
*** xyang1 has quit IRC21:54
*** thorst has quit IRC21:54
openstackgerritMerged openstack-infra/project-config master: Add Gnocchi charm and associated interfaces  https://review.openstack.org/48994621:55
fungii guess as long as the repos form a closed set from a shared gate queue perspective, there should be no issues21:56
*** Sukhdev has quit IRC21:57
pabelangerclarkb: I think we are getting IO issues in citycloud-lon121:58
pabelangerhttp://paste.openstack.org/show/618108/21:58
pabelangermore then 1.5 hours to chown /opt/git21:58
clarkbpabelanger: ya the dstat sdague linked to seemed to agree21:58
clarkbpabelanger: I'm guessing it sa bad hypervisor and related to your original email though21:59
pabelangerya, think so21:59
*** baoli has quit IRC22:01
pabelangerclarkb: fungi: sdague: did we want to consider restoring 49249322:01
pabelangermirror.lon1 is back up to 10 load also22:02
clarkbpabelanger: well the commit message would need rewriting (the mirror is fine aiui)22:02
clarkboh huh22:02
fungiright, if there's an overloaded hypervisor there, then the theory that we got the mirror migrated to a better compute node but are still getting random instances scheduled onto the bad one seems reasonable22:02
*** tesseract has quit IRC22:02
pabelangerand seems to be having IO issues22:02
fungioh, really we have slow performance on the mirror instance again too?22:02
fungii wonder if we're dos'ing their cloud :/22:03
pabelangerwait22:03
clarkbits all in wai too22:03
pabelangerhtcacheclean has multiple processes agan22:03
pabelangermaybe puppet hasn't ran there22:03
pabelangerbut, IO is slow on the mirror22:03
clarkbI don't see any flocks22:04
pabelangerhostkey on mirror.lon1.citycloud.openstack.org changed, so puppet hasn't connected22:04
pabelangerclarkb: fungi: is that expected if we migrated the VM?22:05
fungishouldn't be, no22:05
clarkbno, possibly it wasn't done before (though I thought I got all of them)22:05
*** Apoorva_ has joined #openstack-infra22:05
*** baoli has joined #openstack-infra22:05
clarkbalso did it get the proxy updates?22:06
clarkbif it got proxy updates it should have had working puppet at one time22:06
*** priteau has quit IRC22:07
pabelanger2017-08-10 07:30:30,027 p=21556 u=root |  mirror.lon1.citycloud.openstack.org : ok=5    changed=1    unreachable=1    failed=022:07
pabelangerfirst time is started failing22:07
clarkbit has the proxy config22:07
clarkbso it did update recently if the key did indeed change22:07
clarkbpabelanger: did you test ssh? could just be that server timed out making the ssh connection?22:08
pabelangerclarkb: ya, I can SSH into it22:08
pabelangerjust need to accept new host key22:08
pabelangerbut, not sure why it would have changed22:08
fungiianw: is that around when the server migration happened?22:08
clarkbI mean from the puppetmaster22:08
clarkbjust want to make sure that puppet master does see a new key and this wasn't related to the io problems22:09
pabelangerclarkb: yes, I can hit it from puppet master22:09
*** Apoorva has quit IRC22:09
*** Sukhdev has joined #openstack-infra22:09
pabelangerhttp://paste.openstack.org/show/618110/22:09
clarkbpabelanger: and did it ask toconfirm a new key there?22:09
pabelangerclarkb: yup22:09
pabelangerit wants to remove old and accept new22:10
ianwis 07:30 utc?22:10
clarkbfun so ya definitely changed22:10
pabelangerianw: ya22:10
ianwmigration and reboots happened ~ 10 hours ago22:10
pabelanger/etc/ssh have new timestamps on our host keys22:11
pabelangerlet me see what changed them22:11
mordredI think depending on how the migration is implemented, it can be seen by cloud-init as a boot of a new server from an image snapshot of the server - and thus cloud-init would generate a new host key22:12
pabelangerya, it is cloud-init22:12
pabelangerAug 10 12:16:56 mirror [CLOUDINIT] util.py[DEBUG]: Running command ['/usr/lib/cloud-init/write-ssh-key-fingerprints', '', 'ssh-dss'] with allowed return codes [0] (shell=False, capture=True)22:13
fungioh, so not a live migration at all22:13
pabelangermordred comment appears to be correct22:13
fungijust happened to boot from a snapshot of the old instance and keep the same ip addresses or something?22:13
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Bindmount /etc/lsb-release into bubblewrap  https://review.openstack.org/49020022:13
mordredyah - certainly seems more like a snap/boot22:13
mordredfungi: yah22:13
fungik22:13
fungithen the changing host key isn't a complete surprise at least. that's good22:14
clarkbstep 0 is probably get htcacheclean ing working as expected then we can compare again?22:14
pabelangerwe are confortable to accepting new hostkeys then?22:14
clarkbpabelanger: yes I think so22:14
pabelangerk22:14
clarkbif the large  wai comes back after that is fixed then we should consider restoring sdague's change22:14
*** priteau has joined #openstack-infra22:15
pabelangergoing to kick.sh mirror.lon1.citycloud.openstack.org to confirm puppet runs22:15
pabelangermordred: we should consider using the fact-cache for ansible on puppetmaster.o.o too. shave a few seconds on each time we loop our ansible-playbook commands22:17
*** openstackgerrit has quit IRC22:18
pabelangerclarkb: fungi: ianw: I also confirmed we start htcacheclean from by default with apache2 service, to clean our mod_cache_dir default directory.  Will propose a patch to disable that too, since we are not using it22:19
ianwdmsimard: on the python3.5 ... yeah fedora26 comes with python3.622:19
*** priteau has quit IRC22:20
clarkbdmsimard: ianw where are you having trouble with that? is it devstack? It hardcodes the version of python3 maybe we should just let it run whatever python3 is present?22:20
ianwi haven't quite got to devstack+python3+fedora26 yet, but it's on my list :)  dmsimard just mentioned it before22:21
*** openstackgerrit has joined #openstack-infra22:21
openstackgerritMerged openstack-infra/project-config master: Zuulv3: add the gate pipeline  https://review.openstack.org/49268722:21
pabelangerYa, I think this mirror might need to be migrated or rebuilt, we are at 13+ load with puppetmaster connected via ansible22:22
pabelangersome IO issues for sure22:22
clarkbpabelanger: well it is running a whole bunch of processes trying to stat disk right?22:23
*** slaweq has quit IRC22:23
clarkbpabelanger: the apache run htcacheclean should be fine since its for a dir we don't use22:23
clarkb(basically lets get it to a known good state then evaluate the io problems)22:24
ianwafs_background is writing a couple hundred kb/s on it, but not much else22:26
*** skelso has quit IRC22:27
ianwwatching iotop it's actually more like 1mb/s all up and pretty constant from afs.  it doesn't seem like much, but i'm not sure what to expect22:29
jeblairianw: that may be afs saving data to its cache as fast as it is able to stream it from the server?22:30
openstackgerritRamamani Yeleswarapu proposed openstack-infra/project-config master: Enable TLS in ironic gate jobs except grenade  https://review.openstack.org/49223122:33
*** EricGonc_ has quit IRC22:33
ianwyeah, i'd say ... is there a proc node for it or something22:35
ianweven still, if it can't keep up a 1mb/s, it's not going to be having much fun22:35
*** felipemonteiro has quit IRC22:36
openstackgerritMonty Taylor proposed openstack-infra/shade master: Make get_server_console tests more resilient  https://review.openstack.org/49268322:37
clarkbianw: ya22:37
jeblairclarkb: can you ack https://review.openstack.org/492697  please?22:42
clarkbjeblair: probably worth warning project-config reviewers that layout checks against infra repos will basically be skipped22:43
clarkb(thinking AJaeger_ in particular)22:44
jeblairclarkb: yes, i'll send email22:44
fungiAJaeger_: is travelling this week so yes e-mail will be good22:44
pabelangerAug 10 22:43:42 mirror puppet-user[5313]: (/Stage[main]/Openstack_project::Mirror/Cron[apache-cache-cleanup]/command) command changed 'htcacheclean -n -p /var/cache/apache2/proxy -t -l 81920M > /dev/null' to 'flock -n /var/run/htcacheclean.lock htcacheclean -n -p /var/cache/apache2/proxy -t -l 81920M > /dev/null'22:44
fungialso, not sure whether anybody else noticed (possible i'm the only one here who cares anyway?) but nist's sp 800-63-3 update officially drops the recommendation to periodically change passwords22:45
clarkbalso this shouldn't affect depends on22:45
pabelangerianw: fungi: do either of you mind looking at https://review.openstack.org/492666/ to bring .npmrc back online22:45
mordredfungi: neat22:46
jeblairfungi: oh i didn't notice that.  i had heard about all the other good stuff (like don't require weird chars, allow long passphrases, allow copy/paste, etc)22:46
clarkbjeblair: only other thing I can think of is how worried are you about wedging zuul (are things in flux enough to make that a big concern?)22:46
clarkbI guess worst cas eyou apply fix directly then have that gate the fix22:47
clarkb(if that makes sense22:47
fungiyeah, all it took was decades of security researchers complaining that forcing periodic passwords changes did more harm than good22:47
*** esberglu_ has quit IRC22:47
jeblairclarkb: it is possible, maybe even likely.  but i think it's worth going ahead and exercising it a bit more, and maybe occasionally we have to force push or fix as you suggest.22:48
openstackgerritMerged openstack-infra/openstack-zuul-jobs master: Add openstack-zuul-jobs to gate pipeline  https://review.openstack.org/49269222:48
*** annegentle has quit IRC22:48
mordredfungi: now how long will it take for our of date policies to catch up with the new recommendations?22:48
fungiheh22:49
*** annegentle has joined #openstack-infra22:49
clarkbby the time we get there nist will have decided rotations are a good thing again22:49
fungimordred: i'm at least glad i pushed keystone hard to do pbkdf2 when they changed out their password hash backend22:49
fungiit's one of the couple of key derivations mentioned as recommended now22:49
fungi(the other being balloon)22:50
*** yamamoto has joined #openstack-infra22:51
fungias for hamstringing/confusing zuul, we've just caught up on the job backlog in the past couple hours, so at least as long as we keep an eye on it and notice sudden issues i don't think we're likely to cause significant adverse impact unless we allow it to persist until tomorrow22:51
clarkbfungi: oh I meant for zuul gating with v3 zuul22:52
clarkbI don't expect there will be problems for everyone else22:52
fungioh, got it. thought maybe you were concerned about corner cases in v2 reconfiguration22:52
fungiunknown unknowns22:52
pabelangerodyssey4me: just seen some traffic to images.linuxcontainers reverse proxy cache on a mirror, looks to be working22:53
*** annegentle has quit IRC22:53
jeblairclarkb: email sent, thanks22:54
*** spzala has joined #openstack-infra22:55
*** spzala has quit IRC22:55
*** spzala has joined #openstack-infra22:55
*** spzala has quit IRC22:55
*** spzala has joined #openstack-infra22:56
*** spzala has quit IRC22:56
*** spzala has joined #openstack-infra22:56
*** spzala has quit IRC22:56
openstackgerritMerged openstack-infra/project-config master: Revert "Remove NPM mirror settings"  https://review.openstack.org/49266622:57
*** spzala has joined #openstack-infra22:57
*** spzala has quit IRC22:57
*** spzala has joined #openstack-infra22:57
openstackgerritMerged openstack-infra/zuul-jobs master: Add zuul-jobs to gate pipeline  https://review.openstack.org/49269122:57
*** spzala has quit IRC22:58
*** vhosakot has quit IRC22:58
*** vhosakot has joined #openstack-infra22:59
openstackgerritPaul Belanger proposed openstack-infra/project-config master: Add SSH private key for static.o.o  https://review.openstack.org/49267122:59
mordredjeblair: 492697 has 3 +2s already - so you can pull the trigger whenever23:00
*** abelur_ has joined #openstack-infra23:01
jeblairmordred: ya was going to wait until the final move lands23:01
mordred++23:01
jeblairmordred, pabelanger: while we're thinking about it, https://review.openstack.org/492700 would be really good23:02
pabelangerya, +223:02
openstackgerritMerged openstack-infra/zuul feature/zuulv3: Add Zuul to gate pipeline  https://review.openstack.org/49268923:03
clarkbodyssey4me: pabelanger linux bridge + vxlan is being tested with a neutron change (they have a few jobs that rely on this overlay setup) I've also got one more improvement (to make the setup more symmetrical on the test nodes) that I will push once we have first round of results from neutron jobs23:03
pabelangerclarkb: cool, sounds promissing23:04
*** annegentle has joined #openstack-infra23:05
*** rbrndt has quit IRC23:05
*** Swami has quit IRC23:05
openstackgerritMerged openstack-infra/shade master: Make QoS rules required parameters to be not optional  https://review.openstack.org/49103323:06
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Fix typo in nodepoolv3 config  https://review.openstack.org/49270623:06
jeblairmordred, pabelanger: ^ whoops23:06
clarkbpabelanger: ya if this ends up not making existing jobs unhappy I think we can likely merge it post releaes then update our images to remove rdo repo23:06
fungithat'll be a huuuuge improvement23:08
clarkbpabelanger: do you know if any of the puppet or tripleo jobs use this overlay stuff too? I could depends on in changes to them to double check them (but probably won't worry about that until I know neutron is happy)23:08
clarkbalso 3.8 seems to be the magic kernel that you need23:08
clarkband centos is 3.10 so hopeful it will work there too23:09
jeblairi manually made that change on nl0123:09
*** lihi has quit IRC23:10
*** dimak has quit IRC23:10
*** dimak has joined #openstack-infra23:11
*** pbourke has quit IRC23:11
*** xarses_ has quit IRC23:11
*** lihi has joined #openstack-infra23:11
*** gouthamr has quit IRC23:12
openstackgerritsebastian marcet proposed openstack-infra/openstackid-resources master: External Calendar Sync  https://review.openstack.org/48768323:12
*** pbourke has joined #openstack-infra23:13
*** annegentle has quit IRC23:13
openstackgerritMerged openstack-infra/project-config master: Zuulv3: Remove check/gate jobs from zuul and friends  https://review.openstack.org/49269723:14
jeblairthat's a milestone ^ :)  zuul v3 has moved from "self-hosting" to "self-gating"23:17
*** funzo has joined #openstack-infra23:17
jeblair(we hope; i guess we'll know when we approve the next change ;)23:17
mordredjeblair: \o/23:17
mordredjeblair: that's what I'm looking forward to ...23:17
openstackgerritMerged openstack-infra/project-config master: Zuulv3: Add project-config to check pipeline  https://review.openstack.org/49270023:17
pabelangerianw: clarkb: sign, glean is running after rc.local on centos-7. So /etc/resolv.conf is not pointing to nameserver 127.0.0.123:18
pabelangerianw: clarkb: all centos-7 nodes in infracloud just ping to 8.8.8.823:19
pabelangerpoint*23:19
clarkbpabelanger: huh, I don't think anything in the diff should've affected that right?23:19
clarkbpabelanger: we aren't on fire with that though right? its working just not as we intend?23:19
pabelangerglean==1.9.123:20
pabelangerya, this is an old image23:20
clarkbah ok at least it isn't a new regression from 1.9.223:20
pabelangerclarkb: right, just seen a job fail on DNS in infracloud-vanilla. First time I see that23:20
*** funzo has quit IRC23:21
pabelangerclarkb: we are getting 8.8.8.8 from config-drive, so we should be able to just change it to 127.0.0.1 for now right?23:21
pabelangerthen work on patch to glean for configfile support and disable DNS updates23:21
*** hongbin has quit IRC23:21
clarkbpabelanger: ya thats a network setting in the cloud. Thats an interesting hack to make it do what we want.23:22
ianwi think i'm missing some context; did this just start happening?23:22
pabelangerianw: not sure why it started, but I just seen a job in infracloud-vanilla fail Could not resolve host: mirror.regionone.infracloud-vanilla.openstack.org23:23
clarkbianw: no I don't think it just started, guessing its just been noticed bceause otherwise google dns usually works23:23
pabelangerhttp://logs.openstack.org/63/491463/5/gate/gate-tripleo-ci-centos-7-scenario001-multinode-oooq-puppet/5055004/logs/undercloud/home/jenkins/undercloud_install.log.txt.gz23:23
pabelangerI confirmed ubuntu-xenial is okay23:23
*** slaweq has joined #openstack-infra23:23
clarkb(but I could be wrong maybe something in centos packaging bumped systemd ordering recently?)23:23
pabelangerbut I manually booted a node and confirmed glean is running after rc.local service23:23
pabelangerclarkb: that is possible23:24
openstackgerritMerged openstack-infra/project-config master: Fix typo in nodepoolv3 config  https://review.openstack.org/49270623:26
openstackgerritsebastian marcet proposed openstack-infra/openstackid-resources master: External Calendar Sync  https://review.openstack.org/48768323:27
pabelangerclarkb: okay, maybe tomorrow I'll try creating a new subnet in infracloud-vanilla with 127.0.0.1 as the dns server23:27
clarkbthat may be something you can update in existing subnets? not sure23:28
ianwrc.local is just "After=network.target"23:28
pabelangerclarkb: ya, we can. Wasn't sure if we wanted to do it live23:28
*** slaweq has quit IRC23:28
clarkbpabelanger: if we double check the nodes boot and get an unbound running with the right config and with the right resolvers and listening on localhost I think that is relatively safe23:29
pabelangerclarkb: okay, want me to update it now?23:30
fungiafaik image uploads to infracloud have been paused for a couple days23:30
clarkbpabelanger: one downside to that is you won't be able to boot anything in that subnet unless it is configured to have a local resolver23:30
clarkb(and since it is provider networking I think everything is on that one subnet)23:30
pabelangerYa, we have single subnet provider-subnet-infracloud23:30
*** caphrim007 has quit IRC23:30
*** caphrim007 has joined #openstack-infra23:31
pabelangeractually, I wonder if we could just remove dns server23:32
pabelangerand glean will just skip setting it23:32
pabelangerthen we'll default to 127.0.0.123:32
pabelangerclarkb: ^thoughts on that?23:33
pabelangeropenstack subnet set --name provider-subnet-infracloud --no-dns-nameservers should be the command23:33
*** yamamoto has quit IRC23:33
*** claudiub has quit IRC23:35
*** Swami has joined #openstack-infra23:35
*** caphrim007 has quit IRC23:36
clarkbpabelanger: that may be more friendly to other hosts that may not have an unbound running (though I don't know that we'd do that so maybe its not worth worrying about)23:36
pabelangerclarkb: k, let me try it real quick. It shouldn't affect nodepool, just won't be able to run configure-mirror.sh if failed23:38
*** thorst has joined #openstack-infra23:38
pabelangerokay, centos-7 booted with 127.0.0.123:40
clarkband dns works?23:40
pabelangerand clean ignored writing /etc/resolv.conf23:40
pabelangeryup23:40
clarkbhax23:40
*** yamamoto has joined #openstack-infra23:40
pabelangerlet me look at nodepool debug to make sure nodes are still booting23:41
pabelangerclarkb: ya, nodepool is happy. Going to do the same in chocolate now23:45
clarkbcool23:45
*** vhosakot has quit IRC23:45
openstackgerritJames E. Blair proposed openstack-infra/zuul feature/zuulv3: Bindmount /etc/lsb-release into bubblewrap  https://review.openstack.org/49020023:45
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Override tox requirments with zuul git repos  https://review.openstack.org/48971923:46
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: Rename tox_command_line in docs to tox_extra_args  https://review.openstack.org/48975823:46
openstackgerritMonty Taylor proposed openstack-infra/zuul-jobs master: WIP job for non-OpenStack sphinx build  https://review.openstack.org/49270923:46
pabelangerclarkb: done23:46
openstackgerritJames E. Blair proposed openstack-infra/project-config master: Zuul v3: add description to check pipeline  https://review.openstack.org/49271023:47
pabelanger#status log removed 8.8.8.8 dns servers from both infracloud-chocolate and infracloud-vanilla provider-subnet-infracloud subnet23:47
*** sdague has quit IRC23:47
*** pabelanger has quit IRC23:47
*** pabelanger has joined #openstack-infra23:47
pabelanger#status log removed 8.8.8.8 dns servers from both infracloud-chocolate and infracloud-vanilla provider-subnet-infracloud subnet23:47
openstackstatuspabelanger: finished logging23:47
openstackgerritJeremy Stanley proposed openstack-infra/infra-specs master: Gerrit ContactStore Removal is implemented  https://review.openstack.org/49228723:47
fungipabelanger: i guess that's something we need to encode in the puppet module?23:48
clarkblon1 mirror fixes have been in place for almost an hour, load average is down to ~4 which is an improvement but still high wait23:49
pabelangerfungi: ya, I'm looking to see where we set that up.  Not sure if cloud-launcher or puppet is the place23:49
pabelangerYup, puppet-infracloud23:50
jeblairpabelanger: http://logs.openstack.org/00/490200/5/check/tox-py35/cad03b1/job-output.txt.gz#_2017-08-10_23_49_14_71833623:50
jeblairdoes that mean anything to you?23:50
jeblairthat may be one of our first zuulv3 jobs on internap23:51
openstackgerritIan Wienand proposed openstack/diskimage-builder master: [WIP] LVM support for dib-block-device  https://review.openstack.org/47206523:51
pabelangerjeblair: apt task needs python bindings for apt. It must not be able to install that package for some reason23:51
pabelangerjeblair: I think we could avoid that step by adding it to our DIB?23:52
jeblairpabelanger: hrm... are we in a chicken/egg situation?23:53
jeblairpabelanger: the apt-module needs python-apt to be installed in order to run apt commands, but we're using it to run apt-get update so that apt has a package cache and can function?23:53
pabelangerjeblair: ya, I think so. I am guess our cache apt cache is too old to find python-apt for us to update the cache23:53
pabelangerjeblair: I think so23:53
jeblairpabelanger: well, we did just change the source list under it, so i think that makes it immediately out of date23:54
jeblairpabelanger: should we just make that a shell command?23:54
pabelangerjeblair: I think in this case, adding it infra-packages-needs like we do for python-selinux make sense23:54
pabelangerjeblair: Ya, maybe shell in this case23:54
jeblairpabelanger: since this is in zuul-jobs, that might make it most widely compatible without us having to tell folks they need stuff on their images23:55
openstackgerritPaul Belanger proposed openstack-infra/puppet-infracloud master: Remove dns_servers from provider-subnet-infracloud  https://review.openstack.org/49271223:55
pabelangerjeblair: yes, I agree23:56
jeblairpabelanger: i'm working on a change23:56
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Add publish-openstack-artifacts base job  https://review.openstack.org/49271323:56
openstackgerritMonty Taylor proposed openstack-infra/project-config master: Remove openstack-publish-tarball base job  https://review.openstack.org/49271423:56
openstackgerritMonty Taylor proposed openstack-infra/openstack-zuul-jobs master: Use openstack-publish-artifacts base job  https://review.openstack.org/49271523:56
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Use shell for apt-get update  https://review.openstack.org/49271623:57
jeblairpabelanger, mordred: did i ansible right? ^23:57
mordredjeblair: looking.23:57
mordredjeblair, pabelanger: also - there's the rename stack for the base job23:57
pabelangerHmm, I think ansible-lint will fail. lets see, but looks right23:58
mordredjeblair: yes. although command is preferred to shell unless you actually need shell23:58
jeblairmordred: i can never remember :)23:58
jeblairpabelanger: what's linty?23:58
pabelangerwhat mordred just said23:58
jeblairok i'll just change it then23:59
pabelangertag: skip_ansible_lint I think is how we ignore linting per task23:59
openstackgerritJames E. Blair proposed openstack-infra/zuul-jobs master: Use shell for apt-get update  https://review.openstack.org/49271623:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!