Tuesday, 2018-12-11

corvusi believe it will ignore it00:01
pabelangeryup, just testing on 3.1.1. Not errors on PR00:03
*** efried has quit IRC00:05
*** hwoarang has quit IRC00:06
*** hwoarang has joined #openstack-infra00:08
*** efried has joined #openstack-infra00:10
*** slaweq has joined #openstack-infra00:11
*** efried has quit IRC00:14
*** efried has joined #openstack-infra00:14
*** efried has quit IRC00:15
*** slaweq has quit IRC00:16
*** rh-jelabarre has quit IRC00:18
*** rh-jelabarre has joined #openstack-infra00:23
clarkbI've not seen a successful upload to inap yet, but I think those that have failed may have started before mgagne_ fixed thigns? if they are still failing in another hour so then liekly not fixed sdk side00:28
clarkban image just went ready in inap \o/00:32
clarkbmnaser: ^ fyi we should be closing in on fixing that centos image problem00:32
ianw:/ it looks like the readthedocs isn't triggering correctly any more.  unfortunately to not echo out the password we have no_log on the important bits00:32
mordredclarkb: I was only mildly following - mgagne found a thing?00:33
clarkbmordred: oui00:33
mordredclarkb: awesome00:33
clarkbso this may have been entirely cloud side00:33
clarkbI think it helped that osc was able to reproduce a failure if not the same one00:34
*** wolverineav has quit IRC00:36
*** wolverineav has joined #openstack-infra00:37
clarkb| 0000000040 | 0000000017 | inap-mtl01           | centos-7             | centos-7-1544483708             | 9416a0d2-48f9-43c3-9aed-271635b897dd | ready     | 00:00:02:26 |00:39
clarkbosa centos jobs should be happy now00:39
clarkbif they start on new nodes00:39
*** wolverineav has quit IRC00:42
*** kjackal has joined #openstack-infra00:43
*** jcoufal has quit IRC00:44
ianw... {"detail":"CSRF Failed: CSRF cookie not set."} ... i do not like the look of this, rtd might have broken access to the authenticated endpoint00:45
*** yamamoto has quit IRC00:49
*** _alastor_ has joined #openstack-infra00:59
*** rockyg has quit IRC00:59
*** sthussey has quit IRC01:03
*** wolverineav has joined #openstack-infra01:05
*** rockyg has joined #openstack-infra01:07
*** wolverineav has quit IRC01:07
*** wolverineav has joined #openstack-infra01:07
ianwwell i don't think there's much we can do ... filed https://github.com/rtfd/readthedocs.org/issues/498601:08
*** rkukura has quit IRC01:12
*** rockyg has quit IRC01:14
mnaserclarkb: thank you so much!01:14
*** ianychoi has quit IRC01:20
*** _alastor_ has quit IRC01:25
*** bobh has joined #openstack-infra01:27
*** bobh has quit IRC01:31
*** hwoarang has quit IRC01:36
*** hwoarang has joined #openstack-infra01:37
*** kjackal has quit IRC01:40
*** rkukura has joined #openstack-infra01:55
*** neilsun has joined #openstack-infra01:58
*** _alastor_ has joined #openstack-infra02:01
*** _alastor_ has quit IRC02:06
*** mrsoul has joined #openstack-infra02:07
*** bobh has joined #openstack-infra02:08
*** bobh has quit IRC02:12
*** jistr has quit IRC02:42
*** jistr has joined #openstack-infra02:50
*** psachin has joined #openstack-infra02:52
*** anteaya has quit IRC03:01
*** bhavikdbavishi has joined #openstack-infra03:07
*** apetrich has quit IRC03:15
*** hongbin has joined #openstack-infra03:15
*** rh-jelabarre has quit IRC03:38
*** bobh has joined #openstack-infra03:39
*** ykarel|away has joined #openstack-infra03:42
*** bobh has quit IRC03:43
*** udesale has joined #openstack-infra03:48
*** gyee has quit IRC03:53
*** jamesden_ has joined #openstack-infra03:54
*** agopi_ has joined #openstack-infra03:54
*** agopi has quit IRC03:54
*** jamesdenton has quit IRC03:55
*** markvoelker has joined #openstack-infra03:57
*** ramishra has quit IRC03:59
*** bobh has joined #openstack-infra04:01
*** markvoelker has quit IRC04:02
*** bobh has quit IRC04:06
*** udesale has quit IRC04:08
*** mriedem_away has quit IRC04:15
*** _alastor_ has joined #openstack-infra04:23
*** wolverineav has quit IRC04:29
*** wolverineav has joined #openstack-infra04:30
*** slaweq has joined #openstack-infra04:30
*** jamesmcarthur has joined #openstack-infra04:31
openstackgerritTristan Cacqueray proposed openstack-infra/zuul master: executor: add support for generic build resource  https://review.openstack.org/57066804:38
*** udesale has joined #openstack-infra04:43
*** yamamoto has joined #openstack-infra04:46
*** lpetrut has joined #openstack-infra04:55
*** wolverineav has quit IRC05:05
*** hongbin has quit IRC05:07
*** hwoarang has quit IRC05:10
*** hwoarang has joined #openstack-infra05:12
*** wolverineav has joined #openstack-infra05:13
*** jamesmcarthur has quit IRC05:18
*** jamesmcarthur has joined #openstack-infra05:18
*** ykarel|away has quit IRC05:19
*** chandan_kumar has joined #openstack-infra05:21
*** jamesmcarthur has quit IRC05:23
*** ramishra has joined #openstack-infra05:26
*** agopi_ is now known as agop05:30
*** agop is now known as agopi05:30
*** ykarel|away has joined #openstack-infra05:35
*** lpetrut has quit IRC05:35
*** ykarel|away is now known as ykarel05:42
*** jamesmcarthur has joined #openstack-infra05:50
*** _alastor_ has quit IRC05:50
*** wolverineav has quit IRC05:54
*** yboaron_ has joined #openstack-infra06:00
openstackgerritOpenStack Proposal Bot proposed openstack-infra/project-config master: Normalize projects.yaml  https://review.openstack.org/62427706:06
openstackgerritgaobin proposed openstack-infra/zuul master: Modify some file content errors  https://review.openstack.org/62427806:08
openstackgerritgaobin proposed openstack-infra/zuul master: Modify some file content errors  https://review.openstack.org/62427806:11
*** wolverineav has joined #openstack-infra06:18
*** _alastor_ has joined #openstack-infra06:19
*** _alastor_ has quit IRC06:23
*** betherly has quit IRC06:24
*** ahosam has joined #openstack-infra06:29
*** jmccrory has quit IRC06:34
*** sdake has quit IRC06:35
*** jmccrory has joined #openstack-infra06:40
*** sdake has joined #openstack-infra06:40
*** apetrich has joined #openstack-infra06:40
*** bobh has joined #openstack-infra06:41
*** rcernin has quit IRC06:43
*** wolverineav has quit IRC06:44
*** bobh has quit IRC06:46
*** jamesmcarthur has quit IRC06:50
*** rlandy has quit IRC06:51
openstackgerritTristan Cacqueray proposed openstack-infra/zuul-base-jobs master: Add base openshift job  https://review.openstack.org/57066906:53
*** wolverineav has joined #openstack-infra07:01
*** ahosam has quit IRC07:08
*** e0ne has joined #openstack-infra07:10
*** e0ne has quit IRC07:12
*** e0ne has joined #openstack-infra07:13
*** e0ne has quit IRC07:14
*** quiquell|off is now known as quiquell07:14
*** ramishra has quit IRC07:15
*** e0ne has joined #openstack-infra07:15
*** e0ne has quit IRC07:17
openstackgerritMerged openstack-infra/project-config master: Normalize projects.yaml  https://review.openstack.org/62427707:22
*** bobh has joined #openstack-infra07:25
*** wolverineav has quit IRC07:29
*** jtomasek has joined #openstack-infra07:29
*** bobh has quit IRC07:29
*** ykarel is now known as ykarel|lunch07:31
*** kjackal has joined #openstack-infra07:34
*** yboaron_ has quit IRC07:35
openstackgerritMerged openstack-infra/zuul master: Add spacing to Queue lengths line  https://review.openstack.org/62396007:37
*** jtomasek has quit IRC07:42
*** jtomasek has joined #openstack-infra07:43
*** ahosam has joined #openstack-infra07:43
*** rossella_s has quit IRC07:46
*** ahosam has quit IRC07:49
*** quiquell is now known as quiquell|brb07:53
*** oanson has quit IRC07:54
*** agopi_ has joined #openstack-infra07:56
*** tosky has joined #openstack-infra07:58
*** agopi has quit IRC07:59
*** ramishra has joined #openstack-infra08:01
*** bobh has joined #openstack-infra08:01
*** ginopc has joined #openstack-infra08:02
*** longkb has joined #openstack-infra08:02
*** rossella_s has joined #openstack-infra08:03
*** ccamacho has joined #openstack-infra08:04
*** agopi_ is now known as agopi08:04
*** bobh has quit IRC08:05
*** kjackal has quit IRC08:06
*** mgoddard has quit IRC08:10
*** mgoddard has joined #openstack-infra08:10
*** agopi_ has joined #openstack-infra08:11
*** agopi has quit IRC08:14
*** kjackal has joined #openstack-infra08:15
*** shardy has joined #openstack-infra08:15
amorinhey frickler and others, I am moving your instances on separate hosts08:20
amorinin the meantime, we found a issue on the hypervisors08:20
amorinabout RAM usage08:20
amorinif the instances are not having enough memory, they could be using swap instead, which could cause them to slow down a lot08:21
*** imacdonn has quit IRC08:22
*** imacdonn has joined #openstack-infra08:23
*** _alastor_ has joined #openstack-infra08:25
*** agopi_ is now known as agopi08:27
*** bhavikdbavishi has quit IRC08:28
*** shardy has quit IRC08:28
*** hwoarang has quit IRC08:28
*** shardy has joined #openstack-infra08:29
*** ykarel|lunch is now known as ykarel08:30
*** _alastor_ has quit IRC08:30
*** hwoarang has joined #openstack-infra08:30
*** ahosam has joined #openstack-infra08:32
*** ahosam has quit IRC08:32
*** priteau has joined #openstack-infra08:39
*** quiquell|brb is now known as quiquell08:40
*** ramishra has quit IRC08:44
*** ramishra has joined #openstack-infra08:51
*** bobh has joined #openstack-infra08:51
*** bobh has quit IRC08:56
*** yamamoto has quit IRC09:01
*** ahosam has joined #openstack-infra09:02
*** jpena|off is now known as jpena09:03
*** dpawlik has quit IRC09:03
*** dpawlik has joined #openstack-infra09:04
*** eumel8 has joined #openstack-infra09:05
*** ahosam has quit IRC09:05
*** wolverineav has joined #openstack-infra09:07
*** jtomasek_ has joined #openstack-infra09:08
*** dpawlik has quit IRC09:08
*** jpich has joined #openstack-infra09:09
*** jtomasek has quit IRC09:10
*** wolverineav has quit IRC09:12
*** kjackal has quit IRC09:13
*** kjackal has joined #openstack-infra09:14
*** yamamoto has joined #openstack-infra09:18
*** lpetrut has joined #openstack-infra09:21
AJaegerianw: looking at https://review.openstack.org/621840 - do you have a change that tests it and shows that it does the right thing?09:21
*** yamamoto has quit IRC09:27
*** bobh has joined #openstack-infra09:30
*** derekh has joined #openstack-infra09:36
*** dpawlik has joined #openstack-infra09:39
*** dpawlik has quit IRC09:39
*** dpawlik has joined #openstack-infra09:39
*** aojea has joined #openstack-infra09:40
*** pbourke_ has quit IRC09:54
*** yamamoto has joined #openstack-infra10:06
*** yamamoto has quit IRC10:10
*** e0ne has joined #openstack-infra10:12
*** rossella_s has quit IRC10:21
*** priteau has quit IRC10:24
*** pbourke has joined #openstack-infra10:26
*** electrofelix has joined #openstack-infra10:28
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Only reset working copy when needed  https://review.openstack.org/62434310:31
*** gfidente has joined #openstack-infra10:34
*** rossella_s has joined #openstack-infra10:35
*** yamamoto has joined #openstack-infra10:45
*** yamamoto has quit IRC10:47
*** udesale has quit IRC10:56
*** tobias-urdin is now known as tobias-urdin|lun11:00
*** tobias-urdin|lun is now known as tobias-urdin_afk11:01
*** yamamoto has joined #openstack-infra11:11
*** yamamoto has quit IRC11:16
*** yamamoto has joined #openstack-infra11:16
*** rfolco has quit IRC11:18
*** dtantsur|afk is now known as dtantsur11:18
*** rfolco has joined #openstack-infra11:23
*** tobias-urdin_afk is now known as tobias-urdin11:27
*** rossella_s has quit IRC11:28
*** longkb has quit IRC11:39
*** rossella_s has joined #openstack-infra11:43
*** quiquell is now known as quiquell|brb11:50
*** dpawlik has quit IRC11:56
*** ahosam has joined #openstack-infra11:57
*** dpawlik has joined #openstack-infra11:57
ssbarnea|roveri seen an interesting spike on timeouts which seems to re-occur after exactly one week: http://status.openstack.org/elastic-recheck/12:00
*** ahosam has quit IRC12:01
ssbarnea|roveri am considering adding a new query for POST specific timeouts as the generic one seems too generic and we have a signifiant number of post ones. anyone against?12:02
*** quiquell|brb is now known as quiquell12:13
sean-k-mooneyare there any docs on how to create an elastic recheck query12:16
sean-k-mooneyi want to create one for "os_vif error: [Errno 24] Too many open files" in the nova compute agent log12:17
*** wolverineav has joined #openstack-infra12:18
*** wolverineav has quit IRC12:22
*** yamamoto has quit IRC12:23
*** fresta_ is now known as fresta12:27
*** yamamoto has joined #openstack-infra12:28
*** jamesden_ is now known as jamesdenton12:29
*** jpena is now known as jpena|lunch12:32
*** yamamoto has quit IRC12:32
*** psachin has quit IRC12:35
*** rh-jelabarre has joined #openstack-infra12:39
*** jamesmcarthur has joined #openstack-infra12:42
*** jamesmcarthur has quit IRC12:46
*** e0ne has quit IRC12:47
*** dave-mccowan has joined #openstack-infra12:53
frickleramorin: oh, that could indeed explain our issues. can you work around it by adjusting quota? do you still want us to proceed with the other tests?12:55
openstackgerritChris Dent proposed openstack-infra/project-config master: Change os-resource-classes acl config to placement  https://review.openstack.org/62438713:00
ssbarnea|roversean-k-mooney: just create another file in queries/ folder, that's all. look at existing files to reverse engineer the docs ;)13:00
ssbarnea|roverin the end is a 4-5 lines yaml file13:00
*** dave-mccowan has quit IRC13:01
*** yamamoto has joined #openstack-infra13:05
*** gfidente has quit IRC13:09
sean-k-mooneyya i figured that out but i cant figure out the kibana/elastic serach query13:09
sean-k-mooneytags:"screen-n-cpu.txt" and message:"os_vif error: [Errno 24] Too many open files" and  project:"openstack/neutron"13:09
sean-k-mooneythat does not seam to work13:09
*** boden has joined #openstack-infra13:09
*** panda|off is now known as panda13:10
*** ykarel is now known as ykarel|afk13:12
*** trown|outtypewww is now known as trown13:13
*** jamesmcarthur has joined #openstack-infra13:15
fungiamorin: oh, were the hosts oversubscribed on ram? i agree that could have been an explanation13:18
*** gfidente has joined #openstack-infra13:21
fungisean-k-mooney: do you have a recent example of a job log in which that string appeared?13:23
fungimessage:"os_vif error: [Errno 24] Too many open files" isn't found in any indexed job logs for at least the past week13:23
*** zul has quit IRC13:26
*** zul has joined #openstack-infra13:26
*** rlandy has joined #openstack-infra13:28
sean-k-mooneyfungi: yes so currently its commign up as an uncatogerised issue13:31
sean-k-mooneyonce sec13:31
sean-k-mooneyhttp://logs.openstack.org/49/622449/4/check/neutron-tempest-iptables_hybrid/aa25876/logs/screen-n-cpu.txt.gz?level=TRACE#_Dec_11_10_29_35_87633613:32
*** jpena|lunch is now known as jpena13:32
sean-k-mooneyfungi: the neutron-tempest-iptables_hybrid entry on http://status.openstack.org/elastic-recheck/data/integrated_gate.html13:33
sean-k-mooneyis caused by https://bugs.launchpad.net/os-vif/+bug/180794913:33
openstackLaunchpad bug 1807949 in os-vif "os_vif error: [Errno 24] Too many open files" [High,Triaged] - Assigned to sean mooney (sean-k-mooney)13:33
sean-k-mooneyor rather by pyrout213:33
fungiwe do seem to be indexing that file in that job13:44
fungisince build_name:"neutron-tempest-iptables_hybrid" AND filename:"logs/screen-n-api.txt" returns plenty of hits in the past 6 hours13:45
fungibut appending AND message:"Too many open files" has 0 matches in 24 hours13:46
fungior even 48 hours, so should have caught that run13:47
*** jamesmcarthur has quit IRC13:47
*** jamesmcarthur has joined #openstack-infra13:48
fungibuild_short_uuid:"aa25876" has hits for that file too though13:48
*** kgiusti has joined #openstack-infra13:50
sean-k-mooneyok so atleast i did not completely missunderstand how to use kibana13:50
fungiwell, either that or i completely misunderstand how to use kibana too ;)13:51
fungicertainly not ruling that out13:51
sean-k-mooneyfilenema:"logs/screen-n-api.txt" is the wrong file by the way13:51
sean-k-mooneyit should be screen-n-cpu.txt13:52
fungid'oh, thanks!13:53
fungithat seemed to make a difference, though still no lines indexed with message:"os_vif error"13:54
sean-k-mooneycan you share your query by the way13:54
sean-k-mooneythis could become a neutorn gate blocker or it could just be intermitent so i wanted to get a querry to try and monitor it13:55
fungii'm currently combing through build_name:"neutron-tempest-iptables_hybrid" AND filename:"logs/screen-n-cpu.txt" AND build_short_uuid:"aa25876" AND message:"error"13:56
fungitrying to work out why that line is missing13:56
funginoting the entries are in reverse-chronological order13:56
*** ykarel|afk has quit IRC13:57
fungifound!13:57
fungithe message it parsed out for that line is "error: [Errno 24] Too many open files"13:57
fungiokay, now working to generalize13:58
*** sthussey has joined #openstack-infra14:00
*** jamesmcarthur has quit IRC14:02
*** jamesmcarthur has joined #openstack-infra14:03
fungisean-k-mooney: is the project:"openstack/neutron" part critical to this query?14:03
fungiis this showing up in multiple jobs, but only jobs run on changes to neutron and not to any other projects?14:04
fungitags:"screen-n-cpu.txt" AND message:"error: [Errno 24] Too many open files" AND project:"openstack/neutron" shows up starting around 09:00 utc today14:05
sean-k-mooneyno14:06
fungiif i drop the project filter, it's still the same number of hits14:06
sean-k-mooneyi think i have a query14:06
sean-k-mooneyhttp://logstash.openstack.org/#/dashboard/file/logstash.json?query=tags:%5C%22screen-n-cpu.txt%5C%22%20AND%20message:%5C%22OSError:%20%5BErrno%2024%5D%20Too%20many%20open%20files%5C%22%20AND%20module:%5C%22os_vif%5C%22%20AND%20loglevel:%20%5C%22ERROR%5C%2214:06
*** mriedem has joined #openstack-infra14:06
*** jamesmcarthur has quit IRC14:07
sean-k-mooneyfungi: i catully want to check the nova and neutron and kurry-kubernetes jobs14:07
sean-k-mooneyso droping is fine14:07
fungilgtm14:07
*** dave-mccowan has joined #openstack-infra14:08
*** rossella_s has quit IRC14:08
fungifirst hit does still seem to be around 09:0014:08
sean-k-mooneyya so we did a release yesterday of os-vif14:08
sean-k-mooneythe thing is i dont know if this is intermitent or if it always happens14:08
sean-k-mooneyi think the issue is caused by pyroute2 however14:09
*** rossella_s has joined #openstack-infra14:09
fungithe other litmus test is that appending AND build_status:"SUCCESS" returns 0 hits, which seems to be the case14:09
fungiso we know this pattern is present only in failed job runs14:09
sean-k-mooneycool so this is the tracking bug https://bugs.launchpad.net/os-vif/+bug/180794914:10
openstackLaunchpad bug 1807949 in os-vif "os_vif error: [Errno 24] Too many open files" [High,Triaged] - Assigned to sean mooney (sean-k-mooney)14:10
*** psachin has joined #openstack-infra14:11
sean-k-mooneyif i add tags:"screen-n-cpu.txt" AND message:"OSError: [Errno 24] Too many open files" AND module:"os_vif" AND loglevel: "ERROR" as the query in the elastic serch repo the file just needs the same name as the bug number right14:11
*** dave-mccowan has quit IRC14:14
*** quiquell is now known as quiquell|lunch14:18
*** ykarel|afk has joined #openstack-infra14:18
*** smarcet has joined #openstack-infra14:20
smarcetfungi: clarkb: morming , as i forementioned before, we need to migrate openstackid to latest Laravel version (5.6) and migrate puppet to start using php7.x, u mentioned that newer ubuntu version that u guys support is xenial, but xenial by default only support php 7.0 and i have a hard requirement to use PHP >= 7.1.3 bc https://laravel.com/docs/5.614:22
smarcetits posible for me to update the puppet to use this ppa ppa:ondrej/php and be able to use php 7.2 ?14:22
*** rossella_s has quit IRC14:24
fungii guess laravel has decided they don't support any ubuntu lts other than the latest one at this point? bionic (18.04 lts) seems ot have php 7.2 but we currently have problems using puppet on it and are looking at solutions for deploying containerized services on bionic as a result14:26
fungismarcet: given that ppa:ondrej/php is maintained by one of the official ubuntu php package maintainers, it seems like a safe enough compromise14:27
fungii guess this is his alternative to getting the php7.2 packages into xenial-backports14:28
*** quiquell|lunch is now known as quiquell14:30
openstackgerritsean mooney proposed openstack-infra/elastic-recheck master: add query for os-vif pyroute2 open files  https://review.openstack.org/62441214:30
smarcetfungi: ok cool, if its ok, then i will update puppet to work on that way, may i ask to remove openstackid production server from puppet agent ? so i could test dev server ?14:31
*** rossella_s has joined #openstack-infra14:31
*** udesale has joined #openstack-infra14:33
fungi#status log added openstackid.org to the emergency disable list while smarcet tests out php7.2 on openstackid-dev.openstack.org14:34
openstackstatusfungi: finished logging14:34
fungismarcet: i see that we're still running ubuntu trusty (14.04 lts) on both of those servers too14:35
fungimaybe this is an opportunity to rebuild them on xenial (16.04 lts) too?14:35
*** smarcet has quit IRC14:36
*** ykarel|afk is now known as ykarel14:37
*** smarcet has joined #openstack-infra14:38
smarcetfungi: yes of course14:38
smarceti will test that and we could try first on dev server :)14:38
smarcetthx u14:38
*** rossella_s has quit IRC14:39
*** e0ne has joined #openstack-infra14:45
*** rossella_s has joined #openstack-infra14:46
*** gfidente has quit IRC14:59
*** eharney has joined #openstack-infra15:00
*** markvoelker has joined #openstack-infra15:00
openstackgerritFabien Boucher proposed openstack-infra/zuul master: WIP - Pagure driver  https://review.openstack.org/60440415:05
*** psachin has quit IRC15:05
*** smarcet has quit IRC15:09
*** oanson has joined #openstack-infra15:17
*** smarcet has joined #openstack-infra15:20
*** eharney_ has joined #openstack-infra15:23
*** agopi has quit IRC15:24
*** eharney has quit IRC15:26
*** eharney_ is now known as eharney15:27
*** agopi has joined #openstack-infra15:29
*** geguileo has joined #openstack-infra15:31
geguileodmsimard: hi, I'm trying to run this playbook https://review.openstack.org/#/c/620671/7/playbooks/cinderlib/run.yaml15:32
geguileodmsimard: and it's being called from here https://review.openstack.org/#/c/620671/7/playbooks/legacy/cinder-tempest-dsvm-lvm-lio-barbican/run.yaml15:32
*** bobh has quit IRC15:32
geguileodmsimard: and I'm running into this error http://logs.openstack.org/71/620671/7/check/cinder-tempest-dsvm-lvm-lio-barbican/6de7951/job-output.txt.gz#_2018-12-04_19_52_26_75396915:32
geguileodmsimard: which is a little opaque for me15:33
dmsimardgeguileo: there's a bit more info in the ara report: http://logs.openstack.org/71/620671/7/check/cinder-tempest-dsvm-lvm-lio-barbican/6de7951/ara-report/result/abc1dc34-2d56-43e9-9c11-730cf6ec8d1d/15:33
dmsimard(from http://logs.openstack.org/71/620671/7/check/cinder-tempest-dsvm-lvm-lio-barbican/6de7951/ara-report/ )15:33
geguileodmsimard: thanks!15:34
dmsimarddoes that directory exist or not ? there's the notion of sudoers in your playbook -- do the tests need to run with superuser privileges ?15:34
geguileodmsimard: how can I know where devstack is installed? r:-??15:34
*** agopi has quit IRC15:35
dmsimardgeguileo: the devstack installation occurs in a previous task: http://logs.openstack.org/71/620671/7/check/cinder-tempest-dsvm-lvm-lio-barbican/6de7951/ara-report/result/b1365e39-3d97-48e5-a474-e65e50aba1ff/15:36
dmsimardI'm not super familiar with devstack but it looks like there's stuff in /opt/stack for sure15:37
*** bobh has joined #openstack-infra15:38
*** ykarel is now known as ykarel|away15:38
geguileodmsimard: thanks15:39
geguileodmsimard: I'll try to figure out if there's a variable with the directory15:39
*** bobh has quit IRC15:40
*** neilsun has quit IRC15:41
openstackgerritChris Dent proposed openstack-infra/project-config master: Change os-resource-classes and os-traits acl config to placement  https://review.openstack.org/62438715:47
*** gfidente has joined #openstack-infra15:52
*** wolverineav has joined #openstack-infra15:54
*** markvoelker has quit IRC15:55
clarkbour inap images are all up to date now'15:56
*** markvoelker has joined #openstack-infra15:56
*** bobh has joined #openstack-infra15:57
*** tpsilva has joined #openstack-infra15:58
*** smarcet has quit IRC15:58
*** wolverineav has quit IRC15:58
*** smarcet has joined #openstack-infra15:59
*** markvoelker has quit IRC16:01
*** bobh has quit IRC16:01
*** ccamacho has quit IRC16:09
*** jamesmcarthur has joined #openstack-infra16:10
*** udesale has quit IRC16:14
*** bobh has joined #openstack-infra16:20
*** bhavikdbavishi has joined #openstack-infra16:24
*** bhavikdbavishi has quit IRC16:25
*** bhavikdbavishi has joined #openstack-infra16:31
*** e0ne has quit IRC16:36
*** sean-k-mooney has quit IRC16:43
*** quiquell is now known as quiquell|off16:48
*** sean-k-mooney has joined #openstack-infra16:49
*** eharney has quit IRC16:51
*** d0ugal has quit IRC16:56
*** bhavikdbavishi1 has joined #openstack-infra16:58
*** kjackal has quit IRC16:59
*** kjackal has joined #openstack-infra17:00
*** bhavikdbavishi1 has quit IRC17:00
*** bhavikdbavishi has quit IRC17:02
*** bhavikdbavishi has joined #openstack-infra17:05
*** rossella_s has quit IRC17:07
*** eharney has joined #openstack-infra17:07
*** jamesmcarthur has quit IRC17:13
*** yamamoto has quit IRC17:19
*** zul has quit IRC17:20
*** gyee has joined #openstack-infra17:22
*** jpich has quit IRC17:24
clarkbA lot of email to get through this morning. Probably a fairly slow start for me today between that and our meeting17:35
*** ykarel|away has quit IRC17:35
*** pgaxatte has quit IRC17:37
fungimordred: corvus: clarkb: jpmaxman is hacking on a gerrit backend driver for netlify cms and interested in having a repo in our gerrit for some test content. any concerns?17:43
clarkbfungi: could possibly reuse the sandbox repo? (though that might get abused). I don't see any issues with having a test repo17:43
fungiyeah, i figure it might be cleaner to use a dedicated repo and then just retire it once no longer needed (or keep it around for similar future sorts of netlify backend testing). i think he wants17:44
fungii think he wants to be able to test-drive it with zuul doing gating of content changes and stuff17:45
fungiwhich is why i didn't suggest just using the official gerrit container to test with17:45
*** JpMaxMan has joined #openstack-infra17:45
corvusfungi: no objection here17:46
corvusand also, now that i've read all the requirements -- no better ideas :)17:46
*** xarses has joined #openstack-infra17:46
*** sshnaidm is now known as sshnaidm|afk17:47
fungiand exciting as this may mean easier collaboration on site content for zuul-ci.org and opendev.org17:47
*** xarses has quit IRC17:47
JpMaxManyes that's the dream :)17:48
*** xarses has joined #openstack-infra17:48
fungiclarkb: should it just go in the openstack-infra namespace? seems more related to infra/opendev efforts than to openstack anyway, even if it's not something that would necessarily be an official deliverable repo of the infra team17:49
clarkbthat is fine with me.17:50
JpMaxManright now we're just working up a POC using the starlingx site as it is already in netlify17:50
JpMaxManhttps://github.com/StarlingXWeb/starlingx-website17:52
fungiJpMaxMan: want me to get the project-config change going to create the repository? do you want starlingx-website imported as the initial repository content?17:53
*** lpetrut has quit IRC17:56
JpMaxManI'm happy to take a stab at it - and yes we'd start with the starlingx-website as an initial repo.17:56
*** bobh has quit IRC17:56
*** gyee has quit IRC17:56
fungiJpMaxMan: in that case we have instructions at https://docs.openstack.org/infra/manual/creators.html and are happy to help answer any questions you have17:57
fungiJpMaxMan: i recommend something like openstack-infra/netlify-sandbox to fit with existing naming conventions for other repos in our gerrit17:57
JpMaxManexcellent!  Thank you - will let you know as I proceed.  And yes, any naming conventions suggestions welcome - will use that to start :)17:58
funginote that a lot of what's in there isn't relevant for this particular case so you'll end up skipping some of it (e.g., anything having to do with pypi)17:59
fungiand if you miss something or include something unnecessary, that's why we have automated checks and reviewers18:00
*** derekh has quit IRC18:00
*** aojea has quit IRC18:01
clarkbunrelated but it is really cool that university researchers are starting to figure out we've got all this real world data freely available for research on software development process activity18:01
*** trown is now known as trown|lunch18:02
JpMaxManok good to know - I'll give the automation a run for its money :P18:02
fungiwe all do18:02
fungiclarkb: yes, i love that academic research sees our work as a gold mine of behavioral (both human and systems) data18:06
*** dtantsur is now known as dtantsur|afk18:07
clarkbmwhahaha: ssbarnea|rover EmilienM I'm still in a try to better understand what the afilures are are experiencing are state and looking at http://logs.openstack.org/22/605722/2/gate/tripleo-ci-centos-7-undercloud-containers/d1a7140/logs/undercloud/ I see the undercloud failed due to configuring keepalived? Having a hard time seeing why/where keepalived failed. Can you help me find the appropriate18:07
clarkblogs?18:07
mwhahahaclarkb: error mounting image volumes: unable to find user root: no matching entries in passwd file18:07
mwhahahais a bug in podman (probably runc)18:08
clarkbmwhahaha: which logfile do I look in for that?18:08
mwhahahahttp://logs.openstack.org/22/605722/2/gate/tripleo-ci-centos-7-undercloud-containers/d1a7140/logs/undercloud/home/zuul/undercloud_install.log.txt.gz#_2018-12-11_17_23_1818:09
mwhahahahttps://bugs.launchpad.net/tripleo/+bug/180354418:09
openstackLaunchpad bug 1803544 in tripleo "unable to find user root: no matching entries in passwd file" [High,Triaged]18:09
*** Swami has joined #openstack-infra18:09
clarkbaha I needed to scroll up for more ERROR messages. Thank you18:09
mwhahahahttp://status.openstack.org/elastic-recheck/index.html#180354418:10
*** e0ne has joined #openstack-infra18:10
mwhahahawe're trying to figure it out, it's one of those really obscure bugs18:10
* mwhahaha wanders off18:10
clarkbcool so its being tracked already. Thanks18:10
EmilienMclarkb: hi, yes I'm working with the podman team today and we have a fix already : https://github.com/containers/libpod/pull/197818:11
EmilienMclarkb: I'm working on getting the fix merged and built asap...18:12
*** gfidente has quit IRC18:12
clarkbEmilienM: good to know. FWIW not singling out this specific bug I was jsut going through and trying to find the breadcrumbs and got lost. Thank you for pointing me at the other error messages and the bug and the fix18:12
clarkb(this is me trying to better understand the variety of testing we run so that the infra team can help debug and/or fix things when it is on our end)18:12
EmilienMyeah it makes sense18:13
clarkbmriedem: http://logs.openstack.org/76/582376/8/gate/tempest-full-py3/a8f62b6/job-output.txt.gz#_2018-12-11_10_50_01_185172 is that one you recognize? looks like either the test node ran out of disk or the devstack test flavor is too small for cirros18:16
clarkbunfortunately dstat doesn't capture disk usage18:17
clarkbhttp://logs.openstack.org/76/582376/8/gate/tempest-full-py3/a8f62b6/controller/logs/df.txt.gz whenever that df is run by devstack indicates we have a lot of disk there though18:18
fungiclarkb: could also be bubbling up from lack of disk space at the hypervisor layer, though that build was in inap-mtl01 which isn't somewhere we've seen disk issues like that in the past as far as i'm aware18:18
clarkbfungi: ya the df shows we haev 150GB disk which is a lot more than we promise to have18:18
clarkbmaybe someone can boot the cirros image and check how much disk it ends up using (or is that something we can ask qemu-img)18:19
clarkboh you mean the hypervisor in inap, thats a good point18:19
*** wolverineav has joined #openstack-infra18:19
clarkbsorry misread it as the test node being cirros' hypervisor18:19
fungiyeah, i can see now how that might have been vague on my part18:19
fungithe provider's hypervisor layer/compute host18:19
funginot devstack's hypervisor layer18:20
clarkbyup18:20
fungii think enospc gets plumbed up into the guest anyway18:20
clarkblet us see what logstash says. If its inap specific then ya probably full up hypervisor. If we have more occurences across clouds then maybe cirros is too big18:20
fungigood thinkin18:20
*** _alastor_ has joined #openstack-infra18:22
*** d0ugal has joined #openstack-infra18:22
clarkbthere is a blip of it in inap on the 11th. Then a smaller blip in rax-iad18:23
clarkbthough I'm only searching recent days /me expands search18:23
*** wolverineav has quit IRC18:24
clarkbit happens in rax-iad, ord, inap and ovh gra118:24
clarkbinap is about 2/3 of the occurences and rax ord half that18:24
clarkbmgagne_: ^ if it is easy for you to check, any idea what disk pressure looks like on those hypervisors? Also thank you for the image upload fix. Our images are up to date now18:25
*** wolverineav has joined #openstack-infra18:26
*** wolverineav has quit IRC18:27
*** wolverineav has joined #openstack-infra18:27
*** rkukura_ has joined #openstack-infra18:32
*** rkukura has quit IRC18:32
*** rkukura_ is now known as rkukura18:32
mgagne_clarkb: didn't check all hypervisors but disk is far from being full. and now going into a meeting.18:33
openstackgerritClark Boylan proposed openstack-infra/elastic-recheck master: Add query for bug 1808010  https://review.openstack.org/62445818:35
openstackbug 1808010 in OpenStack-Gate "Tempest cirros boots fail due to lack of disk space" [Undecided,New] https://launchpad.net/bugs/180801018:35
clarkbmgagne_: thanks18:35
clarkbstarted tracking it ^ there18:35
clarkbmriedem: ^ fyi18:35
*** trown|lunch is now known as trown18:36
*** rfolco is now known as rfolco_brb18:38
fungiwe're now down to 15 zuul mergers, and the merger queue seems to be getting backed up more often (though still clears fairly quickly)18:40
*** jpena is now known as jpena|off18:40
clarkbwe expect 20 right? 12 executors + 8 dedicated mergers18:42
fungiwe don't seem to register mergers distinctly in gearman, they just show up in the merger:merge, merger:refstate, merger:fileschanges and merger:cat buckets so hard to tell which ones are missing18:42
fungiyeah, should be 2018:42
mriedemclarkb: ack,18:43
mriedemnote that until https://review.openstack.org/#/c/619319/18:43
mriedemthe flavors used by tempest via devstack specifiy 0 root_gb,18:43
mriedemmeaning compute uses whatever is the size of the image18:43
*** Adri2000 has quit IRC18:43
clarkbmriedem: possible that the image is too small for some of the writes then? you'd expect that to be more consistent though so maby does point to test node or host hypervisor18:44
fungiit was 20 mergers registered for just a split second back on the 6th/7th (when we brought ze12 into production right after restarting everything): http://grafana.openstack.org/d/T6vSHcSik/zuul-status?panelId=30&fullscreen&orgId=1&from=now-7d&to=now18:44
fungilooks like we were already down 2 before the restarts, so probably been going on for a while18:44
mriedemclarkb: hmm, maybe, not sure what size the config drive is18:45
mriedemlooks like vfat is a fixed 64MB18:46
mriedembut we don't use vfat by default18:46
fungiseems two died around utc midnight on november 13th, prior to that we were running with a full compliment since beginning of october at least, so maybe we added something in early-to-mid november which made merger threads crashy?18:47
mriedemoh wait is this config drive in the test node or a nested virt guest created by tempest?18:47
*** armax has joined #openstack-infra18:47
clarkbmriedem: this is the cirros nested "virt" guest created by tempest failing to configure networking because its disk is full (now that could be because the hypervisor running devstack is full disk or the hypervisor running the test node is also running with full disk)18:48
mriedemlooks like this is by far happening in networking-odl-tempest-fluorine18:48
clarkbin particular it appears that it can't set the default route (I'm guessing because that needs disk to write to)18:48
*** Adri2000 has joined #openstack-infra18:48
clarkband without a default route it seems that ssh is failing from tempest to the cirros node18:49
clarkb~10 minutes to the infra meeting18:49
fungi`pgrep -c zuul-executor` returns "2" on all the executors18:50
clarkbfungi: possibly the dedicated mergers have died?18:50
clarkbor maybe haven't reconnected to gearman after restarting the scheduler?18:50
fungi`pgrep -c zuul-merger` returns "1" on all the standalone mergers18:50
fungialso we seem to lose mergers one or two at a time, over time, according to the graph18:51
funginot corresponding to scheduler/geard restarts18:51
fungiwe'll likely need to dig into merger logs on the servers to find out what's going on18:52
*** smarcet has quit IRC18:52
*** bhavikdbavishi has quit IRC18:52
clarkbfungi: check netstat connections to 4730 on all of the executors and mergers?18:52
clarkband vice versa on from the gearman scheduler18:52
clarkbthat should narrow down where the connections don't exist18:52
*** bhavikdbavishi has joined #openstack-infra18:53
fungigood idea18:53
*** _alastor_ has quit IRC18:53
fungiodd that some are ipv4 and some v618:53
fungiwonder if this is network instability in rax-dfw at play18:53
clarkbI did confirm with the logstash switch to just geard that gear will fall back appropriately18:54
clarkbhttps://review.openstack.org/#/c/611920/ was another output of that to make geard a bit more ipv6 friendly18:55
clarkbmriedem: actually /run on cirros isn't necessarily a real fs either. It is possible that that is tmpfs or similar in which case it could be memroy pressure?18:56
fungiconfirmed that all executors see 2 established gearman connections and all standalone mergers 118:56
clarkbI probably need to boot a cirros image locally18:56
fungiwill check from the scheduler end now18:56
*** rlandy is now known as rlandy|brb18:59
mriedemclarkb: on one of the failures i looked at, the config drive was .5 MB18:59
*** _alastor_ has joined #openstack-infra18:59
openstackgerritMerged openstack-infra/nodepool master: Fix race in test_handler_poll_session_expired  https://review.openstack.org/62326919:00
fungiclarkb: i think that got it: http://paste.openstack.org/show/737045/ (we have 5 executors showing only 1 established gearman connection on the geard end)19:01
fungize02, ze03, ze07, ze08 and ze11 seen to have probably lost their gearman connections for their merger threads19:01
fungiShrews: this may also be up your alley to dig into once you get working internets again19:03
fungii guess we don't get separate merger logs on the executors, the messages are just all mixed into the executor logs?19:04
*** bobh has joined #openstack-infra19:07
*** jamesmcarthur has joined #openstack-infra19:09
*** wolverineav has quit IRC19:09
Shrewsfungi: the gearman stuff? maybe alley-adjacent  :)  i can help you poke around in logs after the meeting19:10
*** wolverineav has joined #openstack-infra19:10
fungiyeah, no rush19:10
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Fix node leak when skipping child jobs  https://review.openstack.org/61326119:10
*** bhavikdbavishi has quit IRC19:11
fungichecking ze02 for a start, seems it logged zuul.Merger entries other than "Updating local repository" up until 2018-11-28 23:03:52,646 and then abruptly ceased19:11
openstackgerritMerged openstack-infra/elastic-recheck master: Add query for bug 1808010  https://review.openstack.org/62445819:12
openstackbug 1808010 in OpenStack-Gate "Tempest cirros boots fail due to lack of disk space" [Undecided,New] https://launchpad.net/bugs/180801019:12
*** jamesmcarthur has quit IRC19:14
*** rlandy|brb is now known as rlandy19:14
*** yamamoto has joined #openstack-infra19:17
*** xarses has quit IRC19:19
*** _alastor_ has quit IRC19:25
*** electrofelix has quit IRC19:26
*** xarses has joined #openstack-infra19:27
*** lbragstad has quit IRC19:30
*** lbragstad has joined #openstack-infra19:31
fungidigging into the log around that time, i don't see any exceptions/tracebacks19:32
fungicurrent theory: network issues resulted in geard dropping the connection from the merger, but the merger on the executor still thinks the socket is established. lack of keepalive/dpd(?) means the merger thread is humming along blissfully unaware that it will never see any new requests19:35
*** shardy has quit IRC19:37
*** wolverineav has quit IRC19:37
fungirelated question: why is this only affecting the tag-along merger threads on the executors and not the stand-alone merger daemons?19:40
fungiwe've lost 25% of our mergers, and none are stand-alone even though those account for for 40% of the total19:40
fungistatistically unlikely it's random distribution there19:41
*** jamesmcarthur has joined #openstack-infra19:43
*** wolverineav has joined #openstack-infra19:46
*** smarcet has joined #openstack-infra19:47
*** jamesmcarthur has quit IRC19:47
*** wolverineav has quit IRC19:48
*** wolverineav has joined #openstack-infra19:49
tobiashfungi: related to your theory: https://review.openstack.org/59956719:52
tobiashfungi: we observed the same after a vm crash hosting the scheduler/geard19:52
fungitobiash: thanks!!! that's indeed interesting19:53
fungitobiash: any idea why it might affect the merger threads on our executors but not affect our stand-alone mergers?19:53
tobiashfungi: that's just co-incidence, on our scheduler crash it affected *all* mergers19:54
fungigot it, thanks again19:54
tobiashfungi: the point is that if a merge was in progress while having network issues, the merger will try to send the result and notice that the connection is broken while an idle merger won't notice it19:54
fungimakes sense. perhaps our stand-alone mergers are more active than our tag-along mergers19:55
fungiand so statistically more likely to be in the middle of something when the disconnect occurs, so notice and reconnect19:56
fungino idea if our data backs that up, but one possible explanation anyway19:56
tobiashmaybe19:56
openstackgerritChris Dent proposed openstack-infra/project-config master: Change os-resource-classes and os-traits acl config to placement  https://review.openstack.org/62438719:57
*** wolverineav has quit IRC19:58
corvustobiash, fungi: don't we have keepalives on the server?  shouldn't that be enough?19:59
tobiashcorvus: no, because an idle merger won't notice until it tries to send something19:59
fungicorvus: if we do, then i'm indeed curious why it's not helping19:59
tobiashso you need keepalive in both directions19:59
corvustobiash: oh, i get it.  thanks :)20:00
tobiashcorvus: the server correctly notices that the client is gone, so that's fine20:00
fungiwe definitely seem to have connections which are marked as established on the client but absent on the server20:00
clarkbcorvus: ianw re https://review.openstack.org/#/c/605585/14 I left a comment on what I think is the issue and how to fix it. Do you think that fix is reasonable? if so I can get it up pretty quickly20:00
fungitobiash: yep, that's i think what we're seeing here then20:00
clarkboh wait there is another issue too20:01
corvustobiash, fungi: +320:01
fungithanks!20:01
tobiashcorvus, fungi: the according zuul change is 599567 (which needs an update to the requirements after a geard release)20:01
tobiashcorvus: thanks :)20:01
corvusclarkb: yep; i think you or i may have suggested that originally too20:02
tobiasher 59957320:02
fungii'm just glad this is probably explained (and even known) and i can hopefully stop worrying about the cause now ;)20:02
clarkbcorvus: just left a second comment on another failure20:02
clarkbcorvus: this one will need a little more thought but I think we can safely converge that rule across our control plane20:02
corvusclarkb: that's very amusing, btw -- this was my yesterday: https://review.openstack.org/61964320:03
clarkbha20:04
ianwclarkb: hrm, FORWARD DROP seems safter anyway?20:04
clarkbianw: ya I think FORWARD DROP is currently the more correct rule for how we use our nodes20:04
corvusclarkb: but i agree that -- at least until we're running our own kubernetes clusters on top of our normal infrastructure, that should be fine20:05
*** zul has joined #openstack-infra20:05
clarkbits possible that kubernetes if we switch to it will change that as corvus has found (docker wants it set to DROP as well then it very carefully punches holes for what it passes through NAT)20:05
clarkbsince we'll docker with host network namespace its a noop for our docker20:05
corvusclarkb: re https://review.openstack.org/624246 maybe we should just do it in project-config?20:09
clarkbcorvus: ya we could add stub projects for the tripleo repos20:09
clarkband project config is listed first so will win right?20:09
corvusyep20:09
* corvus lunches20:10
openstackgerritMerged openstack-infra/gear master: Add support for keepalive to client  https://review.openstack.org/59956720:11
clarkbmriedem: that cirros instance seems to boot with 64MB of ram according to http://logs.openstack.org/76/582376/8/gate/tempest-full-py3/a8f62b6/controller/logs/libvirt/qemu/instance-00000022_log.txt.gz20:16
clarkb(I think I mapped the instance id properly from the console log)20:17
*** smarcet has quit IRC20:17
mriedemis that what -m 64 is?20:17
clarkbwhcih seems to be the m1.nano flavor. I'm going to boot cirros here with 64MB memory and see if its unhappy20:18
clarkbmriedem: ya20:18
fungii would rather plan n6420:18
fungier, play n6420:18
mriedemonly if not bond20:18
mriedemclarkb: yeah http://logs.openstack.org/76/582376/8/gate/tempest-full-py3/a8f62b6/controller/logs/devstacklog.txt.gz#_2018-12-11_10_39_47_72120:18
*** yamamoto has quit IRC20:19
mriedemi *think* i might have gotten to the bottom of this multiattach swap volume multinode race bug...20:19
mriedemoh it would be so sweet20:19
*** tobiash has quit IRC20:19
openstackgerritIan Wienand proposed openstack-infra/system-config master: Add a script to generate the static inventory  https://review.openstack.org/62296420:20
clarkbnope I'm not going to boot cirros locally because apparmor says libvirtd is not allowed to start20:20
fungiit knows best20:21
ianwclarkb: ^^ the inventory script was a little too bare-bones i think, suggested updates20:21
clarkbanyone have a quick easy way to boot https://download.cirros-cloud.net/0.3.5/cirros-0.3.5-x86_64-disk.img locally under qemu/kvm with 64MB memory to see if /run is a tmpfs or similar?20:21
clarkbI want to rule out that the low memory environment is itself the source of the cp errors20:21
*** mriedem has quit IRC20:22
*** tobiash has joined #openstack-infra20:23
*** bobh has quit IRC20:23
openstackgerritJean-Philippe Evrard proposed openstack-infra/zuul-jobs master: Add docker insecure registries feature  https://review.openstack.org/62448420:23
clarkbI'm going to find lunch then maybe when I get back I'll figureo ut apparmor20:23
*** mriedem has joined #openstack-infra20:24
fungiclarkb: also board meeting at 2100z if you are interested in dialling in20:24
clarkbya Ill have that in the background likely20:25
*** bobh has joined #openstack-infra20:26
openstackgerritJean-Philippe Evrard proposed openstack-infra/zuul-jobs master: Add docker insecure registries feature  https://review.openstack.org/62448420:26
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Use gearman client keepalive  https://review.openstack.org/59957320:32
*** eharney has quit IRC20:34
*** wolverineav has joined #openstack-infra20:36
*** wolverineav has quit IRC20:36
*** wolverineav has joined #openstack-infra20:36
fricklerclarkb: tmpfs on /run type tmpfs (rw,nosuid,relatime,size=200k,mode=755)20:46
fricklerclarkb: so that is bound to fail if the config drive contains > 200k data20:47
clarkbfrickler: thanks I think that means maybe 64MB isnt big enough20:47
clarkbya20:47
frickler64MB is pretty huge compared to that20:47
clarkboh ya 200k20:47
clarkbmriedem: ^ fyi20:48
*** hamerins has joined #openstack-infra20:48
*** d0ugal has quit IRC20:48
*** eharney has joined #openstack-infra20:49
fungiwhy would we create the tempfs in /run anyway? that's supposed to just be for things like pidfiles during early boot20:49
mriedemclarkb: but the 64MB here http://logs.openstack.org/76/582376/8/gate/tempest-full-py3/a8f62b6/controller/logs/libvirt/qemu/instance-00000022_log.txt.gz is the root disk, not the config drive20:49
fungier, i mean create the configdrive in /run20:49
clarkbmriedem: its ram memory, but that may be orthogonal if the tmpfs is that small20:50
mriedemoh right, was thinking root disk, nvm20:50
clarkb200kb tmpfs us pretty tiny20:50
fungiwe really should never use /run for *anything*20:50
clarkbfungi: thats likely cirros/smoser20:51
fricklerit's the cirros init script that uses it20:51
clarkbsince it doesnt run glean or cloud init it does its own thing20:51
fungiit's for pidfiles and fifos for services starting before /var/run is available20:51
clarkbdoes the 4.0 image change that U wonder20:52
fungiand if you want a reasonable-sized tmpfs for data you generally mount one yourself (like on /tmp)20:52
clarkbcould be a reasoon ti switch if so20:52
clarkbfrickler: ^ maybe you can check the newer 4.0 image too?20:52
*** fuentess has joined #openstack-infra20:52
fricklercirros 0.4 doesn't work in devstack last I checked, so not a short term option20:54
clarkbah20:54
fricklerI'm more wondering why the config-drive gets so large20:54
clarkbmriedem may know20:56
clarkbwe did add a debugging script in tempest as user data20:56
clarkbits not huge but could contribute maybe20:56
*** bobh has quit IRC20:56
clarkbalso the reason not setting the route matters is we ssh via the fip20:57
clarkbso it isnt shared l2 fom cirros perspective20:57
mriedemclarkb: not sure, wondering if something changed in tempest recently20:57
*** rfolco_brb has quit IRC20:57
fricklerclarkb: oh, where was that script added? "df -h /run" gives me 92% used, 16k free, so not much headroom there21:00
clarkbfrickler: its in tempest itself for the heavyweight ssh tests. was added to dumo debug info to console21:00
clarkbi forget where exactly I added it though but its my name on the commit if that helps to find it (eating lunch and listening to board meeting now)21:01
fricklerI found a patch from 2017, so that by itself wouldn't explain any recent breakage21:02
clarkbit may no longerbe helpful and we could rmeove it if it helps21:02
clarkbya it wasnt super re ent21:03
clarkb*recent21:03
fricklerhmm, that only looks to be three lines of script. removing it may help a bit, but if thing are really so tight I think we need some more general measures21:04
frickleranyway, eod for me, will followup tomorrow21:05
clarkb++21:06
clarkbthank you for getting that booted21:06
*** d0ugal has joined #openstack-infra21:06
mriedemclarkb: looks like we need https://review.openstack.org/#/c/623597/ on stable/rocky21:06
mriedembecause grenade on master is failing21:06
mriedemif you want to cherry pick21:07
clarkbI'll look after lunch21:09
clarkbhave a link to failure?21:09
mriedemlogstash still shows it hitting21:10
mriedemin grenade jobs21:10
mriedemso it's probably devstack in stable/rocky21:10
clarkbah21:10
*** bobh has joined #openstack-infra21:18
*** bobh has quit IRC21:19
*** bobh has joined #openstack-infra21:19
*** bobh has quit IRC21:21
*** auristor has quit IRC21:22
openstackgerritMarcH proposed openstack-infra/git-review master: tox.ini: add passenv = http_proxy https_proxy # _JAVA_OPTIONS  https://review.openstack.org/62449621:28
*** kgiusti has left #openstack-infra21:28
JpMaxManHey random question - I'm helping someone get their git review for gerrit going - should this be 404'ing ?  https://git.openstack.org/tools/hooks/commit-msg it's causing an error in the git review.21:30
openstackgerritMarcH proposed openstack-infra/git-review master: tox.ini: add passenv = http_proxy https_proxy # _JAVA_OPTIONS  https://review.openstack.org/62449621:30
clarkbyes that should be served by review.openstack.org21:30
clarkbwhat is your .gitreview file gerrit server value set to?21:31
clarkbJpMaxMan: ^21:31
JpMaxManlemme see21:31
fungiJpMaxMan: when you run, e.g., `git review -s` it should just work. if this is in an empty repository you may need to create a .gitreview file to commit to it21:31
JpMaxManI was having him follow the instructions for the sandbox21:32
JpMaxManhttps://docs.openstack.org/infra/manual/sandbox.html21:32
*** auristor has joined #openstack-infra21:33
fungihttps://git.openstack.org/cgit/openstack-dev/sandbox/tree/.gitreview#n2 looks correct21:33
corvusJpMaxMan: we can get more debug info by running "git review -s -v" and copy/pasting the output to http://paste.openstack.org/21:33
JpMaxManyeah checked the .gitreview it looks right21:34
JpMaxManhost=review.openstack.org21:34
fungiyes, i wonder if something is going sideways/getting guessed wrong due to a problem with a gerrit account21:34
fungiso the verbose output will help21:34
*** eernst has joined #openstack-infra21:34
JpMaxManhttp://paste.openstack.org/show/737053/21:36
JpMaxManhmmm it seems to work if I clone... git clone https://review.openstack.org/openstack-dev/sandbox.git21:37
JpMaxManreview instead of git ...21:38
corvusfungi, JpMaxMan: the first two lines of the debug output are interesting -- apparently gitreview.remote is set21:38
fungicould be set in ~/.gitconfig already?21:39
JpMaxManoh yes sorry I think I did that in my first bit of troubleshooting - it was complaining that there wasn't an remote named gerrit21:39
JpMaxManI looked and the remote was set to origin so I set that21:39
corvusJpMaxMan: where did you set that?21:39
fungiaha, yes if there is already a git remote named "gerrit" then git-review will assume that's what it should use to reach the gerrit server21:40
corvusfungi: JpMaxMan said the opposite of that21:40
JpMaxMangit config --global gitreview.remote origin21:41
fungioh, yep21:41
JpMaxManI first tried renaming the remote to gerrit which produced the same output21:41
corvusJpMaxMan: can you run "git config --global --unset gitreview.remote" please?  and then run 'git review -s -v' and paste the new output?21:41
JpMaxMansure21:42
*** eharney has quit IRC21:42
fungigit review should normally set a git remote named "gerrit" for you based on the content of the .gitreview file and the account name it attempts to determine via a test connection. if something goes wrong with the connection test that's when i've seen users start trying random things21:43
fungiin the future we might want to revisit how it performs username determination21:43
*** jamesmcarthur has joined #openstack-infra21:44
JpMaxManok I think I see what happened one second21:44
*** markvoelker has joined #openstack-infra21:45
*** e0ne has quit IRC21:47
*** eernst has quit IRC21:47
JpMaxManOk - so the initial error was caused by a bad username: "We don't know where your gerrit is. Please manually create a remote named 'gerrit' and try again."21:47
*** jamesmcarthur_ has joined #openstack-infra21:48
JpMaxManand yes @corvus - thank you - unsetting that did fix the issue21:48
JpMaxManbut using the correct username ;)21:48
JpMaxManhe had originally put in email instead of username and I didn't notice21:48
corvusJpMaxMan: aha!  glad it worked :)21:48
clarkbmriedem: remote:   https://review.openstack.org/624499 Set apache proxy-initial-not-pooled env var21:48
*** markvoelker has quit IRC21:49
JpMaxManmakes sense now - appreciate it21:49
*** jamesmcarthur has quit IRC21:50
openstackgerritClark Boylan proposed openstack-infra/system-config master: Import install-docker role  https://review.openstack.org/60558521:54
openstackgerritClark Boylan proposed openstack-infra/system-config master: Set iptables forward drop by default  https://review.openstack.org/62450121:54
*** wolverineav has quit IRC21:54
clarkbcorvus: ianw mordred ^ thats the outcome of the iptables discussion from a bit earlier21:54
*** wolverineav has joined #openstack-infra21:55
*** wolverineav has quit IRC21:55
*** wolverineav has joined #openstack-infra21:55
clarkbjungleboyj: any idea why cinder + lower constraints tests seem to be unhappy fairly often?21:57
jungleboyjclarkb:  No idea.  I was wondering that too.21:58
clarkbjungleboyj: http://logs.openstack.org/42/600442/1/gate/openstack-tox-lower-constraints/6592c5d/job-output.txt.gz#_2018-12-11_21_48_46_655602 seems related to database migrations?21:58
clarkbbut it isn't the old "disk is slow" timeout error. Instead this seems to complain about data types21:59
*** rcernin has joined #openstack-infra21:59
jungleboyjJeez.  I haven't seen that test case fail in a long time.22:00
*** smarcet has joined #openstack-infra22:00
jungleboyjIt is strange that that would be seen more in the LowerConstraints test.22:01
*** hamerins has quit IRC22:01
*** bobh has joined #openstack-infra22:03
fungiif you haven't seen it in a while and it's failing with older versions of deps...22:03
jungleboyj:-)  Yeah.22:03
*** trown is now known as trown|outtypewww22:04
clarkbI've updated https://bugs.launchpad.net/openstack-gate/+bug/1808010 to indicate I think its an interaction with cirros tmpfs and not a cloud issue22:04
openstackLaunchpad bug 1808010 in OpenStack-Gate "Tempest cirros boots fail due to lack of disk space" [Undecided,New]22:04
ianwclarkb:  ++ thanks.  i like it when a change gets like 3 authors ... shows the system is working :)22:04
clarkbianw: I think we are all invested in getting this going :)22:05
ianwclarkb: hrm, this isn't related to a recent change we made calculating tempest disk size?  not sure if that merged ...22:05
clarkbianw: it hasn't mriedem linked to it and its unmerged. But also cirros mounts /run as tmpfs so its actually in memory22:05
clarkbianw: and its only 200kb according to frickler's testing22:05
ianwah, ok, should read the bug22:06
fungiwell, /run is pretty ubiquitously mounted tmpfs by all distros22:06
clarkbfungi: ya thats why it occurred to me it may not be disk when I saw it was /run that had a problem22:06
fungithey don't generally even create a /run directory unless it's going to be used for pre-rootfs situations22:07
*** EmilienM has quit IRC22:07
fungiand it's pretty much always teensy too22:07
clarkbfungi: my hunch here is that cirros is abusing /run this way because config drive can tell you things about what goes into fstab22:08
*** EmilienM has joined #openstack-infra22:08
clarkbso its processing the config drive before it has real disk to write to because it may have to set up those real disks itself22:08
clarkbbut unfortunately it is leading to broken networking due to constraints being run up against22:08
clarkbits also not a super common error. So we may not want to spend too many cycles on it while debugging more common ones first. It is being tracked by e-r now so we should see if it persists or gets worse or bubbles to the top of the list due to us fixing other stuff22:10
clarkbaccording to e-r the top four bugs seem related to timeouts and network issues22:11
clarkbthere was a spiek in those that went away that I haven't debugged because it went away. Guessing a temporary provider issue22:12
clarkbafter that is http://status.openstack.org/elastic-recheck/#1807518 which I just pushed a backport to rocky in devstack for (so hopefully those go away)22:12
clarkbthen its a long long tail of all the random things that are unreliable22:12
clarkbmwhahaha: EmilienM ssbarnea|rover it seems that the centos-ceph-luminous mirror your jobs are talkign to may be getting increasingly flaky22:18
EmilienMdamn22:18
EmilienMweshay: ^22:18
clarkbhttp://mirror.dfw.rax.openstack.org/centos/7/storage/x86_64/ceph-luminous/ is something that I think we do mirror for you22:18
mwhahahawe don't have any of those jobs in the gate anymore22:19
clarkbso may just be a matter of switching to the in region mirrors for ceph-luminous22:19
mwhahahabut yes we should check that out, not sure which we're using22:19
clarkbI'm looking at gate e-r graphs (and the logstash links for them)22:19
clarkbhttp://status.openstack.org/elastic-recheck/gate.html#1708704 specifically that one22:20
mwhahahawe're not running any jobs that should require ceph22:20
mwhahahabut will need to look22:20
clarkbmore than 50% of the failures in the gate are in the last 24 hours and they fail against centos-ceph-luminous against centos.org22:20
mwhahahahmm it's mirror.centos.org22:21
*** anteaya has joined #openstack-infra22:21
mwhahahaFailed to connect to 2607:f130:0:87::10: Network is unreachable22:21
clarkbyup22:21
mwhahahaipv6'd22:21
clarkbbut we mirror it for you locally at http://mirror.dfw.rax.openstack.org/centos/7/storage/x86_64/ceph-luminous/ (replace region specific data as necessar)22:21
mwhahahayea let me go find the config22:22
mwhahahais that build into the image maybe?22:22
mwhahahacause i'm seeing NODEPOOL_CENTOS_MIRROR referenced in quickstart22:23
clarkbzuul drops some hints as to where to find the various mirrors (nodepool did it in the past so the vars say nodepool for compat)22:23
clarkbit writes /etc/ci/mirror_info.sh iirc. Let me see if I can find that22:23
mwhahahayea we use that22:23
mwhahahaso i need to find out why that one isn't set22:23
*** markvoelker has joined #openstack-infra22:24
mwhahahaoh this is before we even get to our config22:24
mwhahahaso yea it's the repos from the image22:24
mwhahahahttp://logs.openstack.org/23/624323/1/gate/tripleo-ci-centos-7-scenario000-multinode-oooq-container-updates/98dc676/job-output.txt#_2018-12-11_21_45_14_65982522:24
clarkbhttp://git.openstack.org/cgit/openstack-infra/openstack-zuul-jobs/tree/roles/mirror-info/templates/mirror_info.sh.j222:24
*** wolverineav has quit IRC22:24
mwhahahathis is in pre-run22:24
clarkbya that should run very early in our base job22:25
mwhahaharight so the pre roles don't properly configure the mirrors22:25
mwhahahanot the tripleo stuff22:25
mwhahahawe're configuring to use the mirrors22:25
mwhahahaso this is likely the repo config of the image22:25
*** wolverineav has joined #openstack-infra22:25
clarkbthe image doesn't have that data, we apply it in the job itself22:25
clarkband our base pre run should run before your pre run does22:26
clarkbyes it is part of the base jobs defined in project-config22:26
mwhahahathe images come with /etc/yum.repos.d cofigured22:26
mwhahahawith the defaults from centos22:26
*** slaweq has quit IRC22:26
mwhahahawe're actually clearing out those configs when our code starts22:27
clarkbwhy would centos have random repos enabled by default22:27
clarkb(I've quickly grepped and project-config dib elements don't add it at least)22:27
* mwhahaha shrugs22:27
clarkbianw: ^ this may interest you22:27
clarkbI wonder if this is new with 7.622:28
mwhahahaso by default all the CentOS-* files are in the cloud iamge22:29
mwhahahahttp://logs.openstack.org/23/624323/1/check/tripleo-ci-centos-7-scenario000-multinode-oooq-container-updates/bcbde3b/logs/undercloud/etc/yum.repos.d/22:29
mwhahahawe turn them off when we run our ci code22:29
clarkbbut those failures are happening before the disabling occurs?22:30
mwhahahayes22:31
mwhahahathis is before any of the tripleo code runs22:31
mwhahahathis is just basic infra prep22:31
mwhahahato install OVS22:31
mwhahahafor the multinode setup22:32
clarkbbut why would it care about the ceph repo in that case? I guess yum has to scan all the repos to see where the most appropriate ovs package lives?22:32
mwhahahayum update tries to get all the metadata22:32
mwhahahaor yum install22:32
mwhahahaif it doesn't exist22:32
mwhahahaso it errors22:32
fungiinfra-root: is someone grooming the openstackadmin account on github right now? seeing some address removals/confirmations and just want to be sure it's one of us (i expect it's related to the discussion in our meeting but would like to be sure)22:33
clarkbfungi: ianw volunteered in the meeting today22:33
*** jamesmcarthur_ has quit IRC22:33
fungicool. ianw: i guess those are you?22:34
fungi(removed root@o.o, confirmed infra-root@o.o...)22:34
mwhahahaclarkb: so it's that role22:34
ianwfungi: yep, poking at it now22:34
mwhahahaclarkb: http://git.openstack.org/cgit/openstack-infra/zuul-jobs/tree/roles/multi-node-bridge/tasks/common.yaml#n10 probably drops that storage repo inplace22:34
*** jamesmcarthur has joined #openstack-infra22:35
fungiianw: perfect. thanks again!22:35
mwhahahaclarkb: but there is no code to swap out the mirrors22:35
mwhahahaclarkb: so it uses what is shipped from centos-release-openstack-queens22:35
*** jamesmcarthur has quit IRC22:35
clarkbmwhahaha: gotcha, fwiw http://logs.openstack.org/18/607318/1/gate/tripleo-ci-centos-7-standalone/fbbd3a3/zuul-info/ also exhibits this behavior and is a single node test22:35
mwhahahayea so it's any centos job that installs OVS22:35
*** jamesmcarthur has joined #openstack-infra22:35
clarkbmwhahaha: not sure why it would be running multinode setup if it is single node (that might be a separate cleanup)22:35
mwhahahaclarkb: we use ovs for fake interfaces22:36
*** jamesmcarthur has quit IRC22:36
mwhahahabut the issue is that the multi-node-bridge role does not properly configure mirrors to install ovs from22:36
clarkbdoes centos-release-openstack-queens imply centos-ceph-lumnious transitively?22:36
mwhahahalikely22:36
fungiclarkb: do we miss setting a mirror url for the ovs packages?22:36
mwhahahaclarkb: yes, https://rpmfind.net/linux/RPM/centos/extras/7.6.1810/x86_64/Packages/centos-release-openstack-queens-1-2.el7.centos.noarch.html22:37
fungiis that the summary?22:37
*** jamesmcarthur has joined #openstack-infra22:37
clarkbfungi: possibly? I'm not sure if we set the mirror properly for the rdo/openstack repo22:37
clarkband then ceph is an unexpected addition22:37
clarkbor if we fail to set both of them22:37
mwhahahaso the repos get added and removed in multi-node-bridge22:37
mwhahahahttp://git.openstack.org/cgit/openstack-infra/zuul-jobs/tree/roles/multi-node-bridge/tasks/common.yaml#n4522:37
clarkblooks like we don't really do much between add repo and install package22:37
mwhahahaso it adds the stock repos, installs ovs, removes the repos22:37
clarkbso likely unset for both repos22:37
*** bobh has quit IRC22:38
*** bobh has joined #openstack-infra22:38
ianwclarkb / fungi: so that turned out to be rather easy ... when you get a minute do you want to look at the password file and try logging into the shared account with 2fa token as described there?22:39
*** jamesmcarthur has quit IRC22:40
clarkbianw: ya I can try when I've paged this ovs/ceph stuff out22:40
*** jtomasek_ has quit IRC22:40
*** bobh has quit IRC22:41
*** _alastor_ has joined #openstack-infra22:41
*** jamesmcarthur has joined #openstack-infra22:42
*** slaweq has joined #openstack-infra22:44
*** bobh has joined #openstack-infra22:44
ianwclarkb: also can you take a look at stein mirroring request, seems straight foward -> https://review.openstack.org/#/c/621231/22:45
*** boden has quit IRC22:46
*** jamesmcarthur has quit IRC22:46
*** slaweq has quit IRC22:48
clarkbmwhahaha: ianw: configure-mirror role tries to do this for centos but only applies it for epel and the base os/ portion of the mirror22:51
clarkbI think it will work if we write out the file that specifies centos-ceph-luminous and disable it like we do with epel. Any idea where I can find a copy of that file?22:52
clarkbhttps://github.com/CentOS-Storage-SIG/centos-release-ceph-luminous/blob/master/CentOS-Ceph-Luminous.repo that maybe?22:54
ianwclarkb: won't the package install overwrite that?  in the epel case, we have epel-release package installed22:55
*** yamamoto has joined #openstack-infra22:55
clarkbianw: maybe? I know very little about how centos is expected to work. Its all a foreign language to me particularly the way everything is in a different repo and you have to do somethign special to install what seems like every other package22:57
clarkbianw: we configure epel with this j2 file https://git.openstack.org/cgit/openstack-infra/zuul-jobs/tree/roles/configure-mirrors/templates/etc/yum.repos.d/epel.repo.j222:58
clarkbseems like we set it to enabled=0 then expect something else to enable it. Can we write out a CentOS-Ceph-Luminous.repo file in a similar way and have the package that installs the repo flip the bit or will it overwrite entirely?22:59
ianwclarkb: i think the package will overwrite it.  for epel, we have pre-installed the package with https://git.openstack.org/cgit/openstack/diskimage-builder/tree/diskimage_builder/elements/epel23:00
clarkbas an alternative we can have multi-node-bridge role do a text substition on that file after the packages install the repo23:00
clarkbbut before we install ovs23:00
clarkboh got it23:00
ianwthe idea for epel is that you do "yum install --enablerepo=epel ..." so we know what we're dragging in explicitly23:00
clarkbin that case maybe the text substition in multi-node-bridge role is better23:00
mwhahahait's really specific to that role, so if the mirrors exist in the ansible vars then do a text substitution between the install of the repo before the package23:01
mwhahahathis is the annoying problems with the CI repo configs that we end up duplicating this same thing all over the place23:01
ianwhrm, i forget, we uninstall the repos after right?23:02
mwhahahain that role, yes23:02
mwhahahahttp://git.openstack.org/cgit/openstack-infra/zuul-jobs/tree/roles/multi-node-bridge/tasks/common.yaml#n4523:02
mwhahahait's litterally to just get the queens version of OVS23:02
ianwyeah, that's right, but if they were there we don't23:02
ianwand i think at one point we used to install RDO in the base package, but that caused problems, which is why we moved it "up" to this point23:03
clarkbmwhahaha: yes, and every other distro avoids this problem by having A repo23:03
clarkb(even fedora has everything in a single repo iirc)23:03
mwhahahapretty sure ubuntu has more than one23:03
mwhahahaUCA is the extra one23:03
mwhahahaanyway23:03
* mwhahaha then points to pypi, yum, docker, etc mirrors23:04
*** eernst has joined #openstack-infra23:04
ianwyeah, not really centos's fault because it's raison d'etre is to be rhel like, so if rhel doesn't have ovs in base then we end up like this23:04
clarkbya it just gets really complicated quickly23:05
*** bobh has quit IRC23:05
ianwwe could install rdo like epel and disable it23:06
clarkbianw: I'm working on a lineinfile patch for multi-node-bridge23:06
*** bobh has joined #openstack-infra23:06
clarkbwhich will replace the remote with the mirror node23:06
clarkb(I hope)23:07
mwhahahawe used to always have the N-1 version installed by default but i think that caused more problems23:07
mwhahahait would be nice if we got OVS from something that only contained OVS23:07
clarkbmwhahaha: ya we removed it from the image because that caused confusion too23:07
ianw++23:07
mwhahahaat this point i think just lineinfile mirror.centos.org with the local mirrors is probably the best bet23:08
ianwyes the KISS approach23:08
mwhahahathough i wonder how that plays in with the uninstall if the file is changed23:08
* mwhahaha shrugs23:08
ianwi think we almost had linuxbridge working for multinode too?  i remember that being a possibility for removing ovs23:08
ianwby "we", i mean clarkb, i didn't do anything useful :)23:09
clarkbianw: neutron assumed ovs unfortunately23:10
clarkbso it got tricky to untable the unfortaunte dep from devstack + neutron that that ovs bridge existing23:11
clarkband I gave up23:11
clarkbits entirely doable at this point if we get devstack + neutron to learn how to plug the linux bridge bridge into its own ovs bridges23:11
clarkbanyone know where to get a copy of /etc/yum.repos.d/CentOS-OpenStack-queens.repo ?23:12
openstackgerritJp Maxwell proposed openstack-infra/project-config master: Adding the netlify-sandbox project  https://review.openstack.org/62452323:13
openstackgerritMatt Riedemann proposed openstack-infra/elastic-recheck master: Add query for glance-api proxy error bug 1808063  https://review.openstack.org/62452423:13
openstackbug 1808063 in OpenStack-Gate "glanceclient.exc.HTTPBadGateway: 502 Proxy Error during server snapshot" [Undecided,Confirmed] https://launchpad.net/bugs/180806323:13
mriedemclarkb: ^23:13
*** slaweq has joined #openstack-infra23:14
ianwclarkb: http://paste.openstack.org/show/737099/ i think, from https://www.rdoproject.org/repos/rdo-release.rpm23:15
*** kjackal has quit IRC23:15
mwhahahahttp://mirror.centos.org/centos/7/extras/x86_64/Packages/centos-release-openstack-queens-1-2.el7.centos.noarch.rpm23:15
* mwhahaha is downloading to fetch23:16
mwhahahahttp://paste.openstack.org/show/737100/23:17
mwhahahait's more than just the rdo-release23:17
*** _alastor_ has quit IRC23:17
mwhahahaif you swap out mirror.centois.org and buildlogs.centos.org i think we have mirrors for those23:17
mwhahahathough only mirror.centos.org is the one that is enabled23:18
openstackgerritClark Boylan proposed openstack-infra/zuul-jobs master: Use mirrors if available when installing OVS on centos  https://review.openstack.org/62452523:18
clarkbya I was just doing mirror.centos.org since it si the only one enabled23:18
clarkbI think something like ^ should work23:18
mwhahahahttp://paste.openstack.org/show/737101/ is the ceph one23:19
clarkbI don't think multi-node-bridge is a trusted role so we should be able to depends on that chagne from a tripelo change to make sure it works23:19
mwhahahayea that should work23:19
*** slaweq has quit IRC23:19
clarkbmwhahaha: care to push that depends on change (I don't know what would be a good representative set)23:20
mwhahahasure23:20
clarkbthanks23:20
mwhahahahttps://review.openstack.org/#/c/624526/23:21
mwhahahawill get an assortment of jobs23:21
*** jamesmcarthur has joined #openstack-infra23:22
*** eernst has quit IRC23:25
*** jamesmcarthur has quit IRC23:26
melwittclarkb: mriedem just told me about https://bugs.launchpad.net/openstack-gate/+bug/1808010 while I was looking at a failed job run, but in the log I see "WARN: failed: route add -net "0.0.0.0/0" gw "10.1.0.1"" but not any messages about no space left. is that a separate known launchpad bug or do you think it's the same thing?23:27
openstackLaunchpad bug 1808010 in OpenStack-Gate "Tempest cirros boots fail due to lack of disk space" [Undecided,New]23:27
melwitthttp://logs.openstack.org/82/623282/3/check/nova-next/a900344/logs/testr_results.html.gz23:27
clarkbmelwitt I thought it was thr same thing23:28
*** smarcet has quit IRC23:28
melwittok, thanks23:28
clarkbmelwitt: in the bug it has messages about the disk errors23:28
clarkbhappena before failing to set the route23:29
melwittyeah, I don't see them in the cirros log excerpt on the job I was looking at (linked above) so I wasn't sure23:29
clarkbhuh maybe disk space isnt the root cause then23:29
clarkbIm pretty sure the broken default route is what breaks ssh23:29
clarkband thought it was caused by the disk issue23:29
melwittbut indeed when I search for it on logstash I see most of the hits coming from the networking-odl-tempest-fluorine job, all failures23:30
melwitt*when I search for the failed route add23:30
*** slaweq has joined #openstack-infra23:36
openstackgerritJp Maxwell proposed openstack-infra/project-config master: Adding the netlify-sandbox project  https://review.openstack.org/62452323:38
clarkbmelwitt: we probably want to better understand what could cause that route add failure23:40
clarkband go from there23:40
clarkbcirros runs busybox so it may be different than whatever distro you have locally too23:40
*** slaweq has quit IRC23:40
*** armax has quit IRC23:42
*** xek_ has joined #openstack-infra23:43
*** xek has quit IRC23:46
*** smarcet has joined #openstack-infra23:46
melwittclarkb: ack, thanks23:49
melwittI added a note to the launchpad23:49
*** dklyle has joined #openstack-infra23:51
openstackgerritIan Wienand proposed openstack-infra/system-config master: Enable github shared admin account  https://review.openstack.org/62453123:52
*** xarses has quit IRC23:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!