Tuesday, 2019-08-27

*** rlandy|ruck is now known as rlandy|ruck|bbl00:11
weshayrlandy|ruck|bbl I think that's default libvirt.. which deployment are you running? standard?00:16
*** aakarsh|2 has joined #oooq00:36
*** dsneddon has quit IRC01:46
*** dsneddon has joined #oooq01:49
*** dsneddon has quit IRC01:54
*** dsneddon has joined #oooq01:59
*** dsneddon has quit IRC02:03
*** apetrich has quit IRC02:10
*** rlandy|ruck|bbl is now known as rlandy|ruck02:29
rlandy|ruckweshay: thanks - got past that error02:29
rlandy|ruckgot the job to start02:29
rlandy|ruckwill debug more tomorrow02:29
*** rlandy|ruck has quit IRC02:29
*** dsneddon has joined #oooq02:32
*** gkadam has joined #oooq03:51
*** gkadam has quit IRC03:51
*** jtomasek has quit IRC04:00
*** rfolco has quit IRC04:03
*** udesale has joined #oooq04:06
*** aakarsh|2 has quit IRC04:06
*** ratailor has joined #oooq04:20
*** raukadah is now known as chkumar|rover04:26
*** jtomasek has joined #oooq04:44
*** dsneddon has quit IRC04:45
*** skramaja has joined #oooq04:54
*** dsneddon has joined #oooq05:07
*** dsneddon has quit IRC05:15
chkumar|roversshnaidm: please merge this https://review.opendev.org/#/c/678622/ our centos7 container build is broken05:33
chkumar|roverbecause of reverts05:34
*** sanjayu_ has joined #oooq05:42
*** dsneddon has joined #oooq05:47
*** ccamacho has quit IRC05:52
*** hamzy has quit IRC05:59
*** jfrancoa has joined #oooq06:05
*** jfrancoa has quit IRC06:09
*** dsneddon has quit IRC06:15
*** dsneddon has joined #oooq06:16
*** surpatil has joined #oooq06:22
*** jfrancoa has joined #oooq06:25
*** brault has joined #oooq06:38
*** dsneddon has quit IRC06:41
*** dsneddon has joined #oooq06:44
*** dsneddon has joined #oooq06:48
*** bogdando has joined #oooq07:14
*** dsneddon has quit IRC07:32
*** dtantsur|afk is now known as dtantsur07:37
*** jpena|off is now known as jpena07:40
*** surpatil has quit IRC07:44
sshnaidmchkumar|rover, ack07:58
*** dsneddon has joined #oooq08:01
*** panda has quit IRC08:02
*** panda has joined #oooq08:02
*** apetrich has joined #oooq08:22
*** ccamacho has joined #oooq08:47
*** derekh has joined #oooq08:50
*** surpatil has joined #oooq09:04
chemchkumar|rover: hi, first time seeing that file /usr/share/ansible/roles/tripleo-hieradata/tasks/hieradata_vars.yaml conflicts between attempted installs of openstack-tripleo-common-11.1.1-0.20190826025903.29b7c8a.el7.noarch and tripleo-ansible-0.2.1-0.20190826144854.bf61a6f.el7.noarch09:12
*** zbr has joined #oooq09:12
chkumar|roverchem: https://review.opendev.org/#/c/673366/ and https://review.opendev.org/#/c/678622/ will fix the issue09:13
chemchkumar|rover: hum ... couldn't find the associated lp. are my lp search skills bad (certainly) or is there somewhere to look ?09:14
zbrhello! i am back. i wonder what I missed as apparently my connection dropped.09:15
pandazbr: so you don't know...09:15
chkumar|roverchem: https://bugs.launchpad.net/tripleo/+bug/184140509:17
openstackLaunchpad bug 1841405 in tripleo "role 'dump_vars' not found leading to logs not getting collect in post" [Critical,Fix released] - Assigned to Kevin Carter (kevin-carter)09:17
chemchkumar|rover: oki, thanks09:20
jfrancoachkumar|rover: hello, I took the freedom to update the depends-on patch here: https://review.rdoproject.org/r/#/c/21946/09:27
jfrancoachkumar|rover: I was debugging the issue in the reproducer environment rlandy lend me and I believe this patch should fix it: https://review.opendev.org/#/c/678767/09:28
chkumar|roverjfrancoa: cool, thanks! so we need only one review09:30
*** dtantsur is now known as dtantsur|bbl09:31
jfrancoachkumar|rover: I think so. I managed to run the overcloud update run passing --ssh-user tripleo-admin, it succeeded. So hopefully this will make it09:32
chkumar|roverjfrancoa: cool.09:32
chkumar|roversshnaidm: Hello09:36
chkumar|roversshnaidm: in order to remove pike jobs, do we need to remove current-tripleo-rdo aka rdo phase 1 jobs also https://github.com/rdo-infra/ci-config/blob/master/ci-scripts/dlrnapi_promoter/config/CentOS-7/pike.ini#L31 ?09:36
sshnaidmchkumar|rover, yeah, we can remove them too from jenkins09:39
chkumar|roversshnaidm: ok, preparing the patch09:39
sshnaidmchkumar|rover, but they don't run if no tripleoci promotion, so not urgent09:39
chkumar|roverack!09:40
sshnaidmchkumar|rover, need also to remove everything from promoter about pike09:40
chkumar|roversshnaidm: https://review.rdoproject.org/r/#/c/21961/ done here09:40
sshnaidmack09:41
*** sshnaidm is now known as sshnaidm|afk09:41
*** jaosorior has quit IRC09:43
*** matbu has joined #oooq09:45
*** apetrich has quit IRC09:55
*** apetrich has joined #oooq10:00
*** pierreprinetti has joined #oooq10:06
zbrpanda: is https://review.opendev.org/#/c/673481/ ready?10:09
zbr@oooq: i am focusing on reviews today, if you have any review where you need help, please ping me here with link.10:10
jfrancoachkumar|rover: the job passed https://review.rdoproject.org/r/#/c/21946/ using https://review.opendev.org/678767, so I'll abandon https://review.opendev.org/#/c/678572/ and we'll try to merge the right fix10:51
chkumar|roverjfrancoa: cool, thanks!10:52
chkumar|roverjfrancoa++10:53
jfrancoachkumar|rover: no problem. happy to help10:53
*** udesale has quit IRC11:02
*** tesseract has joined #oooq11:11
*** hamzy has joined #oooq11:18
*** jaosorior has joined #oooq11:20
*** jpena is now known as jpena|lunch11:30
*** sanjayu_ has quit IRC11:46
zbri see more and more activity on the pro-chaing-dfg document but not from our team.11:52
*** sshnaidm|afk is now known as sshnaidm11:59
pandazbr: TL;DR12:06
*** dtantsur|bbl is now known as dtantsur12:08
*** ratailor_ has joined #oooq12:08
*** ratailor has quit IRC12:08
*** ratailor_ has quit IRC12:18
zbrpanda: can i update your https://review.opendev.org/#/c/673481/ to add the missing part from tox?12:19
*** rfolco has joined #oooq12:22
pandazbr: which missing part ?12:23
pandazbr: don't touch it right now, first I want to understand if oooo is ok with this solution12:23
zbras long they see the -1 vote from rdo, they will not be positive about it.12:24
zbri need to get in touch with cloudnull as he downvoted few changes around molecule, mainly because he used it with delegated and without tox.12:25
*** rlandy has joined #oooq12:28
*** rlandy is now known as rlandy|ruck12:29
rlandy|ruckweshay: chkumar|rover: anything we want to raise at tripleo meeting?12:29
chkumar|roverrlandy|ruck: nope12:30
chkumar|roverrlandy|ruck: just an update on pike job removal12:30
chkumar|roverwe have all the patches up for the same12:30
weshaynot as ruck/rover I think the sprint team had a few things...12:30
rlandy|ruckchkumar|rover: k - saw your jobs - will work on merge today12:30
chkumar|roverrlandy|ruck: https://etherpad.openstack.org/p/ruckroversprint14 line 2612:30
chkumar|roverrlandy|ruck: let me know If I missed anything12:30
chkumar|roverrlandy|ruck: CI is also calm due to this https://bugs.launchpad.net/tripleo/+bug/184156412:31
openstackLaunchpad bug 1841564 in tripleo "/usr/share/ansible/roles/tripleo-hieradata/tasks/hieradata_vars.yaml conflicts between attempted installs of tripleo-ansible and openstack-tripleo-common" [Critical,Confirmed]12:31
rlandy|ruckchkumar|rover: I'll check through those12:31
weshayrlandy|ruck chkumar|rover I think this job, is non-voting12:31
weshayhttp://zuul.openstack.org/builds?job_name=tripleo-ci-centos-7-scenario000-multinode-oooq-container-upgrades&branch=stable%2Fstein12:31
weshayrlandy|ruck I think it should be voting..12:31
chkumar|roverweshay: yes it is nv https://opendev.org/openstack/tripleo-ci/src/branch/master/zuul.d/multinode-jobs.yaml#L30412:32
rlandy|ruckk - can update12:32
weshaychkumar|rover that could be something to mention in the mtg12:32
*** jpena|lunch is now known as jpena12:32
weshaychkumar|rover rlandy|ruck master should be nv, n-* should be voting12:33
weshayk12:33
weshay?12:33
rlandy|ruckweshay: ack12:33
chkumar|roverhttp://zuul.openstack.org/builds?job_name=tripleo-ci-centos-7-scenario000-multinode-oooq-container-upgrades12:33
rlandy|ruckweshay: the default is non-voting12:33
chkumar|roverlooks pretty green12:33
chkumar|roverwith few yellows12:33
rlandy|ruckchkumar|rover: weshay: patch in progress12:33
weshayrlandy|ruck right.. so the upgrade jobs..  for master should be nv, n-* should vote if stable12:33
*** dsneddon has quit IRC12:36
*** jaosorior has quit IRC12:39
*** surpatil has quit IRC12:39
*** jaosorior has joined #oooq12:39
rlandy|ruckhmmm - can we put j2 in job definitions?12:46
zbrrlandy|ruck: can you please wf https://review.opendev.org/#/c/678038/ ?12:52
zbris required for s1-4 work, should have already being workflowed days ago.12:53
zbrrlandy|ruck: thanks!12:53
zbrwhat an useless UI experience on https://zuul.opendev.org/t/openstack/build/5bbd63bc123042d9a435cdf1dc06c0ef/console12:54
zbri need to look at source code to figure-out the name of the log file12:55
rlandy|ruckweshay: https://review.opendev.org/#/c/678814/ - but I would prefer to add a j2 if condition there13:00
rlandy|ruckbecause that would save us changing the name13:00
rlandy|rucknot sure if that is possible13:00
*** aakarsh|2 has joined #oooq13:01
rlandy|ruckjfrancoa: hi - could you reach the nodes held for debug?13:02
chkumar|roverrlandy|ruck: queens upgrade issue now fixed13:02
chkumar|roverrlandy|ruck: https://review.opendev.org/#/c/678767/13:02
jfrancoarlandy|ruck: yep and as chkumar|rover said, it's fixed. thanks a lot for the nodes it was very helpful13:03
*** dsneddon has joined #oooq13:03
rlandy|ruckjfrancoa: k - going to let the admins know they can reclaim them13:03
*** Goneri has joined #oooq13:05
*** yolanda has quit IRC13:06
*** yolanda__ has joined #oooq13:06
zbrrfolco: apparently for s14 work we need to get rid of fluentd, and martin is still working on it https://review.opendev.org/#/c/668851/13:06
rlandy|ruckjfrancoa: chkumar|rover: k - w+ https://review.opendev.org/#/c/678767/13:06
rfolcozbr, ok, please keep chasing scen1-413:07
rfolcozbr, my python guru, do you understand future.wait ?13:08
zbrrfolco: not that kind of guru, yet. but if you point me to some code, i may have a chance of finding more.13:09
*** jaosorior has quit IRC13:09
*** dsneddon has quit IRC13:09
rlandy|ruckchkumar|rover: re: phase 1 master fix - is it worth rerunning phase 1 master or we would need the next promotion to pick that fix up?13:09
rlandy|ruckhttps://etherpad.openstack.org/p/ruckroversprint14 - line 3813:09
rfolcozbr, I think something is wrong with this code - https://opendev.org/openstack/tripleo-common/commit/0be1be779a27d7bb3ba8f5469e391e4c72eee68513:10
chkumar|roverrlandy|ruck: the promotion job for naster is running, let's get it finish, then it will automatically pick up13:10
rfolcozbr, its raising an system exception, mistakenly13:10
rfolcozbr, why? https://3d89b2f66ce8e968c7f7-8b938dd2076b97d235f21ad4df33ebf0.ssl.cf2.rackcdn.com/678058/27/check/tripleo-build-containers-centos-7-buildah/7508cee/logs/build.log.txt.gz13:10
rlandy|ruckchkumar|rover: it will kick phase 1 only of it passes13:11
chkumar|roverrlandy|ruck: ack, thanks!13:11
rfolcozbr, all containers are pushed to local registry, but the code future.wait is still raising exception13:11
rlandy|ruckk - if not then I'll manually kick it13:11
rfolcozbr, line 174... wondering if this is correct:  return_when=futures.FIRST_EXCEPTION13:11
weshayFYI... mtgs most of the morning and afternoon :(13:12
rfolcoweshay, ack13:13
rlandy|ruckhttps://review.rdoproject.org/r/#/c/21961/2/ci-scripts/dlrnapi_promoter/dlrn-promoter.sh - pike is gone an docat ais not :)13:17
rlandy|ruckand ocata13:17
chkumar|roverrlandy|ruck: ocata is still there?13:19
rlandy|ruckchkumar|rover: weshay: so https://review.opendev.org/#/c/678154/ has to merge first before we remove our pike jobs - or, as long a the eol tag is approved, we can go ahead?13:19
rlandy|ruck"CentOS-7/queens" "CentOS-7/ocata" "RedHat-8/master"13:20
rlandy|ruckyep13:20
rlandy|ruckfine though13:20
chkumar|roverarxcruz: please include this one https://review.opendev.org/#/c/678833/ aslo as a deps on fs01 os_tempest patch13:23
weshayrlandy|ruck I approve yes13:23
chkumar|roverit will be a precheck for the same13:23
arxcruzchkumar|rover: is that the problem ?13:24
chkumar|roverarxcruz: it will check early by doing a gateway ping13:25
chkumar|rovermight be the issue, donot know13:25
weshaychkumar|rover rlandy|ruck please add some notes to https://docs.google.com/document/d/1LP23IwpCJLKCMe3wzDOvtqZZ79-eCxhxHVf7NOPeQbI/edit#heading=h.cyleyypknu4t13:25
zbrrfolco: i am looking at the code in i master and I have no idea how an empty set would endup evaluated as true!?13:26
rfolcozbr, there are some unit tests that I am starting to play with... to see if I catch the problem13:27
zbrhave a look at https://github.com/openstack/tripleo-common/blob/9145769a7af55f4402e1ce30b542f7bd5e89e5e3/tripleo_common/image/builder/buildah.py#L18813:27
zbrwhat gets printed is set([]) which on py27 means an empty set.13:28
zbrpy3 prints it as set() but that is a detail, still evaluates the same.13:28
rfolcozbr, parsing your comments.... lets chat on community mtg if time permits13:29
zbrit would not be possible to reach the SystemError if not_done would be an empty set!13:29
zbrit has nothign to do with future.wait13:29
zbrbut personally, I would replce FIRST_EXCEPTION and force it to run all tasks before returning.13:30
rfolcozbr, panda sshnaidm rlandy|ruck arxcruz: community mtg13:32
*** udesale has joined #oooq13:32
chkumar|roverweshay: rlandy|ruck updated the doc13:33
chkumar|roverfeel free to comment13:33
rlandy|ruckthanks13:33
weshaylolz... in managers mtg.... ci creating documentation is coming up13:33
weshay:)13:33
*** jaosorior has joined #oooq13:33
zbrrfolco: sshnaidm please have a look at https://review.opendev.org/#/c/678838/ -- it should fix the issue with buildah SystemError13:35
chkumar|roverrlandy|ruck: container build job passed13:37
rlandy|ruckyep13:38
*** dsneddon has joined #oooq13:42
rfolcozbr, thx for the patch :)14:02
* chkumar|rover headed home, will be back soon14:14
rlandy|ruckchkumar|rover: pls see comment on https://review.rdoproject.org/r/#/c/21961/14:16
rlandy|ruckI can add the cockpit stuff later14:16
rlandy|ruckyour patch is fine for its content14:16
*** Vorrtex has joined #oooq14:17
*** Vorrtex has quit IRC14:17
*** Vorrtex has joined #oooq14:19
*** sshnaidm_ has joined #oooq14:29
*** sshnaidm has quit IRC14:30
*** udesale has quit IRC14:32
*** sshnaidm__ has joined #oooq14:34
*** sshnaidm_ has quit IRC14:36
*** skramaja has quit IRC14:40
*** sshnaidm__ is now known as sshnaidm14:51
sshnaidmweshay, https://review.opendev.org/#/c/678630/ - which readme? which context?14:51
weshaysshnaidm https://github.com/openstack/ansible-role-collect-logs/blob/master/README.rst  don't assume everyone know what sova is, that's basically what I'm poking at14:53
weshayrlandy|ruck can you grep through the featuresets and make sure they all have podman for the appropriate releases please.. eg.  https://review.opendev.org/#/c/678244/3/config/general_config/featureset037.yml14:55
rlandy|ruckweshay: ack14:56
weshayrlandy|ruck or we can make that a default14:56
rlandy|rucklet's see what the damage is14:56
rlandy|ruckie: how the jobs are defined vs the fs14:56
*** dsneddon has quit IRC14:58
*** dsneddon has joined #oooq15:04
*** bogdando has quit IRC15:09
*** dsneddon has quit IRC15:10
rlandy|ruckweshay: will be much easier to add this setting to a common role than put it in every fs15:10
chkumar|roverrlandy|ruck: will remove rrcockpit and pike sova cleanup tomorrow15:12
*** ksambor has quit IRC15:12
rlandy|ruckchkumar|rover: it's fine - better to do it afterwards15:13
weshayrlandy|ruck agree15:13
rlandy|ruckit will remind us if we missed anything15:13
weshayrlandy|ruck so.. I would add to common, and remove from fs15:13
rlandy|ruckweshay: going to put it in common-extras15:13
chkumar|roverrlandy|ruck: weshay ack15:14
rlandy|ruckweshay: this is far from efficient : http://pastebin.test.redhat.com/79217115:14
* rlandy|ruck patches15:15
rlandy|ruck[rlandy@localhost tripleo-quickstart]$ grep -r overcloud_container_cli15:15
rlandy|ruckconfig/general_config/featureset010.yml:overcloud_container_cli: podman15:15
rlandy|ruckconfig/general_config/pacemaker.yml:overcloud_container_cli: docker15:15
rlandy|ruck^^ that is even worse15:15
*** ccamacho has quit IRC15:23
chkumar|roverrlandy|ruck: weshay see ya tomorrow,15:31
*** chkumar|rover is now known as raukadah15:31
weshayraukadah sent an email to you..15:31
weshayhit me up w/ questions if you have them15:31
raukadahweshay: checking, will look into that tobiko stuff15:31
weshayraukadah thanks15:32
raukadahpanda: sshnaidm: weshay https://review.rdoproject.org/r/#/q/topic:remove_pike+(status:open+OR+status:merged) time to say bye bye to pike15:33
pandaso soon ?15:33
raukadahkind of15:34
raukadahpanda: sshnaidm rfolco arxcruz https://review.rdoproject.org/r/#/q/topic:remove_pike+(status:open+OR+status:merged) please vote on this when free15:35
raukadahsorry this one https://lists.rdoproject.org/pipermail/dev/2019-August/009126.html15:35
raukadahzbr: can we make these check jobs to nv https://review.rdoproject.org/r/#/c/21787/15:39
raukadahcurrently15:39
*** dsneddon has joined #oooq15:39
*** jfrancoa has quit IRC15:47
*** brault has quit IRC15:50
*** sanjayu_ has joined #oooq15:51
*** chem` has joined #oooq15:51
*** chem has quit IRC15:52
*** sshnaidm is now known as sshnaidm|afk15:54
*** jpena is now known as jpena|off16:04
weshayzbr fix this please https://review.rdoproject.org/r/#/c/21787/16:15
zbrsure.16:16
weshayrlandy|ruck did you have luck w/ a repro or still need eyes?16:18
rlandy|ruckweshay: I got it to work until the multinode bridge16:19
rlandy|ruckI will try pick it up from there16:19
rlandy|ruckweshay: rdocloud never worked16:19
weshaywow.. 0 gate failures today :))16:19
rlandy|rucklogger issue16:19
weshayI'm not familiar w/ the logger issue, but saw it being discussed16:19
weshaya few days ago16:19
rlandy|ruckstarted working with libvirt to avoid that16:19
rlandy|ruckweshay: you are familiar :) you logged the initial bug16:20
weshayhrm... k16:20
weshayOH?16:20
weshaylolz16:20
weshayooh stein and master promoted too :)16:20
rlandy|ruckyesterday ack16:21
rlandy|ruckgetting them to try promote today as well16:21
rlandy|ruckwell, late last night16:21
zbrweshay: we need to wf https://review.opendev.org/678838 as soon is passing checks, (eta ~1h) -- is needed for s1416:21
weshayzbr  context w/ non-voting is so we can still get a +1 from 3rd party zuul16:22
zbroops.just seen a commen,. need to check it.16:22
weshayclosing unauth bug https://bugs.launchpad.net/tripleo/+bug/183953216:23
openstackLaunchpad bug 1839532 in tripleo "tripleo gate jobs are failing to pull containers when running on ovh provider with "UNAUTHORIZED" error" [Critical,Fix released]16:23
weshayrlandy|ruck sorry to bug you.. I don't see the a logger bug in my list16:24
weshayhelp a brotha out16:24
rlandy|ruckweshay: getting16:26
rlandy|ruckweshay: https://bugs.launchpad.net/tripleo/+bug/183346516:26
openstackLaunchpad bug 1833465 in tripleo "tripleo reproducer fails w/ "waiting on logger"" [Critical,Incomplete]16:26
rlandy|ruck^^ that killed me yesterday16:26
weshayOHHh16:29
weshaythat went away for awhile16:29
weshayI think that is a zuul bug16:29
weshaysshnaidm|afk we may need to update the repro containers16:29
rlandy|ruckit returned16:29
weshayrlandy|ruck well... maybe we just need to refresh16:29
*** ksambor has joined #oooq16:29
rlandy|ruckwith a vengeance16:29
*** ksambor has quit IRC16:30
weshayya.. it's terrible16:30
rlandy|ruckrefresh what?16:30
rlandy|ruckI pick up the shared images16:30
rlandy|ruckwhat else can be refreshed?16:30
weshayrlandy|ruck the builds of the containers themselves16:30
* weshay looks16:30
rlandy|ruckweshay: I was running the reproducer to give the upgrades team a place t debug16:31
rlandy|ruckeventually, I just asked the admins to hold nodes for them16:31
rlandy|ruckso they are all set atm16:31
weshayk k16:32
rlandy|ruckno emergency - but I need to get this sorted16:32
rlandy|ruckso nobody else hits it16:32
weshayrlandy|ruck https://github.com/rdo-infra/ansible-role-tripleo-ci-reproducer/blob/master/defaults/main.yaml#L6016:33
weshayrlandy|ruck I saw that bug in upstream or rdo zuul one time as well16:33
weshayzuul_image: rdoci/zuul:stable16:33
weshayzuul_scheduler_image: rdoci/zuul-scheduler:stable16:33
weshayzuul_web_image: rdoci/zuul-web:stable16:33
weshayzuul_executor_image: rdoci/zuul-executor:stable16:33
weshayzuul_merger_image: rdoci/zuul-merger:stable16:33
weshayzuul_fingergw_image: rdoci/zuul-fingergw:stable16:33
weshayya.. so we're uploading zuul containers to rdoci namespace16:34
weshayI bet they've been updated and probably fixed16:34
weshayhttps://hub.docker.com/u/zuul16:34
rlandy|ruckweshay: there is a review attached to the bug16:34
rlandy|ruckit never merged16:34
rlandy|ruck<weshay> I bet they've been updated and probably fixed16:35
rlandy|ruck^^ who is they?16:35
weshayin docker.io16:35
weshaybecause I think it was a zuul16:35
rlandy|ruckand then we would need to update what?16:35
weshayissue16:35
weshaythe zuul code in the zuul containers16:36
rlandy|ruckwhy am I the only one finding it then?16:36
weshayit wasn't a reproducer issue16:36
weshayrlandy|ruck only one using it16:36
weshayno one is using any of the reproducers atm16:36
rlandy|ruckweshay: sshnaidm|afk tried it16:36
rlandy|ruckbut maybe he has some modified env?16:36
rlandy|ruckalso the vexxhost reproducer ci job keeps failing16:37
weshayugh...16:37
weshayhttps://hub.docker.com/r/zuul/zuul/tags16:37
weshaythey have one tag16:37
rlandy|ruckI just left it until we sort out vexxhost16:37
weshaygeez16:37
weshayrlandy|ruck ya.. it's not on fire16:37
rlandy|ruckit was pretty embarrassing though to struggle so much to produce a test env16:37
rlandy|ruckweshay: suggested action? woudl not want to leave this for the next unsuspecting ruck16:38
weshayrlandy|ruck see if we can setup a job .. like the reproducer job.. that is pulling the latest zuul containers from docker.io16:40
weshaywe we don't have to constantly wonder the state of those containers16:40
weshayrlandy|ruck so something that would override the defaults16:40
rlandy|ruckweshay: no so familiar with where the reproducer container settings are going down - few minutes - will see if I can set up job to override16:41
*** dtantsur is now known as dtantsur|afk16:42
rlandy|rucklunch - back in a few16:45
weshayrlandy|ruck we could probably just have a perm. depends-on  in a job16:55
raukadahhttps://twitter.com/thomasdcameron/status/116638661330538086416:59
weshayhttps://review.rdoproject.org/r/2196417:01
*** derekh has quit IRC17:02
*** pierreprinetti has quit IRC17:12
*** pierreprinetti has joined #oooq17:13
*** pierreprinetti has quit IRC17:15
*** jaosorior has quit IRC17:16
rlandy|ruckweshay: thanks - watching17:20
weshayrlandy|ruck one mroe coming17:20
rlandy|ruckmaybe can try that out in my env17:20
rlandy|ruckk - waiting17:21
*** brault has joined #oooq17:23
weshayrlandy|ruck something like that I think https://review.rdoproject.org/r/2196517:23
weshayrlandy|ruck it may fail on a token17:24
rlandy|ruckhmmm17:25
rlandy|rucklet's see17:26
rlandy|ruckpost failures there17:26
rlandy|ruckweshay: 2019-08-27 16:24:12.952375 | TASK [Run ansible playbook to collect logs]17:27
rlandy|ruck2019-08-27 16:24:18.006183 | Timeout exception waiting for the logger. Please check connectivity to [38.145.35.133:19885]17:27
rlandy|ruck2019-08-27 17:19:13.911826 | primary | ERROR17:27
rlandy|ruck^^ on rdo jobs themselves now17:27
weshayufn17:28
weshayfun17:28
rlandy|rucktristanC is going t be happy when his rotation is over as well17:29
rlandy|ruckpinged17:29
*** tesseract has quit IRC17:30
weshayrlandy|ruck where did he drop the latest link to logreduce17:34
* weshay checks email17:34
weshaywanted to check that out17:34
* rlandy|ruck gets17:34
rlandy|ruck<tristanC> weshay: rlandy: i updated the logreduce filters to remove most of the false positive, here is the most recent report http://logs.rdoproject.org/67/678767/1/openstack-check/tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset001/22b609d/report.html  which looks pretty good17:34
rlandy|ruckthat one>17:34
weshay0.262 | 125904: 2019-08-27 12:01:58 | "InternalServerError: Internal Server Error (HTTP 500)",17:36
weshay0.000 | 125905: 2019-08-27 12:01:58 | "DEBUG:keystoneauth.identity.v3.base:Making authentication request to http://172.17.0.109:5000/v3/auth/tokens",17:36
rlandy|ruckwe see all sorts of logging problems17:37
rlandy|ruckwhen there is a lot of activity17:37
weshayya.,..  I like trying to root cause the issue when I see these logs to see if it's really correct17:38
rlandy|ruckweshay: we dropped the max-servers once17:39
rlandy|ruckmaybe we should drop that number again?17:39
weshayya.. some working jobs is better than no working jobs17:39
rlandy|ruckok - getting review17:39
weshayrlandy|ruck I've updated the program call doc17:54
rlandy|ruckweshay: thanks17:55
rlandy|ruckI will show up though17:55
rlandy|ruckchecking doc17:56
raukadahweshay: can we add a summary on what we found on adding RHEL-8 gating? like the collectd package issue17:56
weshayraukadah where?17:57
weshayand why are you still here?17:57
raukadahmy laptop is still open so lurking around on different channels17:58
raukadahweshay: to prod chain council17:58
weshayraukadah keeping some notes on what gains we've had from running rhel 8 is a smart idea17:59
raukadahweshay: I will drop an email tomorrow17:59
weshayI suspect we *may* want to hold off on a full communication of that until centos8 drops17:59
*** fmount has quit IRC18:00
*** jaosorior has joined #oooq18:00
weshayif things go relatively well w/ centos8 we can claim a big success18:00
rlandy|ruckmax server changed merged18:00
weshayif they don't go well.. well.. don't know :) raukadah18:00
raukadahthen it will be our experience18:00
raukadah!success == experience18:01
openstackraukadah: Error: "success" is not a valid command.18:01
rlandy|ruckhopefully next master run will be better18:01
rlandy|ruckweirdo-master-promote-packstack-scenario003 - still failing18:01
rlandy|ruckthought that wa sfuxed??18:02
rlandy|ruckwas fixed18:02
raukadahrlandy|ruck: one thing I realized today tempest network basic ops tests is too mucn annoying failing with common error ssh timeout18:02
*** fmount has joined #oooq18:02
rlandy|ruckraukadah: on rdocloud?18:02
raukadahyes18:02
rlandy|ruckthere are lots of timeouts there18:02
rlandy|ruckI am not sure it's tempest's fault18:02
rlandy|ruckthat just runs last18:02
rlandy|ruckand gets hit18:02
rlandy|ruckwe dropped the server numbers again18:03
rlandy|ruckto lessen the load18:03
raukadahdonot know may be some real bug hidden there18:03
rlandy|ruckyesterday we found that we kept overwhelming the log server18:03
rlandy|ruckthat is true18:03
rlandy|ruckbut it will still show18:03
rlandy|ruckI'm still running timeout comparisons18:04
raukadahthis sprint we have stats, post_failure and 401 unauthorized access are our night maers18:04
rlandy|ruckI don't have a definitive culprit yet18:04
rlandy|ruckI followed up with caching change18:04
rlandy|ruckEmilien and Alex ditched that as a cause18:04
rlandy|ruckack 401 did us in18:05
rlandy|ruckand may return18:05
raukadahyup, it was awesome , too much learning :-)18:05
rlandy|ruckrdocloud also has less attention than before18:05
rlandy|ruckfocus now being on vexxhost18:05
rlandy|ruckso working within our resources may be our best bet18:05
rlandy|rucktemp18:05
rlandy|ruckraukadah: one more day for us :)18:06
raukadahweshay: are we planning to do our team meeting in india ?18:06
raukadahthat would be fune :-)18:06
raukadah*fun18:07
rlandy|ruckwant to host us all?18:07
raukadahrlandy|ruck: our boss weshay if he wishes :-)18:07
weshayraukadah first let me get everyone in the door, w/ a desk18:08
weshayraukadah get training started...18:08
weshayraukadah I'm planning on stealing the content from bootcamps18:08
weshayand yes.. bringing to Pune18:08
weshayraukadah NOTE: this is all in my head atm, I've only briefly talked about it w/ phil and others18:09
raukadahweshay: :-)18:09
rlandy|ruckit's like three days travel, right?18:10
raukadahrlandy|ruck: depends on where we are based on18:10
raukadahI think In india, we donot have visa issues18:10
raukadahweshay: rlandy|ruck zbr sshnaidm|afk panda arxcruz rfolco https://kubernetes.academy/18:11
rlandy|ruckAll U.S. citizens need a valid passport and valid Indian visa to enter and exit India for any purpose18:12
rlandy|ruck^^ according to website18:12
* raukadah loves this slide https://twitter.com/tobyhede/status/1166200411910365185 on k8s18:12
raukadahrlandy|ruck: yes,18:15
zbri agree, running my k8n cluster with kubespray was a PITA.18:15
raukadahif you are a foodie & travel, India is the best place, with more than 100 + food items and 28 + states with travel hard to complete in an year18:16
raukadah*in an18:16
raukadah*in a18:16
rlandy|ruck13:57:34 TASK [Failure detected when testing packstack-scenario003] *********************18:19
rlandy|ruck13:57:34 task path: /home/jenkins/workspace/weirdo-master-promote-packstack-scenario003/weirdo/playbooks/packstack-scenario003.yml:3518:19
rlandy|ruckstill18:19
raukadahweshay: I am thinking to move collect-logs role to openstack-ansible-sig https://etherpad.openstack.org/p/ansible-sig once our integration work is done, but needs approval18:19
weshayraukadah++++18:20
raukadahI learned one thing today give + take ~= collaboration18:22
rlandy|ruckhttps://review.opendev.org/#/q/topic:000-upgrades-voting+(status:open+OR+status:merged)18:32
rlandy|ruck^^  don;t think I missed anything here18:32
rlandy|ruckbut if anyone notices, pls comment/vote18:33
rlandy|ruckhttps://review.opendev.org/#/q/topic:move-cli-extras+(status:open+OR+status:merged)18:37
rlandy|ruckrequires votes pls18:37
*** ksambor has joined #oooq18:51
*** ksambor has quit IRC18:51
zbrweshay: no invitation to Pune for me? I love Indian food18:55
weshayzbr perhaps :)18:55
weshaywe'll be expecting people to lead training around tripleo..18:56
weshayzbr perhaps we can have some training on molecule etc18:56
weshaywhen we get these folks in the builidng .. November 2019.. we'll start talking about it some more18:57
zbrthat was implied assumption, i do not expect to go there for the food.18:58
zbrmaybe I should change my tagline "making molecule presentations, for food" :D18:58
weshayzbr that would get me excited18:59
zbri need to go now, already 8pm here. as an update, I just raised https://review.opendev.org/#/c/678938/ which is supposed to address the random errors with buildah containers (if it passed the check)19:00
zbrthat is important because we have another change required by scenario 1-4 which was never merged because it was failing on buildah.19:01
*** sanjayu_ has quit IRC19:06
rlandy|ruckweshay: going back to testing reproducer patches19:07
rlandy|ruckwhich ones would I use locally?19:08
weshaycontainers?19:08
rlandy|ruckyep19:08
rlandy|ruckI would shut down my current deployment and start again19:09
rlandy|ruckit pulls from master though19:09
weshayya.. try out changing the registry addr and namespace and tag19:09
weshaymaster?19:09
weshayno19:09
weshaythe reproducer containers..   pull from rdo registry.. not related to openstack brnaches19:10
rlandy|ruckk - let's see19:12
rlandy|rucktesting on tenant19:51
weshaysee you on tues :)20:11
*** weshay is now known as weshay_pto20:11
*** Vorrtex has quit IRC20:15
rlandy|ruckweshay_pto: next monday :)20:21
rlandy|ruckI'm out20:21
rlandy|ruckweshay_pto: ugh - still waiting for logger20:24
rlandy|ruckforget that20:24
*** brault has quit IRC20:48
*** aakarsh|2 has quit IRC21:02
*** Goneri has quit IRC21:13
*** jtomasek has quit IRC21:29
*** aakarsh|2 has joined #oooq22:11
*** brault has joined #oooq22:49
*** brault has quit IRC22:53
*** rlandy|ruck has quit IRC23:57

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!