Monday, 2018-04-23

*** agopi has joined #oooq00:06
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.00:19
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.02:19
*** ykarel_ has joined #oooq03:24
ykarel_quiquell|off, panda when you are back please check https://review.rdoproject.org/r/#/c/13429/, promoter is not promoting queens from long03:55
ykarel_it promoted only once but queens passed multiple runs03:55
*** ykarel_ is now known as ykarel03:55
*** skramaja has joined #oooq04:16
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.04:19
*** pgadiya has joined #oooq04:23
*** pgadiya has quit IRC04:23
*** pgadiya has joined #oooq04:26
*** pgadiya has quit IRC04:26
*** jtomasek has joined #oooq05:12
*** ratailor has joined #oooq05:35
*** jfrancoa has joined #oooq05:46
*** marios has joined #oooq05:52
*** quiquell|off is now known as quiquell|ruck06:08
quiquell|ruckGood morning ykarel06:08
ykarelquiquell|ruck, Good Morning06:08
ykarelquiquell|ruck, please check the promotion issue i mentioned above06:10
*** kopecmartin has joined #oooq06:10
quiquell|ruckykarel: Checking06:10
ykarelOk06:11
ykarelcurrently queens is conflicting with ocata so promotion for queens is not starting06:11
quiquell|ruckPromoter global lock ?06:12
ykarelyes06:12
ykarelif it's possible to stop ocata promotion somehow, it would be good to stop it temporary06:12
quiquell|ruckLet me do some forensics first , to check why ocata is not finishing06:13
*** links has joined #oooq06:13
ykarelquiquell|ruck, it's finishing but it start and stop is between start of queens06:13
ykarelhttp://38.145.34.55/queens.log and http://38.145.34.55/ocata.log06:14
quiquell|ruckAnd queen is not reganing the locking after ?06:14
ykarelno, again and again same situation is commint06:14
quiquell|ruckThat's weird06:14
quiquell|ruckFeels like it's not a semaphore06:14
ykarelLast 5 queens attempt:- 05:24:01, 05:34:01, 05:44:01, 05:54:01, 06:04:0106:16
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.06:19
ykarelLast 5 queens attempt:- 05:24:01, 05:34:01, 05:44:01, 05:54:01, 06:04:0106:20
ykarelLast 5 ocata start, stop:- 05:24:01 - 05:24:14  ,05:33:01 - 05:34:13 ,05:43:01 - 05:44:13 ,05:53:01 - 05:54:14 ,06:03:01 - 06:04:13 ,06:13:02 - 06:14:1406:20
ykarelquiquell|ruck, ^^06:20
ykarelso ocata is winning everytime,06:21
ykarelto acquire lock06:21
quiquell|ruckI will check the code, I think queens is timing out on the lock06:23
quiquell|ruckAnd ocata enters and re-gain again06:23
quiquell|ruckI am going to activate queens by hand06:23
quiquell|ruckTo have a promotion at least06:23
quiquell|ruckMaste is screw with issues06:24
quiquell|ruckmaster06:24
ykarelquiquell|ruck, Ok06:24
*** jaganathan has joined #oooq06:24
quiquell|ruckI have to leave now going back in a few06:24
ykarelyes i pushed a patch for master, but it need +1W06:24
*** jaganathan has quit IRC06:25
quiquell|ruckquiquell|ruck: Let's find someone to +1w that06:25
*** quiquell|ruck is now known as quiquell|ruck|af06:25
*** quiquell|ruck|af is now known as quique|ruck|afk06:25
ykarelOkk06:25
*** jaganathan has joined #oooq06:25
*** apetrich has joined #oooq06:46
*** apetrich has quit IRC06:46
*** jbadiapa has joined #oooq06:51
*** ccamacho has joined #oooq06:56
*** quique|ruck|afk is now known as quiquell|ruck07:01
*** apetrich has joined #oooq07:02
*** tesseract has joined #oooq07:04
quiquell|ruckykarel: I have a WIP for the sequencial of promoter07:10
quiquell|ruckhttps://review.rdoproject.org/r/#/c/13437/07:10
quiquell|ruckIt's just a hack not big refactoring07:10
quiquell|ruckBut doing so we need to remove it from crontab07:10
ykarelquiquell|ruck, hmm adjustment needs to be done if we change the way it used to work07:11
ykarelquiquell|ruck, have you started queens promotion, or stopped ocata one?07:12
quiquell|ruckNot yet07:12
ykarelOkk07:13
quiquell|ruckykarel: Going to do a script to run them  in parallel07:22
ykarelquiquell|ruck, running in parallel won't cause a issue that we saw earlier(layer image issue)?07:23
quiquell|rucks/parallel/sequencial/07:23
quiquell|ruckCoffe is still reaching my blood07:23
ykarelOk:)07:24
*** bogdando has joined #oooq07:24
*** tosky has joined #oooq07:28
quiquell|ruckykarel: Running in sequence now07:33
quiquell|ruckI will give the script a pin07:34
ykarelOk. let's see how it goes07:34
quiquell|ruckRunning now for queens07:37
ykarelNice, promotion started07:38
quiquell|ruckHumm maybe the order is not quite correct07:38
ykarel2018-04-23 07:37:35,485 17070 INFO     promoter Promoting the container images for dlrn hash 5466f249bd36900a1dac573cdc83e7a11493aea2 on queens to current-tripleo07:38
ykareli noticed master --> pike --> ocata --> queens07:38
quiquell|ruckyep07:39
quiquell|ruckThat's wrong07:39
quiquell|ruckmaster -> queens --> pike --> ocata would be the correct07:39
quiquell|ruckLet's queens finish07:39
quiquell|ruckAnd I change the script07:39
ykarelOkk queens need multiple promotions07:39
quiquell|ruckMaster is fucked up so07:39
quiquell|ruckAlso the other ones07:40
quiquell|ruck:-)07:40
quiquell|ruckSo it will do only queens07:40
ykarelyes, we have +1W now on master fix07:40
quiquell|ruckSo let's not fix stuff now to have a lot of queens promotions07:40
quiquell|ruck:-)07:40
ykarelhoping to get that merged before next promotion run07:40
*** amoralej|off is now known as amoralej07:42
quiquell|ruckykarel: Do we want to close https://bugs.launchpad.net/tripleo/+bug/176500807:43
openstackLaunchpad bug 1765008 in tripleo "Tempest API tests failing for stable/queens branch" [Critical,In progress] - Assigned to Gabriele Cerami (gcerami)07:43
quiquell|ruckOr keep opening to progress with segments ?07:43
ykarelquiquell|ruck, i think promotion-blocker tags can be removed, and close once https://review.openstack.org/#/c/562932/ is merged07:45
quiquell|ruckThis also depends on removing from the skip list the test07:46
quiquell|ruckIt's no longer a blocker, the test is in the skip list now07:46
quiquell|ruckWe can work it without hurries07:46
quiquell|ruckJust missing the +2v at https://review.openstack.org/#/c/563443/07:48
quiquell|ruckOK07:48
quiquell|ruckCool :-)07:48
ykarelhmm07:49
*** ykarel is now known as ykarel|lunch07:56
*** agopi has quit IRC08:05
amoralejhi08:05
amoralejwe are finally moving dependencies for master to new repo based on rocky tag08:06
amoraleji don't expect any impact08:06
amoralejbut let me know if you notice anything abnormal08:06
amoralejykarel|lunch, quiquell|ruck ^08:06
quiquell|ruckamoralej: Noted08:06
quiquell|ruckamoralej: newbie question, what are the impacts of it ?08:08
amoralejquiquell|ruck, currently we are using the deps from queens08:09
amoralejwe need to move to rocky, as threre are packages that are only required in rocky08:09
amoralejso we are moving that08:09
amoralejin this particular case we are splitting the dependencies in two repos08:10
amoralejit should be transparent for you08:10
amoralejbut, if you notice something starts failing because of lack of packages or something08:10
amoralejlet me know, in case i mess it up08:10
*** jaosorior has joined #oooq08:15
quiquell|ruckOk will keep you in my mind08:15
*** lucas-afk is now known as lucasagomes08:15
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.08:19
quiquell|ruckykarel|lunch: ERROR:  Error installing fluent-plugin-kubernetes_metadata_filter:08:27
quiquell|ruckINFO:kolla.image.build.fluentd: serverengine requires Ruby version >= 2.1.0.08:27
quiquell|ruckINFO:kolla.image.build.fluentd:[0m08:27
quiquell|ruckINFO:kolla.image.build.fluentd:[91minvalid options: -SHN08:27
quiquell|ruckINFO:kolla.image.build.fluentd:(invalid options are ignored)08:27
quiquell|ruckINFO:kolla.image.build.fluentd:[0m08:27
quiquell|ruckhttps://logs.rdoproject.org/openstack-periodic-24hr/periodic-tripleo-centos-7-pike-containers-build/b7b3eff/kolla/logs/fluentd.log08:27
*** ykarel|lunch is now known as ykarel08:31
*** jaosorior has quit IRC08:32
ykarelquiquell|ruck, since when the job is failing08:33
ykarelamoralej, ack08:34
quiquell|ruck20 of april08:34
quiquell|ruckLooking about -SHN option08:34
quiquell|ruckrdoc has the -SHN opt ion08:35
ykarelalso both master and queens containers build timedout: Build timed out (after 240 minutes). Marking the build as failed.08:38
ykarellooks abnormal08:38
ykarelINFO:kolla.common.utils.zaqar:Trying to push the image08:39
ykarelERROR:kolla.common.utils.zaqar:received unexpected HTTP status: 500 Internal Server Error08:39
quiquell|ruckykarel: Damn08:39
quiquell|ruckHumm about the pikes issue08:40
quiquell|ruckLooks like active support is not being installed08:40
quiquell|ruckactive_support08:40
quiquell|ruckunable to convert "\x84" from ASCII-8BIT to UTF-8 for lib/active_support/values/unicode_tables.dat08:40
quiquell|ruckAnd maybe that's why SHN option is not in place08:40
ykarelquiquell|ruck, /me no idea about this08:40
quiquell|ruckOIk08:40
*** sshnaidm|off is now known as sshnaidm08:41
quiquell|ruckBug openned08:43
quiquell|ruckhttps://bugs.launchpad.net/tripleo/+bug/176619508:43
openstackLaunchpad bug 1766195 in tripleo "Error installing fluent-plugin-kubernetes_metadata_filter at pike" [Critical,New] - Assigned to Gabriele Cerami (gcerami)08:43
pandaquiquell|ruck: morning09:00
quiquell|ruckpanda: Morning09:00
quiquell|ruckpanda: bj to do a sync ?09:01
quiquell|ruckykarel: the kolla timeout feels transtitory, older jobs doesn't have it09:01
quiquell|ruckThe one consistent is the fluentd09:01
quiquell|ruckDamn My bj is close wait09:02
quiquell|ruckOk now is on09:03
ykarelquiquell|ruck, Ok good to focus on fluentd, kolla one can be checked in next run09:07
quiquell|ruckLooking at the timeouts now09:07
arxcruzmyoung: chandankumar kopecmartin weshay hey guys, i have to leave in 2 hours for an appoitment, I should be back in time for the scrum, but if i don't I already update my cards :)09:07
*** jaosorior has joined #oooq09:07
kopecmartinarxcruz, ack09:08
*** moguimar has quit IRC09:08
quiquell|ruckykarel: I see a lot of failures at postin rpm packaging09:09
quiquell|ruckat kolla09:09
*** brault has joined #oooq09:10
*** moguimar has joined #oooq09:10
chandankumararxcruz: ack09:11
ykarelquiquell|ruck, postin?09:12
quiquell|ruckpost installation step of the spec file09:13
ykarelokk09:13
quiquell|ruck200~D-Bus connection: Operation not permitted09:13
quiquell|ruckShit09:13
ykarelahh, that can be a real issue09:13
ykarelnot transient09:13
quiquell|ruckThat is master promotion timeout09:14
pandaquiquell|ruck: ok09:14
quiquell|ruckpanda, ykarel: The master/queens timeout is a 500 Internal Server Error trying to push nova-compute09:33
quiquell|ruckIt's the same that you show at pikes ykarel09:33
*** zoli is now known as zoli|lunch09:33
quiquell|ruckpanda: master/queens/pikes promotion blocker https://bugs.launchpad.net/tripleo/+bug/176620209:41
openstackLaunchpad bug 1766202 in tripleo "Pushing nova-compute kolla image gets HTTP status: 500 Internal Server Error" [Critical,New] - Assigned to Gabriele Cerami (gcerami)09:41
quiquell|ruckpanda: You missing the |rover in the nick09:42
ykarelquiquell|ruck, it not just nova-compute, it's seen in many container images09:45
quiquell|ruckykarel: Ok, renaming it09:45
ykarelquiquell|ruck, where you saw: 200~D-Bus connection: Operation not permitted?09:46
quiquell|ruckykarel: At masters, but maybe it has been there always09:47
quiquell|ruckIn the containers build09:47
quiquell|ruckHave put the info here09:47
quiquell|ruckhttps://review.rdoproject.org/etherpad/p/ruckrover-sprint1209:47
ykarelOkk, then Internal Server can be a transient one09:48
quiquell|ruckCould be...09:48
quiquell|ruckIs there something similar at check or gate ?09:48
ykarelfs007 --> scenario00709:48
quiquell|ruckThey will be run more often09:49
quiquell|ruckUps ok09:49
quiquell|ruck:-/09:49
quiquell|ruckThanks man09:49
ykareland Internal server Error is master and queens, not pike09:49
quiquell|ruckIt's also pike09:50
ykarelOkk, i haven't seen that09:50
quiquell|ruckPike has two issues now09:50
quiquell|ruckHumm this is pushing to docker.io not rdo ?09:52
quiquell|ruckDamn now I don't find the 500 internal error log at master/queens09:59
quiquell|ruck!gatestatus10:00
openstackquiquell|ruck: Error: "gatestatus" is not a valid command.10:00
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.10:00
*** adarazs is now known as adarazs_lunch10:09
*** sshnaidm has quit IRC10:10
*** dtantsur|pto is now known as dtantsur10:16
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.10:19
*** panda is now known as panda|rover10:27
*** zoli|lunch is now known as zoli10:54
*** sshnaidm has joined #oooq11:03
quiquell|ruckykarel:11:04
quiquell|ruckyRevert "Fix neutron-plugin-ml2.yaml puppet base ref11:04
quiquell|ruckmerged11:04
ykarelyup let's wait for it to be packaged11:04
ykarelquiquell|ruck, you can see it here once it's packaged:- https://trunk.rdoproject.org/api-centos-master-uc/api/report.html?package=openstack-tripleo-heat-templates11:06
*** lucasagomes is now known as lucas-hungry11:06
ykarelcommit: 297aac5c255e0c9c55740c7a611f5279dc3a173511:06
quiquell|ruckThen I can move it to released ?11:06
quiquell|ruckat lp ?11:06
ykareli think released is autodone with the new releases11:07
quiquell|ruckbtw queens just promoted11:07
quiquell|ruckOOk11:07
ykarelnice current-tripleo-rdo11:07
*** moguimar has quit IRC11:13
*** dpeacock has quit IRC11:14
ykarelquiquell|ruck, package is built and in consistent repo, so next periodic run will have the fix11:20
quiquell|ruckOk, let's wait for it11:20
chandankumararxcruz: if you are around, do we need more work on this one https://trello.com/c/wwpJfjRA/695-tempest-run-should-take-less-than-5-min-in-tripleo-undercloud-jobs ?11:29
ykarelquiquell|ruck, added alert tag on https://bugs.launchpad.net/tripleo/+bug/176619511:32
openstackLaunchpad bug 1766195 in tripleo "Error installing fluent-plugin-kubernetes_metadata_filter at pike" [Critical,New] - Assigned to Gabriele Cerami (gcerami)11:32
quiquell|ruckykarel: I have to read about the openstack bot11:33
quiquell|ruckThanks btw11:33
ykarelnp11:33
*** amoralej is now known as amoralej|lunch11:36
*** panda|rover is now known as panda|rover|lunc11:41
*** adarazs_lunch is now known as adarazs11:47
*** quiquell|ruck is now known as quique|ruck|food11:47
*** rfolco|off is now known as rfolco11:55
*** atoth has joined #oooq11:56
*** lucas-hungry is now known as lucasagomes12:06
quique|ruck|foodsshnaidm: featureset030 and featureset035 are missing from sova12:09
quique|ruck|foodCould it be possibe for sova to read them from promoter master.ini ?12:09
*** trown|outtypewww is now known as trown12:12
ykarelquique|ruck|food, updated the bug:- https://bugs.launchpad.net/tripleo/+bug/176619512:14
openstackLaunchpad bug 1766195 in tripleo "Error installing fluent-plugin-kubernetes_metadata_filter at pike" [Critical,Confirmed] - Assigned to Gabriele Cerami (gcerami)12:14
sshnaidmquique|ruck|food, both are in config of sova12:15
sshnaidmquique|ruck|food, and both are displayed in promotion page12:16
sshnaidmquique|ruck|food, what the problem do you hit exactly?12:16
quique|ruck|foodsshnaidm: The problem is that I'm totally blind... they are already there :-) sorry12:19
sshnaidmquique|ruck|food, np :)12:19
*** rlandy has joined #oooq12:19
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.12:19
quique|ruck|foodykarel: master container build running now, let's see were we are with the kolla image pushing12:22
ykarelquique|ruck|food, ack12:22
ykarelhoping it to not reproduce and have a promotion :)12:22
*** amoralej|lunch is now known as amoralej12:25
*** tcw has joined #oooq12:39
*** tcw has quit IRC12:39
*** tcw has joined #oooq12:40
*** faceman- is now known as faceman12:44
sshnaidmquique|ruck|food, ykarel do you know about such problem with gperftools-libs? https://logs.rdoproject.org/26/563526/1/openstack-check/gate-tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset001-master/Zd49e57a183c44c3c871ca322d7f5e63c/undercloud/home/jenkins/install_packages.sh.log.txt.gz#_2018-04-23_11_43_2912:54
quique|ruck|foodsshnaidm: They are chaning the master's dependencies from queens to rocky12:56
*** adarazs is now known as adarazs_brb12:56
quique|ruck|foodTry a recheck12:56
quique|ruck|foodamoralej: It's related ?12:56
amoralejquique|ruck|food, yes, it's related12:57
amoralejto my change12:57
*** quique|ruck|food is now known as quiquell|ruck12:57
amoralejit will be fixed whtn mirror is in sync12:57
amoralejquiquell|ruck, you are seeing many jobs failing?12:57
amoraleji could apply a fix temporarily until synchronization is done12:58
quiquell|ruckNot much12:58
quiquell|ruckI think it's starting to fail in the gates12:58
amoralejif it's critical12:58
quiquell|ruckperiodic ones take so long12:58
amoraleji think it should be ok now, in fact12:58
amoralejfor jobs starting now12:58
quiquell|ruckApply the fix will take a lot of time12:58
quiquell|ruckLet's wait a little, and send and advise to one of the mailing list12:59
sshnaidmquiquell|ruck, amoralej ack, thanks, will recheck\13:00
amoralejsshnaidm, quiquell|ruck i'll push a fix in the repos and keep it for one or two days13:01
amoralejmmmm13:02
amoralejalthough i may break other things...13:02
amoralejquiquell|ruck, sshnaidm please let me know if you still see jobs failing with that error13:02
quiquell|ruckamoralej: ack13:03
*** Goneri has joined #oooq13:04
*** ratailor has quit IRC13:21
quiquell|ruckamoralej: Now it's a promotion blocker13:21
quiquell|ruckhttps://logs.rdoproject.org/openstack-periodic/periodic-tripleo-centos-7-master-containers-build/eadb914/undercloud/home/jenkins/install_packages.sh.log.txt.gz13:21
ykarel:(13:21
amoralejyeah, i saw it13:22
amoraleji'm testing again if that's fixed now13:22
quiquell|ruckamoralej: It's a transitory problem ?13:22
amoralejquiquell|ruck, it's something that should be already fixed13:23
amoralejbut i'm testing it13:23
amoralejtesting it in https://review.rdoproject.org/jenkins/job/rdoinfo-tripleo-master-testing-centos-7-multinode-1ctlr-featureset016-nv/96/console13:24
ykarelhmm the failures from master job is 1 hour back, so should be good now13:25
quiquell|ruckbad luck13:25
quiquell|ruckLet's just wait for another periodic run13:25
quiquell|ruckamoralej: Let also check the amoralej running job13:26
*** moguimar has joined #oooq13:26
weshaythanks for checking that out quiquell|ruck  :)13:26
quiquell|ruckHello there13:27
quiquell|ruckrecursive hchecking13:27
weshayamoralej, are the changes to master/rocky impacting the gate while the sync is happening?13:29
weshaye.g. http://logs.openstack.org/63/507963/54/gate/tripleo-ci-centos-7-undercloud-oooq/54add67/logs/undercloud/home/zuul/repo_setup.log.txt.gz13:29
amoralejweshay, yes it has impacted13:30
amoralejbut impact should be low13:30
weshayamoralej, do we have any sort of tech debt card / bug to capture that13:30
amoralejlet me see that job13:30
*** adarazs_brb is now known as adarazs13:30
weshayamoralej, FYI.. anytime the gate goes down due to ci/infra we should make sure we're working to elliminate the root cause if / when possible13:31
amoralejweshay, yes, that job failure is related to new rocky13:31
amoralejweshay, ok, should i fill a card somewhere?13:32
amoraleji think i have an idea to fix that13:32
amoralejin future13:32
amoralejdon't cache repomd.xml files13:32
amoraleji need to use expiration headers for it13:32
weshayamoralej, that sounds good13:32
weshayya13:32
amoralejweshay, do you have the full list of mirrors upstream?13:33
amoraleji could force refresh13:33
weshayamoralej, I don't but I bet david or paul has it13:33
amoralejmmmm, it's still failing13:34
amoralejin https://review.rdoproject.org/jenkins/job/rdoinfo-tripleo-master-testing-centos-7-multinode-1ctlr-featureset016-nv/96/console13:34
amoralejmmmm13:34
amoraleji don't know why13:34
ykarelamoralej, this time also gperftools?13:35
amoraleji'm waiting for the logs13:36
amoralejbut i guess so13:36
amoralejlet's see if i can hold the node13:36
arxcruzchandankumar: checking13:36
ykarelamoralej, logs available now: https://logs.rdoproject.org/42/12942/6/check/rdoinfo-tripleo-master-testing-centos-7-multinode-1ctlr-featureset016-nv/Za85b635804154c8582c917a5bd949b60/undercloud/home/jenkins/install_packages.sh.log.txt.gz13:36
amoralejyeah13:37
amoralejlet me try to reproduce it13:37
weshaysshnaidm, any thoughts on https://review.openstack.org/#/c/562347/13:39
sshnaidmweshay, seems fine to me13:41
weshayhttps://review.openstack.org/#/c/563244/13:46
arxcruzomg, look how beautifull is all these SUCCESS jobs... https://review.openstack.org/#/c/561920/13:48
chandankumararxcruz: yay!13:50
*** panda|rover|lunc is now known as panda|ruck13:51
quiquell|ruckarxcruz: Better if you start like "You will not believe what this review does... "13:51
*** quiquell|ruck is now known as quiquell|rover13:51
arxcruzquiquell|rover: hahaha13:52
amoralejquiquell|rover, ykarel it should be fine now13:52
amoralejrechecking my job13:52
ykarelamoralej, hmm locally it's working now13:52
ykarelbefore 1 minute not working13:52
amoraleji'm updated the repos manually13:53
amoralejwith the needed packages13:53
amoraleji couldn't force mirror sync13:53
amoralejand in my browser repomd.xml is fine13:53
amoralejbut not in yum13:53
amoralejanyway, it's good to add those packages13:53
amoralejto the repo13:53
amoraleji'll send a rdoinfo review for it13:53
ykarelamoralej, Ok, so only gperftools is missing, or there are others as well13:54
amoralejother two which are needed by gperftools13:54
ykarelokk13:54
amoralejgv and Xaw3d13:54
ykarelack13:54
amoralejin fact, we may have others13:54
amoralejthat's why i want to rerun a job13:55
ykarelahh, Ok go on13:55
*** trown is now known as trown|brb13:56
ykarelany idea why queens phase1 is not promoting for:- 5466f249bd36900a1dac573cdc83e7a11493aea2_0c8f7f9513:57
ykarelweshay, quiquell|rover panda|ruck ^^13:58
* quiquell|rover checking13:58
myounggood morning, TripleO CI standup/scrum start shortly!  https://bluejeans.com/7050859455, https://etherpad.openstack.org/p/tripleo-ci-squad-meeting13:59
ykarelweshay, any idea why CIX card not created for https://bugs.launchpad.net/tripleo/+bug/1766195?14:00
openstackLaunchpad bug 1766195 in tripleo "Error installing fluent-plugin-kubernetes_metadata_filter at pike" [Critical,Triaged] - Assigned to Gabriele Cerami (gcerami)14:00
weshayykarel, it will, we wait for 5hrs to give us a chance to fix it before we escalate it14:01
*** trown|brb is now known as trown14:01
ykarelweshay, ack14:01
ykareli saw sometime it's created soon, so just asked14:01
quiquell|roverykarel: Promoted to tripleo-current14:03
quiquell|roverhttps://dashboards.rdoproject.org/queens14:03
quiquell|roverTo current-tripleo-rdo will take time...14:03
ykarelquiquell|rover, i can see all required jobs for phase 1 passed:- https://trunk.rdoproject.org/api-centos-queens/api/civotes_detail.html?commit_hash=5466f249bd36900a1dac573cdc83e7a11493aea2&distro_hash=0c8f7f95a9dace8504905c03862036fe270f154514:06
ykarelfor what promotion is waiting?14:07
ykarelhttp://38.145.34.55/queens.log14:07
weshaytrown, fyi https://review.openstack.org/#/q/topic:gate_update+(status:open+OR+status:merged)14:07
weshaysshnaidm, trown I made the tht zuul change non-voting but I can think of a good reason not to have it voting..14:07
weshayjust treading lightly I guess14:08
sshnaidmweshay, acc. to policy it should run some period as non-voting to prove a stability14:09
weshaysshnaidm, ya.. I think we have done that already though14:11
weshaysshnaidm, so is that true for the same job on the diff repo?14:11
weshaythat's where I'm not clear..14:12
weshaythe job is def.. stable14:12
*** matbu has quit IRC14:12
sshnaidmweshay, I think it's overall triple policy: experimental -> non-voting -> voting14:12
weshaysshnaidm, yes.. I get that .. but if you have job_A voting on repo_1 do you then have to start over for repo_2?14:13
weshayand set job_A back to experimental14:13
sshnaidmweshay, I don't think so..14:13
weshayhuh.. starting to see some success on queens w/ the update job14:14
weshaypreviously only saw it working on master14:14
*** moguimar has quit IRC14:14
*** matbu has joined #oooq14:14
ykarelquiquell|rover, containers-build for queens is running for around 2 hours, so looks like https://bugs.launchpad.net/tripleo/+bug/1766202 is not a transient issue and needs to be checked14:15
openstackLaunchpad bug 1766202 in tripleo "Pushing kolla image gets HTTP status: 500 Internal Server Error" [Critical,New] - Assigned to Gabriele Cerami (gcerami)14:15
ykarelIt would be good to hold the node: upstream-centos-7-rdo-cloud-tripleo-143846 to check what's going on14:16
ykarelamoralej, ^^14:16
*** skramaja has quit IRC14:16
amoralejthat's pushing to RDO's registry?14:17
quiquell|roverykarel: ack14:17
ykarelamoralej, yes14:18
amoralejlet me see if i see some error in logs14:18
quiquell|roverIt's running in RDO infra so yes14:18
ykarelamoralej, ack14:19
quiquell|roverykarel: Going to ask the RDO guys14:19
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.14:19
*** links has quit IRC14:30
amoralejquiquell|rover, ykarel my test job has just passed the step where it was failing before, at least14:30
ykarelamoralej, nice, let's see full run to confirm we don't have anything else missing14:31
amoralejyeah14:31
amoraleji think i know why i missed these14:31
amoralejit's kind of corner case14:31
amoralejbecause it's a requirement for ceph, and i didn't take it into account in my script as ceph shouldn't depend on things in rdo repo14:32
ykarelOkk14:32
quiquell|roverWe have to open a bug to track this14:33
quiquell|roverFailing one for 3 hours ago14:33
quiquell|roverhttps://review.rdoproject.org/jenkins/job/gate-tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset001-master/11739/14:33
amoralejquiquell|rover, let me know if you open a bug for that and i'll update it14:38
quiquell|roveramoralej: ack14:38
quiquell|roveramoralej: Do you have around the review with the changes to get deps from rocky ?14:39
amoralejquiquell|rover, not yet, but i first did it manually to workaround14:40
quiquell|roverOk14:40
quiquell|roveramoralej, ykarel: https://bugs.launchpad.net/tripleo/+bug/176627114:42
openstackLaunchpad bug 1766271 in tripleo "Installing packages: gperftools-libs-2.4.91-1.el7.x86_64.rpm not found" [Critical,Confirmed] - Assigned to Alfredo Moralejo (amoralej)14:42
panda|ruckrlandy: so this is maybe I don't get: with the RH nameserver it doesn't work. If you have both the RH nameserver and another it does ?14:47
ykarel5466f249bd36900a1dac573cdc83e7a11493aea2_0c8f7f95 queens phase1 promotion still not started(all criteria jobs passed), can this be a bug in promotion script?14:47
ykarelwill look tomorrow if it's not promoted by then14:48
ykarelquiquell|rover, ^^14:48
quiquell|roverykarel: Will check14:48
*** agopi has joined #oooq14:48
rlandypanda|ruck: if you look at what you get on the vm nodes14:48
rlandyusing dhcp14:49
rlandywhen you set up your minidell attached to rh network14:49
rlandyif you add an additional dns server14:49
rlandyyes, it works14:49
rlandyhence I sued append14:49
rlandyused14:49
rlandybut supercede will also work14:50
*** ykarel is now known as ykarel|away14:50
ykarel|awayquiquell|rover, Thanks14:50
quiquell|roverykarel|away: Thanks to you man14:50
panda|ruckrlandy: probably because the first nameserver does not respond in this case. If it starts responding and saying it doesn't exist, the second will not be queried14:51
panda|ruckIf I understood what trown was saying14:51
rlandypanda|ruck: that was my understanding14:53
rlandyalso14:53
panda|ruckrlandy: so no appending seems a safer solution14:54
rlandyI wasn't sure if all user would hit the dns resolution problem - as such forcing a supersede seemed too much14:54
panda|ruckmmhh14:55
trownya14:55
rlandybut since we are narrowing the scope to just our use case14:55
rlandyI am ok with the supersede14:55
*** quiquell|rover is now known as quiquell|off14:56
* rlandy updates review14:56
panda|ruckrlandy: thanks.14:58
*** skramaja has joined #oooq15:04
*** skramaja has quit IRC15:07
adarazsmyoung, panda|ruck, weshay: which stable check jobs do you want to monitor with hubbot? stable/{ocata,pike,queens}?15:10
panda|ruckadarazs: can I get stable/austin ?15:27
trownnope15:29
adarazspanda|ruck: we need to reach 88mph at least and start THT early :)15:29
adarazsso that I can submit a change on it :)15:29
trowncant even get stable/grizzly15:29
trownor folsom15:30
trownessex... I think that is as far back as I can go without looking at google15:30
panda|ruckto be honest I'm not sure austin was even released to the public15:30
*** chandankumar has quit IRC15:35
*** sshnaidm is now known as sshnaidm|afk15:41
*** chandankumar has joined #oooq15:42
*** sshnaidm|afk has quit IRC15:46
*** sshnaidm has joined #oooq15:48
*** sshnaidm has quit IRC15:49
*** ykarel|away has quit IRC15:49
*** moguimar has joined #oooq15:50
*** panda|ruck is now known as panda|bbl15:51
*** bogdando has quit IRC16:01
trownrlandy: I dont think we need your timeout patch if we remove cloud-init16:04
trownrlandy: I think that is actually what is causing vms to be so slow16:04
trownto boot16:04
*** ccamacho has quit IRC16:04
rlandytrown; that's fine, I can abandon the patch if the time to boot improves16:05
trownya was more just giving a heads up, becuase for iterative testing it is annoying that vms are taking more than 3 minutes to boot :P16:06
trownso if you have that issue in your testing too, try removing cloud-init16:07
*** marios has quit IRC16:11
*** ccamacho has joined #oooq16:11
*** zoli is now known as zoli|gone16:14
*** zoli|gone is now known as zoli16:14
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.16:20
myoungkopecmartin, arxcruz, chandankumar, weshay, tempest squad scrum in 3 min, https://etherpad.openstack.org/p/tripleo-tempest-squad-meeting, https://bluejeans.com/705085945516:27
*** lucasagomes is now known as lucas-afk16:29
arxcruzrlandy: weshay trown can one of you guys +w https://review.openstack.org/#/c/562155/ ?16:32
*** trown is now known as trown|lunch17:00
*** sshnaidm has joined #oooq17:00
*** moguimar has quit IRC17:02
*** sshnaidm is now known as sshnaidm|off17:02
*** jaosorior has quit IRC17:05
*** kopecmartin has quit IRC17:12
*** tesseract has quit IRC17:23
*** dtantsur is now known as dtantsur|afk17:33
*** agopi is now known as agopi|lunch17:42
*** jaosorior has joined #oooq17:49
*** trown|lunch is now known as trown18:05
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 and fix them ASAP.18:20
*** jaosorior has quit IRC18:55
*** Goneri has quit IRC18:56
*** amoralej is now known as amoralej|off19:06
*** agopi|lunch is now known as agopi19:07
*** atoth has quit IRC19:20
*** jfrancoa has quit IRC19:29
*** jfrancoa has joined #oooq19:42
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, gate-tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset001-master, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, gate-tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset035-master, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 a20:20
rlandytrown: wrt removing cloud-init, do you have something for that change yet? if not will put up a patch to with another virt-customize line in https://github.com/openstack/tripleo-quickstart/blob/master/roles/libvirt/setup/overcloud/tasks/fake_nodepool.yml20:27
rlandythe libvirt reproducer is working so far for me with the extra wait time20:28
*** dougbtv_ has joined #oooq20:37
trownrlandy: ya i dont have a patch for it... I just did it in one of my snapshots20:39
*** dougbtv has quit IRC20:40
trownrlandy: but that is the right spot i think20:40
rlandytrown: cool - will put something up for review20:40
rlandyreproducer is going fine on fs01020:40
rlandywill move on to other sets afterwards20:40
rlandythe reproducer patch is updated20:41
rlandytrown: I can test the snapshot stuff tomorrow20:41
trownrlandy: sure.. I have still been ironing out some kinks in the way I checked for stopped vms today20:42
trownrlandy: but it should be ready for testing tomorrow20:42
*** jfrancoa has quit IRC20:42
rlandyI didn't sign up for it yet as I thought someone else might volunteer but if nobody signs by then, I'll be QE20:42
*** jtomasek has quit IRC20:43
trownI was able to figure out the fs037 issue is something to do with iptables... havent checked exactly what yet though20:44
rlandyoh - that's good20:44
trownya I think we will likely have that featureset working in reproducer by the end of the sprint20:45
trownwhich is the one that gave me so much trouble last sprint20:45
rlandywell then we accomplished something20:55
*** Goneri has joined #oooq21:00
*** dougbtv__ has joined #oooq21:01
*** trown is now known as trown|outtypewww21:02
*** dougbtv_ has quit IRC21:03
rlandytrown|outtypewww: ha - disabling cloud-init does fix the long reboot21:21
*** Goneri has quit IRC22:07
*** jrist has quit IRC22:08
*** jrist has joined #oooq22:12
*** jrist has quit IRC22:12
*** jrist has joined #oooq22:12
hubbotFAILING CHECK JOBS: gate-tripleo-ci-centos-7-container-to-container-upgrades-master-nv, gate-tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset001-master, tripleo-quickstart-extras-gate-newton-delorean-full-minimal, gate-tripleo-ci-centos-7-ovb-3ctlr_1comp-featureset035-master, tripleo-ci-centos-7-scenario007-multinode-oooq-container | check logs @ https://review.openstack.org/472607 a22:20
*** tosky has quit IRC23:54

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!