*** rlandy|bbl is now known as rlandy|out | 01:10 | |
*** dviroel|rover|afk is now known as dviroel|rover | 01:17 | |
*** ysandeep|out is now known as ysandeep | 01:37 | |
ysandeep | hello oooci o/ morning | 01:38 |
---|---|---|
*** ysandeep is now known as ysandeep|afk | 03:45 | |
marios | fridaaaaayyyy \o/ | 05:25 |
*** ysandeep|afk is now known as ysandeep | 06:16 | |
ysandeep | happy friday marios and all o/ | 06:20 |
marios | \o | 06:24 |
jm1 | o/ | 06:46 |
marios | chandankumar: o/ please vote when you get a chance (the two should go together) https://review.opendev.org/c/openstack/tripleo-ci/+/849883 | 06:53 |
*** chandankumar is now known as chkumar|rover | 06:59 | |
chkumar|rover | chkumar|rover: done | 07:02 |
chkumar|rover | jm1: \o | 07:03 |
chkumar|rover | jm1: please address the rlandy and doug suggestion on this review https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/847419 when free, thanks! | 07:03 |
jm1 | chkumar|rover: working on it.. | 07:04 |
marios | https://logserver.rdoproject.org/58/43558/13/check/tripleo-ci-centos-8-9-multinode-mixed-os/e621742/logs/undercloud/var/log/tempest/stestr_results.html.gz \o/ | 07:43 |
* marios collapses | 07:43 | |
jm1 | marios: woohooo 🥳 | 09:04 |
marios | jm1: ;) | 09:09 |
chkumar|rover | marios: awesome \o/ | 09:13 |
marios | chkumar|rover: \o/ | 09:19 |
bhagyashris | marios, woohoo \o/ | 09:21 |
marios | :) | 09:22 |
ysandeep | marios, so in this job we have centos9 controller and centos8 compute right? | 09:22 |
marios | ysandeep: why do i feel like i am not going to like your next statement :) | 09:23 |
marios | ysandeep: yes | 09:23 |
marios | ysandeep: ... did you spot something? | 09:26 |
ysandeep | marios: cool thanks.. nah all good, just wondering will we support instance migration from c8 compute to c9 compute and vice versa. | 09:26 |
marios | ysandeep: not sure about that and don't know if it has been fully decided. not just instance migration, but generally there will be a 'degraded' support when in this state (this will happen during an upgrade so the deployment is still 'in the upgrade state' the whole time until all the 8 computes are done) | 09:27 |
marios | ysandeep: so don't know if will be supported or what will be supported | 09:27 |
marios | ysandeep: like upgrades/compute/df have to agree and sign off on what they can support for day2 operations in that state | 09:28 |
marios | (scale out etc) | 09:28 |
ysandeep | ah okay | 09:29 |
ysandeep | so we deploying ctlr stack first and then overcloud deploy | 09:30 |
ysandeep | nice we have logs for both https://logserver.rdoproject.org/58/43558/13/check/tripleo-ci-centos-8-9-multinode-mixed-os/e621742/logs/undercloud/home/zuul/ctrl_overcloud-deploy.sh.txt.gz and https://logserver.rdoproject.org/58/43558/13/check/tripleo-ci-centos-8-9-multinode-mixed-os/e621742/logs/undercloud/home/zuul/overcloud-deploy.sh.txt.gz | 09:31 |
marios | yeah i actually just kept the default name (as i didn't need to change it/didn't give a value) so i just have 'overcloud' for both. but yes, first deployment of control then compute | 09:31 |
marios | ysandeep: yeah i just kept some of those logs but we can revisit/add whatever | 09:31 |
marios | ysandeep: for eaxmple maybe using different stack name instaed of 'overcloud' for both will give us better log files? perhaps that is a value | 09:32 |
marios | but anyway there will be lots to fix and improve but first would like to start getting the base merged https://review.opendev.org/q/topic:oooci_mixed_rhel which will take a few days at least i think | 09:32 |
chkumar|rover | marios: I have left some initial comments, I am not sure you saw that | 09:38 |
chkumar|rover | on mixed rhel patches | 09:38 |
marios | chkumar|rover: thank you I saw but was working on the final bug for tempest first ... will start to work on those | 09:39 |
chkumar|rover | marios: thank you :-) | 09:39 |
ysandeep | marios, very creative of you to think out of the box here, deploying single controller in one stack and in second stack use controller of first stack.. I know about multi stack / DCN extended to edge , looks like you mixed both DCN and multi stack in a single approach for this CI job. | 09:41 |
ysandeep | afair.. In DCN deployment we use single overcloud stack | 09:42 |
ysandeep | ah.. no my memory failed me - IN dcn we use multiple stacks: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/16.1/html/distributed_compute_node_and_storage_deployment/assembly_deploy-edge-without-storage | 09:43 |
ysandeep | so basically this job is DCN deployment in CI with different base OS in central and edge site | 09:44 |
marios | ysandeep: 'creative' ;) yeah i was happy to find the multi stack stuff otherwise not sure how we would do this | 09:45 |
ysandeep | so we covered.. controller son 9 and computes on 8 scenario with this job.. Do we also need to cover 1 controller on 9 and other on 8 scenario :)? | 09:50 |
ysandeep | controller on* 9 | 09:50 |
marios | ysandeep: in future.. for mvp the scenario is specifically 9 control 8 compute | 09:52 |
ysandeep | okay, thanks for info o/ marios++ great work | 09:53 |
marios | thanks ysandeep | 09:54 |
ysandeep | intesting.. https://logserver.rdoproject.org/58/43558/13/check/tripleo-ci-centos-8-9-multinode-mixed-os/e621742/logs/subnode-1/etc/hosts.txt.gz don't have compute node entry but still everything worked.. not sure how it resolved compute node to its ip.. normally controller have entry for compute nodes in /etc/hosts. | 10:04 |
ysandeep | but may be things work differently in DCN(I don't have exp with dcn deployments) | 10:05 |
marios | ysandeep: well not quite 'everything worked' - we had to run nova cell discover_hosts after compute dpeloyment, for controller (nova_api) to discover the compute node and be able to finally boot vms | 10:06 |
marios | ysandeep: there https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/841764/29/playbooks/multinode-overcloud-mixed-os-deploy-compute.yml#67 | 10:07 |
ysandeep | ack, I see we have mentioned similiar steps here to connect to remote compute node | 10:20 |
ysandeep | https://opendev.org/openstack/tripleo-docs/src/branch/master/deploy-guide/source/deployment/standalone.rst#discover-the-remote-compute-node-from-the-central-controller-node | 10:20 |
marios | ysandeep: ha i wish i found that first :) i just saw what the 'normal' deployment would do with https://opendev.org/openstack/tripleo-heat-templates/src/commit/48deb4cbb53d187454c7de82e7125900e93926d1/deployment/nova/nova-compute-common-container-puppet.yaml#L59-L78 and recreated it | 10:21 |
marios | ysandeep: (https://bugs.launchpad.net/tripleo/+bug/1981459 ) | 10:21 |
ysandeep | :) I wonder if we had to do similiar step in dcn deployment but I don't find the same workaround in docs for dcn.. anyway +1 (I will revisit again) | 10:28 |
*** rlandy|out is now known as rlandy | 10:31 | |
rlandy | chkumar|rover: hey - how are things? | 10:32 |
chkumar|rover | rlandy: everything seems to be normal | 10:32 |
rlandy | want to sync or wait for dasm to come line? | 10:33 |
chkumar|rover | rlandy: rhos-16.2 container image update looks good | 10:33 |
chkumar|rover | nothing seen in the logs | 10:33 |
rlandy | chkumar|rover: yay | 10:33 |
chkumar|rover | rlandy: let's wait for dasm|off to come online then we can sync | 10:33 |
rlandy | chkumar|rover: going to rerun the two failed jobs in 16.2 | 10:34 |
rlandy | so we can promote the new base image | 10:34 |
chkumar|rover | rlandy: already re-runned those | 10:34 |
chkumar|rover | rlandy: https://hackmd.io/KrGAxws2QiaHuJmj8TH6Jg?view | 10:34 |
chkumar|rover | for all release failed jobs which seems not good have testproject running | 10:35 |
rlandy | chkumar|rover: https://code.engineering.redhat.com/gerrit/c/testproject/+/419956/2/.zuul.yaml are 17 reruns | 10:36 |
rlandy | we need 16.2 | 10:36 |
rlandy | the same testproject is linked in 16.2 and 17 lines | 10:37 |
rlandy | is there another one? | 10:37 |
rlandy | RH8 RHOS-16.2 Last promotion: 12th july | 10:37 |
rlandy | Re-running failed job: https://code.engineering.redhat.com/gerrit/c/testproject/+/419956 | 10:37 |
rlandy | RH8 RHOS-17 Last promotion: 13th july | 10:37 |
rlandy | Re-running failed job: https://code.engineering.redhat.com/gerrit/c/testproject/+/419956 | 10:37 |
rlandy | same job | 10:37 |
chkumar|rover | oh sorry | 10:40 |
chkumar|rover | i updated the same patchset | 10:40 |
chkumar|rover | let me re0run 16.2 one | 10:40 |
rlandy | great - thanks | 10:41 |
rlandy | chkumar|rover: dviroel|rover said using the bigger flavor for train did not help | 10:46 |
chkumar|rover | rlandy: I was looking at master fs035 | 10:46 |
rlandy | it may help for master if you want to try that out | 10:46 |
chkumar|rover | https://paste.centos.org/view/raw/09698f75 | 10:46 |
chkumar|rover | I picked one ipv6 test which is failing in 3 last runs | 10:46 |
chkumar|rover | trying to find out the issue | 10:46 |
rlandy | yep - would be good to try see what is going on | 10:47 |
rlandy | frenzy_friday: hello | 11:00 |
rlandy | frenzy_friday: how are things with 17.1? | 11:00 |
rlandy | chkumar|rover: +2'ed https://code.engineering.redhat.com/gerrit/c/openstack/tripleo-ci-internal-jobs/+/419819 - but needs rebase | 11:01 |
rlandy | frenzy_friday: reviewed your jobs | 11:01 |
rlandy | can we run them now? | 11:01 |
rlandy | frenzy_friday: all merged | 11:02 |
rlandy | let's run those today and get the hash moved along | 11:03 |
rlandy | ping if you need help | 11:03 |
chkumar|rover | rlandy: fixed merged conflict https://code.engineering.redhat.com/gerrit/c/openstack/tripleo-ci-internal-jobs/+/419819 needs your vote again | 11:07 |
rlandy | chkumar|rover: done | 11:10 |
chkumar|rover | rlandy: I am marking this bug https://bugs.launchpad.net/tripleo/+bug/1981634 as a duplicate of https://bugs.launchpad.net/tripleo/+bug/1981467 | 11:24 |
rlandy | will look in a sec - on review time | 11:24 |
frenzy_friday | rlandy, yep, running them now | 11:25 |
rlandy | frenzy_friday: thanks | 11:25 |
*** dviroel|rover is now known as dviroel | 11:28 | |
* dviroel not rover anymore | 11:29 | |
rlandy | frenzy_friday: need to fix ... | 11:47 |
frenzy_friday | rlandy, which one? | 11:47 |
rlandy | frenzy_friday: see error on job | 11:47 |
rlandy | getting | 11:47 |
frenzy_friday | oh, checking | 11:48 |
marios | food biab | 11:48 |
rlandy | The error was: {{ dlrnapi_protocol }}://{{ dlrnapi_host }}/{{ dlrn_baseurl[distro][osp_release|default(release)] }}: 'dict object' has no attribute 'rhos-17.1'\n\nThe error appears to be in '/var/lib/zuul/builds/7eaac3717da54e87a87529efe91eb758/trusted/project_4/review.rdoproject.org/rdo-infra/ci-config/ci-scripts/infra-setup/roles/get_hash/tasks/get_hash.yaml': line 16, column 3, but may\nbe elsewhere in the file depending on | 11:48 |
rlandy | the exact syntax problem.\n\nThe offending line appears to be:\n\n# commit/distro hashes for component pipeline (an unique commit hash for each component)\n- name: discover the hashes url (component)\n ^ here\n" | 11:48 |
rlandy | we need to add this option | 11:49 |
rlandy | frenzy_friday: https://github.com/rdo-infra/ci-config/blob/master/ci-scripts/infra-setup/roles/get_hash/defaults/main.yml | 11:50 |
rlandy | you need to add 17.1 options there | 11:50 |
frenzy_friday | oohh, ack, adding | 11:52 |
*** ysandeep is now known as ysandeep|afk | 11:56 | |
chkumar|rover | dviroel: it is just tracking the overcloud deployment failures https://bugs.launchpad.net/tripleo/+bug/1981478 na? | 11:57 |
dviroel | chkumar|rover: hi o/ - no, happens also in tempest tests, when you look into controller's errors, lots of db connections appears | 11:59 |
chkumar|rover | dviroel: ok, thanks I will track it here | 12:00 |
dviroel | chkumar|rover: yeah, if we identify a specific issue on one of the these jobs, we can create another one | 12:08 |
dviroel | but so far, everything seems related to mysql/network issues | 12:08 |
frenzy_friday | chkumar|rover, I did not get your comment on https://review.rdoproject.org/r/c/rdo-infra/ci-config/+/44022 . RHOS 17 on RHEL8 is already added right? | 12:09 |
frenzy_friday | Are we running rhosp 17.1 on RHEL8 as well? | 12:10 |
chkumar|rover | frenzy_friday: sorry wanted to add the comment for rhos-17.1 | 12:11 |
chkumar|rover | rlandy: in future is there any plan to run rhos-17.1 on rhel-8 also? | 12:11 |
* bhagyashris stepping out for bit | 12:26 | |
*** amoralej is now known as amoralej|lunch | 12:31 | |
jm1 | dviroel: thanks for your comments here :) https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/849945 | 12:38 |
jm1 | chkumar|rover: may want to have a look here? edited as requested ;) https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/849945 | 12:38 |
dviroel | jm1: thanks for trying the other approach | 12:42 |
chkumar|rover | jm1: thanks ;-) | 12:43 |
marios | interesting discussion in tc meet about stream/stability fyi https://meetings.opendev.org/meetings/tc/2022/tc.2022-07-14-15.00.log.html#l-31 | 12:44 |
marios | 15:21:00 <gmann> #agree to make centos stream jobs testing in periodic way but keep it in testing runtime. monitor, debug, and report the failure to centos stream team | 12:44 |
jm1 | marios: this one has 3x +2, maybe you could give it a +w? 🥺 https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/849945 | 12:48 |
marios | jm1: sure but i think it has enough votes you don't need more perhaps one of the folks that +2 can +w it ? | 12:50 |
marios | done | 12:50 |
marios | jm1: when you have time please review https://review.opendev.org/q/topic:oooci_mixed_rhel | 12:51 |
jm1 | marios: getting to your +2/+w is like reaching boss level 😉 thanks for merging it, now we can work on porting tripleo to lastest aoc :) | 12:56 |
marios | jm1: hah why i am probably the easiest person to give +2 ! | 12:56 |
marios | like here | 12:56 |
marios | please +w *this* | 12:56 |
* marios OK! | 12:56 | |
marios | ;D | 12:57 |
marios | jm1: but seriously i expect payment this is not free, so please add to your reviews list 15:51 < marios> jm1: when you have time please review https://review.opendev.org/q/topic:oooci_mixed_rhel | 12:57 |
reviewbot | Do you want me to add your patch to the Review list? Please type something like add to review list <your_patch> so that I can understand. Thanks. | 12:57 |
marios | jm1: otherwise you'll be hearing from my lawyer | 12:57 |
marios | good day sir | 12:57 |
jm1 | marios: yes, sir! 💂 | 12:58 |
chkumar|rover | rlandy: dasm|off train cs8 promoted | 13:11 |
rlandy | chkumar|rover: wow | 13:14 |
rlandy | chkumar|rover: did you skip? | 13:14 |
chkumar|rover | rlandy: nothing | 13:14 |
rlandy | oh - we used the internal | 13:14 |
rlandy | no - we didn't https://github.com/rdo-infra/ci-config/blob/master/ci-scripts/dlrnapi_promoter/config_environments/rdo/CentOS-8/train.yaml | 13:15 |
rlandy | wow | 13:15 |
chkumar|rover | I just re-runned the failed jobs, fs035 01 | 13:16 |
chkumar|rover | it passed | 13:16 |
rlandy | ttps://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-8-ovb-3ctlr_1comp-featureset001-train&pipeline=openstack-periodic-integration-stable4&skip=0 | 13:16 |
rlandy | wow | 13:16 |
marios | thank you for review chkumar|rover++ | 13:16 |
rlandy | never passed in a while | 13:16 |
marios | review*s* | 13:16 |
jm1 | marios: your patch series will take a while to review, it's massive 😅 | 13:20 |
rlandy | chkumar|rover: wow | 13:20 |
jm1 | marios: working on in | 13:20 |
jm1 | *it | 13:20 |
rlandy | dviroel: ^^ https://review.rdoproject.org/r/c/testproject/+/44018 - check that out | 13:20 |
rlandy | passing train fs001 | 13:20 |
rlandy | frenzy_friday: hey - looking better on the promote jobs? | 13:20 |
frenzy_friday | rlandy, are we planning to run rhosp 17.1 on rhel 8 in the future? | 13:21 |
dviroel | rlandy: chkumar|rover: this is impossible | 13:22 |
dviroel | lol | 13:22 |
*** amoralej|lunch is now known as amoralej | 13:22 | |
chkumar|rover | it's the mercy of CI | 13:24 |
dviroel | chkumar|rover: the ci respects you | 13:25 |
rlandy | frenzy_friday: ack | 13:25 |
* rlandy bows to chkumar|rover | 13:25 | |
rlandy | dviroel: we don't have the magic :( | 13:25 |
rlandy | frenzy_friday: pls add both options | 13:25 |
* chkumar|rover bows back | 13:26 | |
dviroel | rlandy: no, we don't, we are lucky to have chkumar|rover with us | 13:26 |
frenzy_friday | ack, rlandy, chkumar|rover updated https://review.rdoproject.org/r/c/rdo-infra/ci-config/+/44022 | 13:27 |
marios | jm1: yeah no worries and i understand a lot of it will be unkown but feel free to add questions there so it can be useful for you as learning (not about mypatch, but about our ci code/mechanisms more generally) | 13:27 |
marios | jm1: and you can always "just review the ansible" kind of thing maybe you can spot something to fix | 13:28 |
marios | thanks | 13:28 |
rlandy | frenzy_friday: looks good - pls try testproject with that depends on | 13:29 |
rlandy | let's see if we are missing anything else | 13:29 |
*** dasm|off is now known as dasm|ruck | 13:33 | |
dasm|ruck | o/ | 13:33 |
chkumar|rover | dasm|ruck: \o good morning :-) | 13:33 |
dviroel | \o | 13:33 |
dasm|ruck | chkumar|rover: o/ anything important you're fighting with right now? | 13:34 |
chkumar|rover | dasm|ruck: nothing looking at fs035 | 13:35 |
dasm|ruck | ack | 13:35 |
chkumar|rover | https://review.opendev.org/c/openstack/tripleo-quickstart/+/849988 test results | 13:35 |
dasm|ruck | chkumar|rover: cs8 wallaby looks like promoted too | 13:37 |
dasm|ruck | neat | 13:37 |
chkumar|rover | dasm|ruck: yes and train also | 13:37 |
chkumar|rover | just promoted right now | 13:37 |
dasm|ruck | ++ | 13:37 |
chkumar|rover | I think we can focus on master | 13:38 |
dasm|ruck | k | 13:38 |
rlandy | chkumar|rover: rekicked the rhos-17 patch | 13:39 |
rlandy | added failed envD for 16.2 | 13:40 |
rlandy | dasm|ruck: ^^ | 13:40 |
dasm|ruck | rlandy: "failed envD for 16.2" what do you mean? | 13:41 |
dasm|ruck | chkumar|rover: do you have cs9 master testproject running? | 13:41 |
chkumar|rover | it just failed | 13:41 |
dasm|ruck | chkumar|rover: if not, i'm gonna kick off one now for ebad96608af28084b7aee40526cf4760 | 13:41 |
chkumar|rover | I have rechecked soniya tp | 13:41 |
dasm|ruck | ah | 13:42 |
chkumar|rover | dasm|ruck: I think we can rekick a fresh tp | 13:42 |
dasm|ruck | this https://review.rdoproject.org/r/c/testproject/+/42692 ? | 13:42 |
dasm|ruck | k, i'm gonna start a new one | 13:42 |
chkumar|rover | https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-9-scenario010-ovn-provider-standalone-master | 13:42 |
chkumar|rover | dasm|ruck: give me 5 mins | 13:42 |
chkumar|rover | I think I can fix this job | 13:42 |
dasm|ruck | chkumar|rover: https://review.rdoproject.org/r/c/testproject/+/42374 | 13:43 |
dviroel | chkumar|rover: dasm|ruck: https://review.rdoproject.org/r/c/rdo-infra/ci-config/+/43689 - feel free to decide on merging or not, based on current results :P | 13:44 |
dasm|ruck | ok, updated the patch one more time -- it's gonna stop the test | 13:44 |
rlandy | dasm|ruck: https://code.engineering.redhat.com/gerrit/c/testproject/+/419956 | 13:44 |
dasm|ruck | rlandy: ack | 13:44 |
dasm|ruck | dviroel: ack | 13:44 |
chkumar|rover | bhagyashris: hello | 13:52 |
chkumar|rover | bhagyashris: https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-9-scenario010-ovn-provider-standalone-master is failing | 13:52 |
chkumar|rover | bhagyashris: I think https://github.com/rdo-infra/rdo-jobs/blob/master/zuul.d/standalone-jobs.yaml#L269 is wrong | 13:53 |
chkumar|rover | we need to add tags here https://opendev.org/openstack/tripleo-ci/src/branch/master/zuul.d/standalone-jobs.yaml#L724 just like this | 13:53 |
chkumar|rover | Do you know about this one? | 13:53 |
chkumar|rover | rlandy: dasm|ruck dviroel marios https://review.rdoproject.org/r/c/rdo-jobs/+/44024 please have a look when free, thanks! | 14:01 |
bhagyashris | chkumar|rover, hey when i tested it was passing https://review.rdoproject.org/r/c/testproject/+/39932/33#message-57afa997db1286cece3def5fa6c6e5830aaa4d05 | 14:02 |
dasm|ruck | chkumar|rover: we have separate trunks for those? | 14:02 |
dasm|ruck | yeah, seems like it | 14:03 |
bhagyashris | chkumar|rover, so looks like it may be different issue otherwise it should should that time only | 14:03 |
chkumar|rover | bhagyashris: let me send a testproject to confirm that | 14:04 |
bhagyashris | chkumar|rover, and sc010-ovn-provider is inheritated from sc010-standalone job and that job contains the tag here https://github.com/rdo-infra/rdo-jobs/blob/master/zuul.d/standalone-jobs.yaml#L211 | 14:05 |
marios | chkumar|rover: ok but do we realy need it? added a comment | 14:06 |
marios | chkumar|rover: i mean it should have tags there via inheritance already? | 14:06 |
chkumar|rover | bhagyashris: I think you are correct | 14:06 |
chkumar|rover | I missed that line | 14:06 |
chkumar|rover | marios: bhagyashris yes | 14:07 |
marios | chkumar|rover: k | 14:07 |
dasm|ruck | chkumar|rover: in the meantime, do you want me to chase a new hash for promition? | 14:10 |
chkumar|rover | dasm|ruck: yes please | 14:11 |
dasm|ruck | on it | 14:12 |
marios | rlandy: first green runs https://zuul.opendev.org/t/openstack/build/a3fbbdfba721457d8301e105164be2da (1 hr 54 mins 35 secs) & https://review.rdoproject.org/zuul/build/4205031ebc834dabbfb05757108143f5 (2 hrs 21 mins 3 secs) \o/ | 14:12 |
marios | that upstream is well within timeout but still gonna be our longer upstream job we have to be careful with files: or otherwise restrict it somehow | 14:13 |
marios | times in brackets wrong way aroudn but you get the jist ;) | 14:13 |
marios | interesting its like half an hour faster in rdo/vexx env | 14:14 |
rlandy | marios++ very nice | 14:15 |
chkumar|rover | dasm|ruck: we need a seperate tp for this periodic-tripleo-ci-centos-9-scenario010-kvm-internal-standalone-master | 14:16 |
chkumar|rover | in downstream | 14:16 |
*** ysandeep|afk is now known as ysandeep | 14:16 | |
ysandeep | chkumar|rover, rlandy marios dviroel mtg time | 14:16 |
dasm|ruck | chkumar|rover: started: https://code.engineering.redhat.com/gerrit/c/testproject/+/408119 | 14:17 |
chkumar|rover | dasm|ruck: cool | 14:17 |
jm1 | rcastillo, rlandy, chkumar|rover: quick! stop this merge!!! https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/847419 | 14:25 |
jm1 | ysandeep, marios: ^ | 14:25 |
rlandy | jm1; abondon the patch | 14:25 |
jm1 | rlandy: thanks! | 14:25 |
rlandy | jm1; and w-1 it if you don;t want it merged | 14:26 |
jm1 | rlandy: i dont have permissions | 14:26 |
dasm|ruck | it's rcastillo's patch. rcastillo needs to -w it | 14:26 |
rlandy | any core can | 14:26 |
jm1 | rlandy: please abandon, i am not core there | 14:27 |
rlandy | jm1: done | 14:28 |
jm1 | rlandy: thank you very much. that was close 😅 | 14:28 |
frenzy_friday | https://code.engineering.redhat.com/gerrit/c/testproject/+/420019 failed again with the same error | 14:33 |
frenzy_friday | Maybe the depends on didnt work | 14:33 |
marios | ysandeep: reminder you're sharing screen # | 14:37 |
marios | ysandeep: don't share you email secrets | 14:37 |
ysandeep | no secrets gere, finding one draw.io from chandan | 14:38 |
marios | ;) | 14:38 |
dasm|ruck | btw centos breaking openstack: https://lists.openstack.org/pipermail/openstack-discuss/2022-July/029468.html cc rlandy | 14:40 |
marios | jm1: sorry in call but i see it is abandoned now | 14:51 |
jm1 | marios: no worry, rlandy abandoned it. i just pinged all because it had to be done quick | 14:52 |
rlandy | frenzy_friday: ERROR: Got error 404, probably endpoint https://osp-trunk.hosted.upshift.rdu2.redhat.com/api-rhel9-osp17-1 is not available | 14:53 |
rlandy | frenzy_friday: pls check with release -delivery on that | 14:54 |
frenzy_friday | rlandy, ack | 15:05 |
rlandy | frenzy_friday: jon should be able to confirm that | 15:05 |
frenzy_friday | rlandy, I see https://osp-trunk.hosted.upshift.rdu2.redhat.com/api-rhel9-osp17 returns 404 as well. Am I using a wrong url? | 15:09 |
rlandy | frenzy_friday: doens't mean that the endpoint is working | 15:10 |
rlandy | correct | 15:10 |
rlandy | but two diff things | 15:10 |
frenzy_friday | no, I mean foe 16.2, 17 - for all of them I am getting 404 | 15:10 |
rlandy | checking | 15:11 |
chkumar|rover | dasm|ruck: https://bugs.launchpad.net/tripleo/+bug/1981829 | 15:12 |
chkumar|rover | dasm|ruck: added it to cix, if you get more info about this bug please add | 15:12 |
rlandy | https://sf.hosted.upshift.rdu2.redhat.com/logs/openstack-component-network/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-rhel-8-rhos-17-component-network-promote-consistent-to-component-ci-testing/f05645b/job-output.txt | 15:12 |
rlandy | frenzy_friday: ^^ works there | 15:12 |
dasm|ruck | chkumar|rover: ack | 15:12 |
frenzy_friday | rlandy, are you able to access https://osp-trunk.hosted.upshift.rdu2.redhat.com/api-rhel8-osp17 from the browser? | 15:13 |
rlandy | frenzy_friday: try hit with with api | 15:14 |
chkumar|rover | gthiemonge: hello, please have a look at this bug https://bugs.launchpad.net/tripleo/+bug/1981829 when around, I have added as a promotion blocker | 15:14 |
chkumar|rover | thanks! | 15:14 |
chkumar|rover | dasm|ruck: rlandy everything is on hackmd, see ya on Monday! | 15:15 |
rlandy | frenzy_friday: they are fixing it | 15:15 |
dasm|ruck | chkumar|rover: thanks, see ya | 15:15 |
dviroel | chkumar|rover: happy fridayo/ | 15:16 |
* dviroel lunch | 15:16 | |
frenzy_friday | yep | 15:16 |
rlandy | chkumar|rover: have a good weekend | 15:16 |
*** dviroel is now known as dviroel|lunch | 15:16 | |
rlandy | frenzy_friday: pls try again | 15:16 |
frenzy_friday | I am waiting for the patch from amoralej to merge | 15:17 |
rlandy | ack - thanks | 15:20 |
*** ysandeep is now known as ysandeep|out | 15:28 | |
* rlandy needs to pick up package - brb | 15:42 | |
*** rlandy is now known as rlandy|brb | 15:42 | |
marios | happy weekend tripleo-ci o/ | 15:57 |
*** marios is now known as marios|out | 15:57 | |
dasm|ruck | marios|out: o/ | 16:04 |
*** rlandy|brb is now known as rlandy | 16:05 | |
rlandy | dasm|ruck: you all set? | 16:05 |
dasm|ruck | yup, chasing cs9 master | 16:05 |
dasm|ruck | seems like we might have a valid kvm issue. but still verifying | 16:06 |
*** dviroel|lunch is now known as dviroel | 16:32 | |
*** amoralej is now known as amoralej|off | 16:53 | |
rlandy | lunch brb | 16:59 |
dasm|ruck | k | 17:01 |
dasm|ruck | we have two issues with periodic-tripleo-ci-centos-9-scenario010-kvm-internal-standalone-master | 17:10 |
dasm|ruck | > octavia_tempest_plugin.tests.scenario.v2.test_pool.PoolScenarioTest)testtools.testresult | 17:10 |
dasm|ruck | > (octavia_tempest_plugin.tests.scenario.v2.test_traffic_ops.TrafficOperationsScenarioTest)testtools.testresult | 17:10 |
dasm|ruck | checking notes if we've noticed that already | 17:11 |
dasm|ruck | hmm... seems that we can have something already: https://bugs.launchpad.net/tripleo/+bug/1979546 | 17:12 |
dviroel | dasm|ruck: there is a CIX for this issue | 17:12 |
dviroel | dasm|ruck: the KVM error one | 17:13 |
dviroel | " show_loadbalancer provisioning_status updated to an invalid state of ERROR" | 17:13 |
dasm|ruck | last time it passed 2022-07-05 and it's continuously failing since then | 17:13 |
dasm|ruck | dviroel: i believe that might be either nova or kvm issue, not octavia | 17:13 |
dviroel | dasm|ruck: yes, there are more info in the BZs | 17:14 |
dasm|ruck | k | 17:14 |
dasm|ruck | > 2022-07-15 16:34:00.420+0000: 191452: info : qemuMonitorJSONIOProcessLine:218 : QEMU_MONITOR_RECV_REPLY: mon=0x7fa7800852f0 reply={"id": "libvirt-48", "error": {"class": "GenericError", "desc": "SEV is not enabled in KVM"}} | 17:14 |
dasm|ruck | hmm | 17:15 |
dasm|ruck | rlandy: when did you enable security enabled VMs for sbaker? | 17:15 |
dasm|ruck | dviroel: but you're talking about your 16.2 bz? | 17:16 |
dviroel | dasm|ruck: it was logged against it | 17:17 |
dviroel | check instance log on logs/undercloud/var/log/containers/libvirt/qemu/ | 17:17 |
dviroel | "KVM: entry failed, hardware error 0x8" | 17:17 |
dviroel | same errkr | 17:17 |
dasm|ruck | ack | 17:18 |
dasm|ruck | i'm checking this: https://sf.hosted.upshift.rdu2.redhat.com/logs/19/408119/11/check/periodic-tripleo-ci-centos-9-scenario010-kvm-internal-standalone-master/ca7e7c3/logs/undercloud/var/log/containers/libvirt/virtqemud.log | 17:18 |
dasm|ruck | yeah, i see the KVM error | 17:18 |
dasm|ruck | wait a sec... it's cs9 sc10 kvm master | 17:20 |
dasm|ruck | it's not rhos at all, it just runs on internal infra. | 17:20 |
rlandy | dasm|ruck: I uploaded the image a few days ago | 17:38 |
dasm|ruck | ack | 17:41 |
dasm|ruck | i added this: https://review.rdoproject.org/r/c/rdo-infra/ci-config/+/44029 | 17:41 |
dasm|ruck | rlandy: dviroel can you check? ^ | 17:41 |
dviroel | dasm|ruck: ack, there is no other way, just kvm missing? | 17:44 |
dasm|ruck | dviroel: it's not the only one atm, but it's definitely gonna fail until KVM bug is fixed | 17:45 |
dasm|ruck | i'm checking other missing jobs | 17:45 |
dviroel | ack | 17:45 |
dasm|ruck | that's the test run for cs9 master: https://review.rdoproject.org/r/c/testproject/+/42374 | 17:45 |
rlandy | looking | 17:46 |
dasm|ruck | i suu cs9 main integration is running right now | 17:47 |
rlandy | https://sf.hosted.upshift.rdu2.redhat.com/zuul/t/tripleo-ci-internal/builds?job_name=periodic-tripleo-ci-centos-9-scenario010-kvm-internal-standalone-master&skip=0 | 17:47 |
dasm|ruck | yes | 17:47 |
rlandy | dasm|ruck: is that is the only job missing - ok | 17:48 |
rlandy | pls revert after promo | 17:48 |
rlandy | done | 17:49 |
dasm|ruck | k, thx | 17:49 |
dasm|ruck | ok... i'm shocked: https://sf.hosted.upshift.rdu2.redhat.com/zuul/t/tripleo-ci-internal/build/e0f2f5fee5df461c8b5aa3f67706ecd6 is green | 18:06 |
dasm|ruck | it's so inconclusive... | 18:06 |
dviroel | lol | 18:21 |
dasm|ruck | it is not funny :P | 18:21 |
rlandy | not as funny as train fs001 passing | 18:24 |
rlandy | dasm|ruck: ask chkumar|rover to run it - he has the magic touch | 18:24 |
rlandy | dviroel and I are no good | 18:24 |
dviroel | exactly | 18:24 |
dasm|ruck | hmm... maybe it means noisy neighbors? | 18:25 |
dasm|ruck | if it's working early India TZ? | 18:25 |
dasm|ruck | idk, just speculating :) | 18:25 |
dviroel | https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-8-ovb-3ctlr_1comp-featureset001-train | 18:25 |
dviroel | sonyia was trying the same testproject in india tz | 18:25 |
dasm|ruck | hmm | 18:26 |
dviroel | worked only with chandan | 18:26 |
rlandy | he spins some charm | 18:40 |
dasm|ruck | cs8 train & wallaby are nicely promoted. cs9 wallaby and master aren't so nice | 18:44 |
dviroel | nice | 18:54 |
rlandy | frenzy_friday: you're good to go now | 18:54 |
rlandy | 17.1 endpoint is functional | 18:54 |
rlandy | promote jobs are running | 18:55 |
dasm|ruck | nice | 18:55 |
rlandy | dasm|ruck: we need to start tracking the OVB investigation | 18:55 |
rlandy | need to deal with the ceph and cinder issue on internal now | 18:56 |
rlandy | but we need to start collating all this to stake steps one way or another | 18:56 |
dasm|ruck | i haven't started doing anything yet. | 18:56 |
dasm|ruck | rlandy: i've never integrated anything. is it something similar to enabling "nodepool"? | 18:57 |
rlandy | integrated? | 19:11 |
dasm|ruck | enabled new cloud | 19:22 |
jm1 | have a nice weekend, team 🍻 | 19:43 |
dviroel | o/ | 19:43 |
dasm|ruck | jm1: o/ | 19:47 |
rlandy | dasm|ruck: you need help with ibm cloud? | 20:07 |
rlandy | dasm|ruck: master is not promoting | 20:08 |
* dviroel biab | 20:09 | |
rlandy | what hash were you chasing | 20:09 |
*** dviroel is now known as dviroel|biab | 20:09 | |
dasm|ruck | rlandy: it's having its issue | 20:19 |
dasm|ruck | rlandy: https://review.rdoproject.org/r/c/testproject/+/42374 | 20:20 |
dasm|ruck | 55921623476163598a5dba0658e34aa3 | 20:20 |
dasm|ruck | oh, pipeline must've finished | 20:20 |
dasm|ruck | oh no. it just moved in different place | 20:21 |
dasm|ruck | that's one of issues with cs9 master: https://bugs.launchpad.net/tripleo/+bug/1981829 | 20:22 |
dasm|ruck | cs9 wallaby also is not happy: https://review.rdoproject.org/r/c/testproject/+/41367 | 20:24 |
* rlandy looks | 20:25 | |
rlandy | dasm|ruck: you said it was the only missing job | 20:26 |
rlandy | what hash did you check for that???: | 20:26 |
rlandy | master ran again | 20:26 |
dasm|ruck | no, i said it's one job which was definitely gonna fail, because of ongoin kvm problem | 20:26 |
rlandy | what was the previous hash? | 20:26 |
dasm|ruck | previous: ebad96608af28084b7aee40526cf4760 | 20:27 |
rlandy | so no - then we can't skip until we know it's a good hash for everything else | 20:27 |
dasm|ruck | hmm | 20:27 |
dasm|ruck | even if kvm won't work at all? | 20:27 |
dasm|ruck | i didn't want to skip it just for promo | 20:27 |
dasm|ruck | my intention was to disable it from promo for time being until there's gonna be fix | 20:28 |
rlandy | well then we need to keep skipping and unskipping | 20:28 |
rlandy | no | 20:28 |
dasm|ruck | i didn't want to go back and forth with enabling/disabling | 20:28 |
rlandy | we need to keep that file clean | 20:28 |
rlandy | and just skip and unskip as needed | 20:28 |
dasm|ruck | it's broken by design. | 20:29 |
rlandy | <rlandy> dasm|ruck: is that is the only job missing - ok | 20:29 |
rlandy | <rlandy> pls revert after promo | 20:29 |
dasm|ruck | rlandy: and i responded "no", it was not the only one missing | 20:29 |
dasm|ruck | rlandy: revert: https://review.rdoproject.org/r/c/rdo-infra/ci-config/+/43690 | 20:30 |
rlandy | maybe the current hash will be better | 20:30 |
rlandy | running now | 20:30 |
dasm|ruck | current hash is still computing. promo line is in progresss | 20:31 |
rlandy | ack | 20:32 |
rlandy | still working | 20:32 |
dasm|ruck | rlandy: this can be reverted: https://review.rdoproject.org/r/c/rdo-infra/ci-config/+/43689 | 20:38 |
rlandy | done | 20:39 |
dasm|ruck | k, thx | 20:39 |
rlandy | sorry - just need to get the ceph patches in for 17 | 20:43 |
dasm|ruck | no worries | 20:45 |
rlandy | dasm|ruck: ok- new master hash looks better | 21:06 |
rlandy | let's see if we can get a pass on rerun of those failures | 21:06 |
dasm|ruck | still 5 broken, including ovn. | 21:06 |
dasm|ruck | i'm gonna rerun them in a sec, just about to wrap up one, currently running hash | 21:06 |
dasm|ruck | https://review.rdoproject.org/zuul/status/change/42374,27 | 21:07 |
dasm|ruck | it passed, just waiting for results to being posted | 21:07 |
*** dviroel|biab is now known as dviroel | 21:08 | |
dasm|ruck | chkumar|rover: sc010 ovn across most (if not all) releases is being affected by KVM issue. I left note in the hackmd | 21:20 |
rlandy | k - one less to rerun | 21:20 |
dasm|ruck | rlandy: kvm is such a pain right now ^ | 21:21 |
dasm|ruck | or, more detailed: sc010 | 21:21 |
rlandy | https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-9-scenario010-ovn-provider-standalone-master&skip=0 | 21:21 |
rlandy | dasm|ruck'; ^^ new failure | 21:21 |
rlandy | was working cleanly a few days ago | 21:21 |
dasm|ruck | not new | 21:21 |
dasm|ruck | ond one | 21:22 |
rlandy | real issue | 21:22 |
dasm|ruck | *old | 21:22 |
dasm|ruck | there is a bug on that | 21:22 |
dasm|ruck | https://bugs.launchpad.net/tripleo/+bug/1981829 | 21:22 |
rlandy | yeah 0 but the last 3 days? | 21:22 |
rlandy | seems weird | 21:22 |
rlandy | was clean for two weeks before that | 21:22 |
dasm|ruck | under the hood kvm is failing | 21:23 |
dasm|ruck | oh wait. i'm not sure if i'm not mixing two things right now | 21:23 |
rlandy | Details: {'faultcode': 'Client', 'faultstring': 'Policy does not allow this request to be performed.', 'debuginfo': None} | 21:23 |
rlandy | ^^ different error | 21:23 |
dasm|ruck | right. i was thinking about kvm one | 21:24 |
rlandy | periodic-tripleo-ci-centos-9-scenario010-kvm-standalone-masteropenstack/tripleo-cimasteropenstack-periodic-integration-main1 hr 41 mins 54 secs2022-07-15 17:17:58SUCCESS | 21:24 |
dasm|ruck | but it's submitted already | 21:24 |
rlandy | dasm|ruck: ^^ passes cleanly | 21:24 |
dasm|ruck | the one on psi does not | 21:25 |
dasm|ruck | chkumar|rover asked me to kick off one there, and it failed | 21:25 |
rlandy | ack | 21:25 |
rlandy | dasm: but the error is not a poicy issue | 21:25 |
rlandy | looks different to me | 21:25 |
dasm|ruck | this: https://bugs.launchpad.net/tripleo/+bug/1981829 | 21:25 |
dasm|ruck | you're right. both are different. one is for kvm, another for policy | 21:26 |
rlandy | Details: (TrafficOperationsScenarioTest:setUpClass) show_server status failed to update to ACTIVE within the required time 300. Current status of show_server: BUILD | 21:26 |
rlandy | dasm|ruck: pls log a bug for the policy one | 21:27 |
rlandy | it's new | 21:27 |
rlandy | the old bug is the active state one | 21:27 |
dasm|ruck | 21:22:22 dasm|ruck | https://bugs.launchpad.net/tripleo/+bug/1981829 | 21:27 |
dasm|ruck | it's already logged | 21:27 |
dasm|ruck | what do you mean by "active state one"? | 21:27 |
dasm|ruck | oh, lovely | 21:30 |
dasm|ruck | > periodic-tripleo-ci-rhel-8-ovb-3ctlr_1comp-featureset001-internal-clients-rhos-17 https://sf.hosted.upshift.rdu2.redhat.com/zuul/t/tripleo-ci-internal/build/ec7883b754ba499688896269fddd75aa : NODE_FAILURE in 0s | 21:30 |
rlandy | dasm|ruck: the previous hash is in rerun | 21:32 |
rlandy | let's see if that one passes | 21:32 |
dasm|ruck | k | 21:32 |
rlandy | at tempest test stage | 21:32 |
rlandy | for fs001 17 on 8 | 21:32 |
rlandy | dasm|ruck: I rekicked the jenkins job for 16.2 | 21:33 |
rlandy | missing from promo | 21:33 |
rlandy | 17 on 9 - no new hash | 21:33 |
dasm|ruck | 17 on 9 is up to date | 21:33 |
rlandy | trying to figure out the new ceph stuff | 21:33 |
dasm|ruck | i'm slowly disconnecting. mixing cases. | 21:34 |
dasm|ruck | i'm calling it a day. | 21:34 |
dviroel | have a great weekend team | 21:34 |
dviroel | o/ | 21:34 |
dasm|ruck | chkumar|rover: i tried to clear components for all outstanding releases. Links are added in the hackmd | 21:34 |
dasm|ruck | dviroel: take care o/ | 21:34 |
dasm|ruck | chkumar|rover: the only outliers are cs9 wallaby and master. it's mainly due to kvm job. i'm rekicking promo hash | 21:35 |
dasm|ruck | chkumar|rover: i hope it's gonna be clearer by Morning | 21:36 |
* dasm|ruck => offline | 21:36 | |
dasm|ruck | have a good one! | 21:36 |
dasm|ruck | rlandy have a good weekend! | 21:37 |
*** dasm|ruck is now known as dasm|off | 21:37 | |
*** dviroel is now known as dviroel|out | 21:37 | |
rlandy | dasm|off: you too | 21:38 |
rcastillo | leaving as well | 22:28 |
rcastillo | have a good weekend | 22:28 |
rlandy | bye all | 22:50 |
Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!