Tuesday, 2022-04-19

*** chandankumar is now known as chkumar|ruck04:47
*** chkumar|ruck is now known as chandankumar04:51
*** ysandeep|out is now known as ysandeep05:49
*** ysandeep is now known as ysandeep|lunch07:29
jrossermorning07:39
*** arxcruz|out is now known as arxcruz08:11
johnnydHey there, I'm currently facing a problem with the OS-ansible inventory. We have some inconsistencies in it. More precisely, the services cinder-volumes is declared in 3 lxc-containers (this is normal, we have 3 controllers) but also on 2 of the 3 controllers (so it's also installed in bare metal). If we delete the inventory and regenerate it, this disappear but obviously we generate something incorrect because IP and hostnames of every lxc 08:22
johnnydSo, the question is, how to clean this properly except removing it by hand ? Thanks :)08:22
damiandabrowski[m]hey jrosser!08:28
damiandabrowski[m]johnnyd: I'm not 100% sure, but i think there's no good way to remove host from the inventory. IIRC the `scripts/inventory-manage.py` will remove it only from openstack_inventory.json, but removed host will still exist in openstack_hostnames_ips.yml08:29
johnnydtypically, this is normal: https://pastebin.com/raw/pxZhcQ4t08:32
johnnydand this is not: https://pastebin.com/raw/ACcaqpsY08:33
johnnydI don't really want to remove the os-controller-3 declaration because we need it for HAproxy, just remove the cinder_backends part and storage address08:35
noonedeadpunkmornings08:42
jrosserjohnnyd: keep /etc/openstack_deploy under git then you can use a combination of inventory_manage.py and some grep/hand editing with the ability to roll back if it goes wrong08:48
noonedeadpunkjust use ansible-vault for encrypting user_secrets.yml :)08:51
johnnydyup, we already have our setup with git :)08:52
johnnydthat was our intuition, we'll edit that by hand08:53
jrossercinder volume service is optionally on metal nodes for iscsi, as that doesnt work in containers08:53
johnnydactually, we are using a Ceph backend08:54
jrossernoonedeadpunk: this is a wierd error https://paste.opendev.org/show/bIcy9AiNGEjvlqbCKh6K/08:59
jrosseri was sanity checking the gluster stuff by trying to mount from the repo container into the glance container, just to make sure all the systemd_mount stuff is robust08:59
jrosserand it seems that systemd_mount calls systemd_service, which has conditionals on vars which only exist in systemd_mount.....09:00
noonedeadpunksounds like systemd_mounts defined wierdly?09:14
jrosseryes - i'm looking at that now09:15
jrosserpretty hard to debug09:15
noonedeadpunkmaybe ansible_hostname is undefined?09:15
noonedeadpunklike stale facts or smth?09:15
noonedeadpunkah, but it fails in glance?09:19
jrosserwith that little test playbook, yes09:20
jrosserjust using glance as an empty container which is not a repo server09:20
jrosserto test that the glusterfs client works with just systemd_mount09:21
jrosseroh i see why09:26
jrosseri have https://paste.opendev.org/show/bkis6o2S20aA9hx7xOPz/ in /etc/ansible/roles/systemd_mount/tasks/systemd_install.yml09:27
jrosserso thats a block: with a when:, which calls a role that runs a handler09:27
jrosserand the handler gets the when: condition applied as part of import_role09:28
noonedeadpunkbrrrrr09:36
noonedeadpunkThen at this level variable shouldn't be as `systemd_mount_types` at all, since it's getting defined only inside role?09:37
jrosseri think thats right - i can't think of a clean way to do this except set_fact something for the when: in a previous task09:37
jrosseras systemd_mount role vars won't be available inside systemd_service09:38
jrosserhuh well set_fact: `_configure_glusterfs: "{{ 'glusterfs' in systemd_mount_types }}"` makes it work straight away09:41
jrosserseems that the root cause is lazy vs. immediate evaluation and there are lots of ansible issues related to that incl https://github.com/ansible/ansible/issues/1037409:48
noonedeadpunkwell.... yeah...09:50
*** ysandeep|lunch is now known as ysandeep09:50
jrosserits easy to forget that block: and include_<> with when: will apply the when condition everywhere09:50
jrosserusually it's OK but for including roles inside roles its tricky09:51
opendevreviewJonathan Rosser proposed openstack/ansible-role-systemd_mount master: Add support for glusterfs mounts  https://review.opendev.org/c/openstack/ansible-role-systemd_mount/+/83775010:28
SiavashSardarimorning10:43
SiavashSardariwould someone please help me with https://zuul.opendev.org/t/openstack/build/b1b5b45fa02c4bf180a68f2da11bb127/10:43
noonedeadpunkfolks, can we please review backports to stable branches? https://review.opendev.org/q/parentproject:openstack/openstack-ansible+branch:%255Estable/.*+status:open+,2510:45
noonedeadpunkSiavashSardari: huh, that's interesting....10:47
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Add ability to define bridge type for containers  https://review.opendev.org/c/openstack/openstack-ansible/+/83774210:47
SiavashSardari@noonedeadpunk I checked the parent commit. it was after the CI issue got merged. have I missed something?10:50
noonedeadpunkWell it's not related to commit itself, it's more likely some packaging issue or some braking change has landed...10:51
SiavashSardariwell thanks10:55
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Cleanup distro vars  https://review.opendev.org/c/openstack/openstack-ansible/+/83784510:56
noonedeadpunkSiavashSardari: I will try to check what's wrong if it will fail again10:57
jrosser[Errno 2] No such file or directory: b'ceph-volume'" suggests some upstream change there11:04
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-lxc_hosts master: Ensure systemd-udev is present in centos containers  https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/83786511:07
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-repo_server master: Add facility to store repo contents on a remote mount  https://review.opendev.org/c/openstack/openstack-ansible-repo_server/+/83758711:10
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-repo_server master: Clean up legacy lsycnd, rsync and ssh key config  https://review.opendev.org/c/openstack/openstack-ansible-repo_server/+/83785911:11
opendevreviewMerged openstack/ansible-role-python_venv_build stable/stein: Remove tripleo-ci-centos-7-standalone-os-tempest from gate layout  https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/83817011:19
*** dviroel|out is now known as dviroel11:31
*** ysandeep is now known as ysandeep|afk11:41
SiavashSardarithere is an issue regarding the default_bind_mount_logs variable in osa, the default value is true, which cause the lxc config file to be like https://paste.opendev.org/show/bvUwH0C2pxAYg2M7ejt4/.12:04
SiavashSardarithe issue is that there is two mount entry for /var/log and mostly it doesn't create an issue till first restart. after a restart there would be a race condition between which dir gets mounted first.12:04
opendevreviewSiavash Sardari proposed openstack/openstack-ansible master: Disable the default_bind_mount_logs  https://review.opendev.org/c/openstack/openstack-ansible/+/83850112:05
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-repo_server master: Remove all code for lsync, rsync and ssh  https://review.opendev.org/c/openstack/openstack-ansible-repo_server/+/83758812:06
SiavashSardariI believe removing the extra mount entry requires restarting the container if the /var/log entry has won the race.12:07
jrosserSiavashSardari: i think i remember something like this before12:08
jrosserthere were some patches for a similar problem12:08
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-plugins master: Add role for creating simple glusterfs servers  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/83758212:15
SiavashSardari@jrosser I will search the lauchpad. I'd appreciate if you could help me find the patches you mentioned12:18
opendevreviewJonathan Rosser proposed openstack/openstack-ansible master: Use glusterfs to synchronise repo server contents  https://review.opendev.org/c/openstack/openstack-ansible/+/83758912:19
opendevreviewMerged openstack/openstack-ansible-rabbitmq_server stable/xena: Verify if hosts file already managed with OSA  https://review.opendev.org/c/openstack/openstack-ansible-rabbitmq_server/+/83616612:21
jrosserSiavashSardari: is this related to your issue? https://review.opendev.org/c/openstack/openstack-ansible/+/75205912:22
SiavashSardarithis is related 12:39
opendevreviewMerged openstack/openstack-ansible-os_cinder stable/xena: Add the ability to disable send_actions option in cinder-volume-usage-audit service.  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/83821012:40
SiavashSardari@jrosser but the I believe the patch you mentioned does not solve the issue. there is always one of the mount entries work12:42
SiavashSardarilet me restart one container and send you the log.12:42
opendevreviewMerged openstack/openstack-ansible-os_cinder stable/wallaby: Add the ability to disable send_actions option in cinder-volume-usage-audit service.  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/83821112:43
SiavashSardarihttps://paste.opendev.org/show/boifQ4Qg3WhVRtztTuwK/12:46
noonedeadpunkwell, that worked for me at least, as I see mounts ordered properly now as well as journals from containers be preset on metal hosts12:51
noonedeadpunkbut more interesting would be to see your lxc config12:51
SiavashSardari@noonedeadpunk please take a look at this https://paste.opendev.org/show/bVbQK7AMkWnvNPao4zDD/12:53
*** ysandeep|afk is now known as ysandeep12:53
SiavashSardarithis is my lxc config https://paste.opendev.org/show/b6YjEnL0Ue6MMkkDulsu/12:54
noonedeadpunkare you sure you have patch mentioned?12:56
SiavashSardariyep i do have that12:59
*** spotz__ is now known as spotz13:12
opendevreviewMerged openstack/openstack-ansible stable/ussuri: Mark OSA repository as safe in git.config in CI  https://review.opendev.org/c/openstack/openstack-ansible/+/83802113:23
noonedeadpunkThinking about that, I'm not sure that that patch would re-order things correctly. So it's likely doing right things only on freshly spawned containers13:30
noonedeadpunkAs once line is present, task will do nothing I guess13:30
*** spotz is now known as spotz__13:54
*** spotz__ is now known as spotz13:54
SiavashSardari@noonedeadpuck exactly. actually now that I'm looking at things more closely, having /var/log mounted on host is usefull in case of some services like rabbit. I'm not sure but I think the root cause is that /var/log is the parent of /var/log/journal and mounting one would cause problem for mounting the other one14:22
SiavashSardarilet me try something on my dev. I will keep you posted14:23
*** dviroel is now known as dviroel|afk14:24
noonedeadpunkSiavashSardari: it's just ordering issue in the config14:30
noonedeadpunkif you would deploy from scratch you would have /var/log mounted and /var/log/journal inisde it which prefectly makes sense 14:31
SiavashSardariyep you're right. i tried to regenerate the issue with my hypothesis but no luck yet14:35
noonedeadpunkI should have wroted a release note for that commit to handle upgrade path... or well, abscence of it 14:36
noonedeadpunkbut if you do OS upgrade, you will have container re-generated and issue fixed :p14:36
SiavashSardarido we support ubuntu 22 yet? :D14:37
noonedeadpunknope...14:37
noonedeadpunkit hasn't even released yet :)14:37
noonedeadpunkdoes it?14:38
noonedeadpunkand don't re-ask in 2 days ! :p14:38
SiavashSardarino but maybe we can start with nightly builds :))14:38
*** dviroel|afk is now known as dviroel|mtg14:39
SiavashSardariEureka Eureka finally I could regenerate the issue14:42
SiavashSardariI believe the order of the mount point is wrong14:42
SiavashSardarithe /var/log should be before /var/log/journal.14:43
SiavashSardariat least my test shows that.14:43
*** ysandeep- is now known as ysandeep|out14:45
SiavashSardarimounting parent dir before the child scenario https://paste.opendev.org/show/bbeVnxOeEisyFzizsVVr/14:47
SiavashSardarimounting the child dir before parent scenario https://paste.opendev.org/show/bbeVnxOeEisyFzizsVVr/14:49
noonedeadpunk`the /var/log should be before /var/log/journal.` yes and that excatly what happens for me on all envs...14:49
noonedeadpunkand that actually what changes patch we're talking about14:50
noonedeadpunkBUT, if container already exist with wrong ordering, it won't fix it14:50
*** dviroel|mtg is now known as dviroel14:54
SiavashSardarioh I get it now it has the wrong order from the beginning.14:54
SiavashSardari@noonedeadpunk ci job failed for https://review.opendev.org/c/openstack/openstack-ansible/+/837742 again.14:57
spatelUbuntu 22 - Release date is April 21, 202214:59
spatelnoonedeadpunk i am interested in Ubuntu 22 so hope we can get it setup on CI-CD asap after release :) 15:00
noonedeadpunk#startmeeting openstack_ansible_meeting15:00
opendevmeetMeeting started Tue Apr 19 15:00:38 2022 UTC and is due to finish in 60 minutes.  The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot.15:00
opendevmeetUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.15:00
opendevmeetThe meeting name has been set to 'openstack_ansible_meeting'15:00
noonedeadpunk#topic rollcall15:00
spatel\o/15:00
damiandabrowski[m]hey!15:01
noonedeadpunko/15:01
mgariepyhey ! o/15:01
jrosserhello15:01
noonedeadpunk#topic office hours15:02
damiandabrowski[m]and thanks for nominating me! (i just saw the message on ML)15:02
noonedeadpunkso. seems we have issue with ceph-ansible version... Will need to take a look at the actuall thing first time tomorrow in the morning....15:03
noonedeadpunkbut that sounds really weird, I havn't heard a thing that they're replacing cinder-volume with anything else...15:04
noonedeadpunk*ceph-volume15:04
damiandabrowski[m]FYI, i'm not sure if it's related, but I've hit this error when deploying AIO+ceph today https://paste.openstack.org/show/bRMsII6P2s1YyjugFuWk/15:04
damiandabrowski[m]but haven't looked much into this yet15:05
noonedeadpunkthat's interesting actually.15:05
noonedeadpunkI can recall we were having apt pinning for ceph somehow....15:05
jrosserwe should have apt pins to make sure that the ceph version comes from where we expect15:05
jrosserwhich iirc on ubuntu should always be download.ceph.com15:06
noonedeadpunkbut I think we did that only for ceph-client role?15:06
jrossercould be, yes15:06
*** dviroel is now known as dviroel|lunch15:06
jrosserhttps://github.com/openstack/openstack-ansible-ceph_client/blob/master/defaults/main.yml#L4315:08
noonedeadpunkhuh and we don't really include role anywhere15:08
noonedeadpunkunless we rely on that? https://opendev.org/openstack/openstack-ansible-ceph_client/src/branch/master/meta/main.yml#L40-L4515:09
jrosseryes it comes from meta15:09
noonedeadpunkbut, um....15:09
noonedeadpunkbut anyway failure is during ceph-ansible runtime15:09
noonedeadpunk(in CI at least)15:09
noonedeadpunkAlso I spent time previous week for unplanned activity of fixing run of roles in check mode15:10
noonedeadpunkwas interesting how much doable is that...15:10
noonedeadpunkhosts and infra seems quite fine, openstack services are tricky15:11
noonedeadpunkI also thought it would be easy to have CI job but then realized that to run in check mode it has plenty of dependencies in terms of missing services and certificates. 15:11
noonedeadpunkStill doable but appeared a bit harder then expected15:12
noonedeadpunkWill likely continue that after more important things are done:)15:12
noonedeadpunkBTW we kind of need to make last release of V and move it to EM15:12
noonedeadpunkjrosser: do we need wip here? https://review.opendev.org/c/openstack/openstack-ansible-tests/+/83633515:14
jrosserno - i think it was just to try to get things merged in order down the stable branches15:15
jrosserthis has all been a quite big mess with not really any cherry-picks that work15:15
jrosserso i had to patch each branch pretty much individually15:15
jrosserand really the patch we need is https://review.opendev.org/c/openstack/openstack-ansible-tests/+/83736815:16
jrosserthere is a bunch of broken things due to that ^15:16
jrosserthe gluster patches are nearly ready for proper review15:18
jrosseri have to make a hack^W fix for rocky15:19
noonedeadpunkBut hopefully NeilHanlon would ping for some packaging changes one day?15:20
jrosserhopefully yes15:20
noonedeadpunkok great. And then I can pick CentOS 9 deployment on top of gluster :)15:21
NeilHanlonYes, i am working on fixing up those release packages for, e.g., gluster15:22
jrosseri just need to decide if we lineinfile/regex the repo file or vendor a copy of it and just copy: it into place15:22
noonedeadpunkgreat)15:22
jrosserdo you think we should be more specific about pinning the ceph repo on ceph hosts?15:24
jrosserif we pin the client to ceph.com but not the server, we still have potential conflicts between ubuntu / uca / ceph.com repos15:25
noonedeadpunkjrosser: Um, can you point to that Rocky hack as I can't find it quickly :)15:26
jrosserno i still try to write it :)15:26
noonedeadpunkah, ok then :)15:26
jrosserbut i believe we have to adjust mirrorlist=http://mirrorlist.centos.org/?release=$stream&arch=$basearch&repo=ResilientStorage&infra=$infra15:27
noonedeadpunkregarding ceph pinning - that kind of make sense... Despite that should be part of ceph-ansible to be fair...15:27
jrosserso that always release=8-stream15:27
damiandabrowski[m]i need to spend some time trying to figure out why i get this downgrade error, then I'll have a better view on how do we want to pin ceph repo. Planning to look into this during the evening15:29
noonedeadpunkceph-ansible already has requirements.yml in place, so technically we can push them a change to handle pinning and add another dependency on our roles :)15:29
jrosserNeilHanlon: do you have a bug link i can put in my patch for glusterfs?15:29
noonedeadpunkdon't think they're gonna like it though...15:31
noonedeadpunkI kind of more and more thinking that integration of ceph-ansible should be more for CI purposes only.15:32
noonedeadpunkwhile it's good to have playbooks for that and some sort of integration, I'm not really sure if anybody should do that in production :)15:33
noonedeadpunk(at least as is)15:33
damiandabrowski[m]I'm going to use it in production in a few weeks, i can share my thoughts then :D 15:34
noonedeadpunkthe questions come when you're trying to upgrade stuff :)15:34
noonedeadpunkas we pin specific ceph-ansible version that is tighten to ceph releases quite hardly15:35
SiavashSardarii have it in production. (he types with shaking hands :D )15:35
noonedeadpunkas then when you upgrade osa you should either override ceph-ansible or do upgrade ceph at same time kind of.15:36
mgariepyi had it with osa for exactly 1 deploy then i move the ceph-ansible on the side right after 15:36
SiavashSardariwe are thinking about replacing it with ceph-adm, not starting the procedure though 15:37
jrosserceph-ansible is fine if you use it on its own15:37
jrosseri wouldnt copy the approach from OSA CI into a production deployment15:37
noonedeadpunknah, we're not talking about ceph-adm, but more about at what level we should provide ceph deployment15:38
noonedeadpunkand if we explicit enough about how to use that properly15:38
SiavashSardaritbh I like the idea of having one repo for both but at the end of the day we are using them kinda separately and will override the role-requirement if needed15:41
damiandabrowski[m]i just think that ceph integration is a huge advantage of openstack-ansible project(especially for a newcomer who is is comparing different deployment tools and trying to choose one)15:41
damiandabrowski[m]so maybe we can list our main issues with it and think what we can do about them?15:42
jrosserceph-ansible sometimes has very specific requirements for the ansible version (though this is less important with modern ansible)15:42
jrosserand versions of ceph-ansible only support very specific releases of ceph15:42
mgariepyspecific major rel15:43
jrosserso you ended up in very difficult places at upgrade time when the openstack, ansible and ceph (and maybe also OS) versions all had to change simultaneously15:43
jrosserand thats a total nightmare for upgrades15:43
damiandabrowski[m]but You don't have to run `ceph-install.yml` in the middle of openstack upgrade, right? :D 15:44
noonedeadpunkno, if you're not running /run-upgrade.sh :p15:45
noonedeadpunk(which what newcommers will do)15:45
jrosserand we test exactly "zero" of that approach :)15:45
noonedeadpunkand you will get newer clients anyway unless explicitly set in user_variables15:45
noonedeadpunkI'm not talking about dropping support, but likely more being more explicit in documentation about possible caveats by doing that15:46
damiandabrowski[m]so maybe what we need to do is to show a warning message when running run-upgrade.sh with ceph enabled? 15:46
damiandabrowski[m]ah yeah, i was just going to say that15:46
damiandabrowski[m]that another option is put some information in docs15:47
jrosserwell - back to where we started there is really only benefit in pinning to ceph.com15:47
jrosserthe UCA packages didnt used to have debug symbols and that was a problem for us when things were broken15:47
damiandabrowski[m]saying that OSA upgrades may be tricky when it manages ceph as well15:47
noonedeadpunkI can recall some early on bionic distro provided ceph version was treated as newer one comparing to one that was coming from ceph.com15:48
jrosseryes, i remember having to fix that15:48
noonedeadpunkI guess that's when we added pinning :)15:49
jrosseri think it was the local patch that ubuntu applied made the version number sematically greater15:49
noonedeadpunkyup15:49
noonedeadpunkwhile in fact it wasn't and dependencies were broken15:49
noonedeadpunkanyway - let's see what the current issues are.15:50
jrosserdamiandabrowski[m]: you might want to look at this as it addresses similar errors to those you were seeing https://github.com/openstack/openstack-ansible-os_cinder/blob/9f2bf29db8ef921cfad7857dcb7652436d0d887b/tasks/main.yml#L183-L20915:51
noonedeadpunkofc we can add pinning in playbook alone15:51
damiandabrowski[m]thanks, I'll have a look15:51
jrosseri had to split the list of packages into those to install before, and after the ceph_client role is run15:52
damiandabrowski[m]btw. we already managed to merge few of my tempest patches, but many of them still remain unreviewed.15:53
damiandabrowski[m]it would be awesome to close this topic soon ;) https://review.opendev.org/q/topic:tempest-damian-2021-1215:53
mgariepyi'll take a look15:55
noonedeadpunkdamiandabrowski[m]: aren't most of them need rebase?15:56
noonedeadpunksee tons of `Indirect ancestor` in gerrit15:57
noonedeadpunklikely they won't merge cleanly but not sure15:57
damiandabrowski[m]that's right, but the plan was to get initial reviews ->  apply Your suggestions -> rebase -> merge15:58
damiandabrowski[m]by doing it like this, we'll avoid running too much unnecessary zuul jobs15:58
damiandabrowski[m]but I can rebase them now if You want :D 15:58
damiandabrowski[m]but main drawback of this plan is that You'll need to leave Your reviews twice on these patches(before and after rebase)16:00
noonedeadpunkwell, rebase in gerrit doesn't remove Code-Review label16:01
noonedeadpunkit would remove jsut Workflow16:01
noonedeadpunkbut ok, indeed:)16:01
damiandabrowski[m]ahhh okay16:01
noonedeadpunk#endmeeting16:01
opendevmeetMeeting ended Tue Apr 19 16:01:40 2022 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)16:01
opendevmeetMinutes:        https://meetings.opendev.org/meetings/openstack_ansible_meeting/2022/openstack_ansible_meeting.2022-04-19-15.00.html16:01
opendevmeetMinutes (text): https://meetings.opendev.org/meetings/openstack_ansible_meeting/2022/openstack_ansible_meeting.2022-04-19-15.00.txt16:01
opendevmeetLog:            https://meetings.opendev.org/meetings/openstack_ansible_meeting/2022/openstack_ansible_meeting.2022-04-19-15.00.log.html16:01
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-plugins master: Add role for creating simple glusterfs servers  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/83758216:04
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-plugins master: Add role for creating simple glusterfs servers  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/83758216:05
*** dviroel|lunch is now known as dviroel16:14
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-plugins master: Add role for creating simple glusterfs servers  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/83758216:56
jrosserhrrm the ceph job is breaking things in the gate17:15
mgariepyhmm let me have a look17:18
mgariepyi'll biuld an aio to check it out.17:18
jrosserits failing on msg": "[Errno 2] No such file or directory: b'ceph-volume'"17:20
mgariepyyep i saw that part but it will be easier to see what is going on in a vm17:20
jrosserif we were using the ubuntu package that should be in ceph-osd https://packages.ubuntu.com/focal/amd64/ceph-osd/filelist17:20
jrosserand in that CI jobs we get the package from ceph-osd/focal-updates,now 17.1.0-0ubuntu3~cloud0 amd64 [installed,automatic]17:22
jrosseralso i see upgrade jobs failing on git clone for ansible-sshd fatal: reference is not a tree: 59a3cb190f7913e32b044dc2733cfe7e46c4c99a17:25
damiandabrowski[m]FYI I have found out why I'm hitting librados2/librbd1 downgrade issue on my AIO17:33
damiandabrowski[m]lxc_hosts installs lxc-templates package, which has some dependency to librados2/librbd1, so these 2 packages are firstly installed from UCA repo17:34
jrosseroh wow17:34
damiandabrowski[m]and that's why they need to be downgraded afterwards17:34
damiandabrowski[m]now I'm trying to figure out the exact dependency tree and how to solve it properly17:34
jrosserthis definatly means we should pin to ceph.com17:34
jrosserreally early, like in openstack_hostts17:34
damiandabrowski[m]+117:35
jrosserwe already run the apt_package_pinning role there https://github.com/openstack/openstack-ansible-openstack_hosts/blob/master/meta/main.yml#L38-L4117:36
damiandabrowski[m]great, so it should be fairly easy to implement17:41
jrosseryeah, hopefully just needs a var adding in the right place into inventory/group_vars/....17:42
damiandabrowski[m]if I understand it correctly, the dependency chain looks like this: lxc-templates->cloud-image-utils->qemu-utils->qemu-block-extra->librbd1/librados217:43
damiandabrowski[m]I need to leave for ~3h, can continue later if needed17:43
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-plugins master: Add role for creating simple glusterfs servers  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/83758217:46
mgariepyceph-volume is splitted in uca.  https://paste.openstack.org/show/bLMvIdro1Wp0eMiWh0rd/ 19:38
mgariepylxc host role does pull the librados2 and librbd1 pkg which then breaks the install of ceph via ceph community :/20:30
mgariepynot sure what is the best way to fix this, appliing the pin before without the repo will have a net 0 effect.20:31
mgariepyhttps://zuul.opendev.org/t/openstack/build/7ee28ca3d39e472fab4c55798b4dec8b/log/logs/etc/host/apt/preferences.d/ceph_community_pin.pref.txt20:31
mgariepyso i guess we need the ceph repo before running lxc-container-host playbook.20:33
damiandabrowski[m]ouh, You're right20:37
mgariepywe migt need to run : https://github.com/openstack/openstack-ansible-ceph_client/blob/master/tasks/ceph_preinstall_apt.yml and https://github.com/openstack/openstack-ansible/blob/master/playbooks/common-tasks/ceph-server.yml20:37
mgariepyto apply the repo + the pin20:37
damiandabrowski[m]and I assume this repo is added by ceph-ansible(/etc/ansible/roles/ceph-ansible/roles/ceph-common/tasks/installs/install_debian_packages.yml)20:37
mgariepythe repo can be added by both roles20:38
mgariepyeither ceph_client or ceph-ansible 20:38
mgariepyall repos and pins should be applied to a server prior installing any software :D20:40
mgariepyhaha20:40
*** dviroel is now known as dviroel|out20:41
damiandabrowski[m]hmm, why do You think we also need to run https://github.com/openstack/openstack-ansible/blob/master/playbooks/common-tasks/ceph-server.yml early in setup-hosts?20:45
damiandabrowski[m]I may be missing something but apt pinning can be done directly via ceph_client role20:46
damiandabrowski[m]so I'm just looking for another reason to run ceph-server.yml20:46
damiandabrowski[m]i only wonder where exactly do we want to run ceph_client role(probably we need to implement 'preinstall' tag as well to avoid installing ceph packages and only add repo/pins?). 21:48
damiandabrowski[m]Only on lxc_hosts as 'lxc-templates' package triggers ceph-related packages(librbd1,librados2) to be installed which breaks things or maybe on all baremetal nodes?21:49
damiandabrowski[m]the other question is when do we want to run it? only when OSA manages ceph, so `groups['ceph_all'] | length > 0`?21:50
damiandabrowski[m]ehh, it's way more complicated than i thought :D 21:50
* damiandabrowski[m] heading to bed, good night21:50

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!