Wednesday, 2024-06-05

opendevreviewMerged openstack/openstack-ansible stable/2023.2: Enable RabbitMQ feature flags pre-upgrade  https://review.opendev.org/c/openstack/openstack-ansible/+/92099800:18
opendevreviewMerged openstack/openstack-ansible master: Bump SHAs for 2024.1 (Caracal) RC1  https://review.opendev.org/c/openstack/openstack-ansible/+/92127100:52
noonedeadpunkSome reviews for https://review.opendev.org/c/openstack/openstack-ansible/+/918979 would be appreciated07:34
noonedeadpunkuntil it passes CI and branches not dropped....07:34
hamidlotfi_Hi there,07:35
hamidlotfi_An instance whose volume is on CEPH, after abnormally restarting the Compute server where that instance is located, that instance does not come up and displays the following error:07:35
hamidlotfi_https://www.irccloud.com/pastebin/lQoVeUo0/07:35
hamidlotfi_The unsaid point is that I put the disk_cachemodes parameter in all three mentioned cases, but the problem persists.07:35
hamidlotfi_How can I solve this problem?07:35
noonedeadpunkhamidlotfi_: and you tried disk_cachemodes = network=directsync ?07:39
noonedeadpunk(or network=none)07:39
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_trove master: DNM  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/91505207:44
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_keystone master: Set X-Forwarded-Proto header in apache conditionally  https://review.opendev.org/c/openstack/openstack-ansible-os_keystone/+/92131207:46
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_keystone master: Set X-Forwarded-Proto header in apache conditionally  https://review.opendev.org/c/openstack/openstack-ansible-os_keystone/+/92131207:52
jrossergood morning07:54
noonedeadpunko/07:57
noonedeadpunkI've proposed branching of integrated repo: https://review.opendev.org/c/openstack/releases/+/91948407:58
noonedeadpunkand I think there was at least 1 backport to 2024.1 for cloudkitty07:59
hamidlotfi_noonedeadpunk: I try these  "file=directsync,block=none,network=writeback"08:00
hamidlotfi_but I didn't check the one you mentioned,08:00
hamidlotfi_let me check it.08:00
noonedeadpunkactually 2: https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/921284 and https://review.opendev.org/c/openstack/openstack-ansible-os_cloudkitty/+/92128508:00
noonedeadpunkhamidlotfi_: um, so writeback will crash FS :)08:00
jrosserandrew is away for today/tomorrow and all of next week so we are going to be struggling for reviews a bit08:00
noonedeadpunkand only `network` is valid for Ceph backend08:00
noonedeadpunkDamian is also having vacation :/08:00
noonedeadpunkwell, will summon folks then08:01
hamidlotfi_noonedeadpunk: Oh, this is an important point.08:01
noonedeadpunkhamidlotfi_: there's quite good explanation what each mode means: https://docs.openstack.org/nova/latest/configuration/config.html#libvirt.disk_cachemodes08:01
noonedeadpunkand what you defined is just default value08:02
noonedeadpunkso yeah, basically you did not change any behaviour :)08:02
hamidlotfi_noonedeadpunk:  Yes ;)08:06
noonedeadpunkhamidlotfi_: but frankly, we're pretty much fine with writethrough for Ceph08:32
noonedeadpunkwhich kinda balances risk and performance08:33
jrosserit might be better trying out the ceph caching stuff too08:33
noonedeadpunkyeah, true08:33
jrosseras you get to optimise if you care about caching the underlying image, or writes08:33
noonedeadpunkthough you'd need extra drives on computes afaik08:34
jrossertrue08:34
jrosserand also tbh ceph was never about perforamance to a single vm08:34
opendevreviewOpenStack Release Bot proposed openstack/openstack-ansible stable/2024.1: Update .gitreview for stable/2024.1  https://review.opendev.org/c/openstack/openstack-ansible/+/92135708:46
opendevreviewOpenStack Release Bot proposed openstack/openstack-ansible stable/2024.1: Update TOX_CONSTRAINTS_FILE for stable/2024.1  https://review.opendev.org/c/openstack/openstack-ansible/+/92135808:46
opendevreviewOpenStack Release Bot proposed openstack/openstack-ansible master: Update master for stable/2024.1  https://review.opendev.org/c/openstack/openstack-ansible/+/92135908:46
noonedeadpunkalso, this would be nice fix kinda: https://review.opendev.org/c/openstack/openstack-ansible-os_horizon/+/92131408:55
noonedeadpunkmgariepy: NeilHanlon if/when around can you check on it? ^08:55
noonedeadpunkand also these backports to 2024.1: https://review.opendev.org/c/openstack/openstack-ansible-os_cloudkitty/+/921285 and https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/92128409:03
noonedeadpunkand also, as I've scrwwed deploy guide - we need to land and backport https://review.opendev.org/c/openstack/openstack-ansible/+/92123709:05
noonedeadpunkand https://review.opendev.org/c/openstack/openstack-ansible-os_horizon/+/921314?tab=change-view-tab-header-zuul-results-summary fixes distro installation of horizon for EL as it;'s broken now09:09
jrosseri was looking at that and its a bit 8-009:10
noonedeadpunkyeah....09:15
noonedeadpunkit kinda is.09:15
noonedeadpunkas rdo started (at some point?) to place static files and openstack_dashboard not inside site-packages (as that python packages after all), but in /usr/share...09:16
noonedeadpunkthe only question I had if we wanna have condition or always add libdir to manage.py script09:25
noonedeadpunkthis kinda doesn't hurt overall09:26
noonedeadpunkah, and when horizon_lib_dir is not defined - it will get detected on it's own by python path for lib files09:29
opendevreviewMerged openstack/openstack-ansible stable/2024.1: Update .gitreview for stable/2024.1  https://review.opendev.org/c/openstack/openstack-ansible/+/92135709:40
opendevreviewMerged openstack/openstack-ansible stable/2024.1: Update TOX_CONSTRAINTS_FILE for stable/2024.1  https://review.opendev.org/c/openstack/openstack-ansible/+/92135809:42
opendevreviewMerged openstack/openstack-ansible master: Update master for stable/2024.1  https://review.opendev.org/c/openstack/openstack-ansible/+/92135909:54
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Start 2024.2 (Dalmatian) development  https://review.opendev.org/c/openstack/openstack-ansible/+/92136709:57
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Start 2024.2 (Dalmatian) development  https://review.opendev.org/c/openstack/openstack-ansible/+/92136710:04
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Switch services SHAs to track master  https://review.opendev.org/c/openstack/openstack-ansible/+/92136810:06
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Drop support for Debian 11 (bullseye)  https://review.opendev.org/c/openstack/openstack-ansible/+/92137010:30
jrosserdo we need to go back through the stable/unmaintained branches and strip out centos 8-stream?10:38
opendevreviewMerged openstack/openstack-ansible-os_cloudkitty stable/2023.2: Define region and interface for collector/fetcher  https://review.opendev.org/c/openstack/openstack-ansible-os_cloudkitty/+/92105411:00
noonedeadpunkwe do...11:03
opendevreviewMerged openstack/openstack-ansible-openstack_hosts stable/2023.2: Do not mark all directories as safe for git clone by default  https://review.opendev.org/c/openstack/openstack-ansible-openstack_hosts/+/92132311:14
opendevreviewMerged openstack/openstack-ansible-openstack_hosts stable/2023.1: Do not mark all directories as safe for git clone by default  https://review.opendev.org/c/openstack/openstack-ansible-openstack_hosts/+/92132411:24
hamidlotfi_noonedeadpunk: I checked all of those values for the disk_cachemodes parameter,11:31
hamidlotfi_network=directsync 11:31
hamidlotfi_network=none11:31
hamidlotfi_network=writethrough11:31
hamidlotfi_but in all of the values, the result is the same and I had that error.11:31
fricklerhamidlotfi_: none of these can repair your broken fs, those option only help to repeat the issue from happening again on the next host crash11:34
noonedeadpunkyes, absolutely ^11:37
hamidlotfi_Yes, I know, I have nothing to do with the previous file system in that instance, deleted it, the test I am doing is on a newly created instance and I expect this will not happen to the new instance, but this will happen again. I cold boot my compute host to simulate this event.11:40
hamidlotfi_After changing the value of this parameter every time, I create an instance and cold reboot the compute server.11:42
hamidlotfi_I hope I have explained well.11:42
noonedeadpunkaha, ok, then it's quite surprising frankly speaking...11:45
hamidlotfi_unfortunately11:46
hamidlotfi_Do you have any other recommendations? Maybe another parameter?11:58
noonedeadpunkfrankly - not really12:13
hamidlotfi_Ok, Thanks12:13
mgariepygood momrning ! went through the reviews you posted 3 hours ago noonedeadpunk. 12:37
noonedeadpunkawesome!12:37
noonedeadpunkthanks12:37
noonedeadpunkso, upgrade jobs looks quite borked for 2024.112:52
noonedeadpunkie https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/92128412:52
noonedeadpunkfails on missing tasks/rabbitmq_feature_flags.yml12:53
noonedeadpunkwonder if we need to have sha bump on stable branches for that...12:54
noonedeadpunkas highly likely we do12:54
mgariepy2023.1 works but not 2023.2 ?13:00
mgariepyupgrade from ^^13:01
mgariepyha. the patch for 2023.1 is not merged.13:04
noonedeadpunkyeah, I guess we need to bump shas to get the rabbitmq role with the task13:07
jrosseri guess we should be careful about merging feature flag things further back13:13
mgariepythe bump should have been added to :https://review.opendev.org/c/openstack/openstack-ansible/+/92099813:13
jrosserelse the upgrade jobs will also be broken further back and prevent merging a sha bump13:14
jrosseryeah13:14
mgariepysince the tests are not quite complete.. we missed it.13:15
mgariepyupgrade job are always tested on tip of the previous release i guess ?13:16
jrosserwell i guess that is chicken/egg because you cant test the bumped sha until you merge the bump?13:17
mgariepyindeed13:18
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible stable/2023.2: Update RabbitMQ SHA  https://review.opendev.org/c/openstack/openstack-ansible/+/92137913:26
noonedeadpunkhopefully this will fix it ^13:26
mgariepywas the depends-on root cause of the upgrade test pass on merge? in the 2023.2 branch ?13:47
noonedeadpunkwell, CI always kinda gets zuul-prepared repos for current branch13:49
noonedeadpunkbut for N-1 it does bootstrap process as user13:49
noonedeadpunkand then on N uses what zuul has13:49
noonedeadpunkand zuul has latest for current branch13:49
mgariepyokie13:50
opendevreviewMerged openstack/openstack-ansible-os_horizon master: Fix horizon distro installation for EL  https://review.opendev.org/c/openstack/openstack-ansible-os_horizon/+/92131414:58
opendevreviewMerged openstack/openstack-ansible-rabbitmq_server stable/2023.1: Enable feature flags post-upgrade  https://review.opendev.org/c/openstack/openstack-ansible-rabbitmq_server/+/92099915:28
deflatedHi all, trying to deploy 29.0.0rc1 on debian bookworm but getting an apt key depreciated errorr, found the bug report but its still wip, is there a way round this or is it not possible to deploy 29 on bookworm? TiA15:49
opendevreviewMarc Gariépy proposed openstack/openstack-ansible stable/2023.1: Update RabbitMQ SHA  https://review.opendev.org/c/openstack/openstack-ansible/+/92139615:51
mgariepynoonedeadpunk, copied your commit message :) hehe15:51
noonedeadpunkdeflated: hey, it's weird you;re getting error. can you please paste it?15:53
noonedeadpunkmgariepy: I think it's better to update https://review.opendev.org/c/openstack/openstack-ansible/+/921000 ?15:54
deflatedsure, where should i paste it?15:56
deflatedNevermind, got it15:56
noonedeadpunkie https://paste.openstack.org/15:56
mgariepyhmm sure.15:57
mgariepysorry for that.15:57
deflatedthe catchpa seems broken on there15:58
deflatedso its here instead: https://pastebin.com/6kmZjhN515:58
deflatedthat goes on for miles and happens on every container it seems15:59
noonedeadpunkdeflated: so, apt key depreciated is warning from what I see15:59
noonedeadpunkbut it's followed by some different error I've never seen before15:59
noonedeadpunk"E: Repository 'http://osbpo.debian.net/osbpo bullseye-zed-backports-nochange InRelease' changed its 'Origin' value from 'Infomaniak' to 'osbpo'"16:00
opendevreviewMarc Gariépy proposed openstack/openstack-ansible stable/2023.1: Enable RabbitMQ feature flags pre-upgrade  https://review.opendev.org/c/openstack/openstack-ansible/+/92100016:00
noonedeadpunkdeflated: also I kinda wonder how you ended up with both bullseye and bookworm repos16:01
deflatedit was an upgrade from it but i have checked all repos and removed everything related to it so i'll quickly chgeck my configs16:02
noonedeadpunkum, maybe you left ansible facts?16:03
deflatedin /etc/opensdtack_deploy?16:03
noonedeadpunkyeah16:04
jrosser(before you delete them just take a look at one of the files to properly understand this)16:04
noonedeadpunkso, we do test bookworm in CI, and that's recent results for LXC test: https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_98b/921367/2/check/openstack-ansible-deploy-aio_lxc-debian-bookworm/98bfb20/logs/etc/16:04
noonedeadpunkyou can see configured repos there and compare with what you have16:04
noonedeadpunkbut the one that's role not happy with is https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_98b/921367/2/check/openstack-ansible-deploy-aio_lxc-debian-bookworm/98bfb20/logs/etc/openstack/aio1-horizon-container-3e2f35c5/apt/sources.list.d/osbpo.list.txt16:05
noonedeadpunkalso, keep in mind that some "base" repos are copied to containers from host IIRC16:05
jrosserwe basically don't support having containers that are not the same as the host OS16:06
noonedeadpunkwell... I guess, that facts are container specific?16:09
noonedeadpunkso if you've spawned containers on bullseye, then upgraded underlying OS, ansible will correctly manage existing bullseye on top of bookworm16:10
noonedeadpunkbut yes, in case you re-create containers, there's no OS choice - it will be same as host OS16:10
noonedeadpunksmth I wanted to work to improve at some point...16:10
noonedeadpunkanother vote on this one would be nice: https://review.opendev.org/c/openstack/openstack-ansible/+/92137916:12
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-ops master: Update ansible-collection-kubernetes version pin  https://review.opendev.org/c/openstack/openstack-ansible-ops/+/92139916:45
opendevreviewJonathan Rosser proposed openstack/openstack-ansible-ops master: Revert "Ensure that python3-cryptography is present in k8s control plane hosts"  https://review.opendev.org/c/openstack/openstack-ansible-ops/+/92140016:50
deflatederok so i've removed all containers as well as files from /etc/openstack_deploy (apart from what is needed anyway), checked apt/sources for anything relating to bullseye (there's nothing), deleted unused apt keys, ran cleanup_venvs from ansible-ops just in case its held there and reran, same issue, did the same again, rebooted, updated, same issue.18:11
deflatederJust to be clear i deleted /openstack /opt/*ansible* and a few others, not sure why at this point, nothing on my system points towards bullseye18:14
mgariepydeflateder, anything in /etc/openstack_deploy/ansible_facts ?18:20
deflatedernot the second time, i deleted the folder and everything i could find relating to openstack and ansible18:20
mgariepyafter the run what are the fact for nova container?18:21
mgariepylike: ansible_distribution_release18:22
deflatedergive me 5 and i'll let you know18:22
deflatederwell. 15 mor elike18:23
mgariepyno worries18:24
deflateder"ansible_distribution_release": "bookworm" _version "12.4"18:44
mgariepystill have the zed repo in the container?18:45
deflatederyes18:46
deflatederwhich is just weird at this point18:47
mgariepywhen was the repo added ?18:47
mgariepymight come from the lxc cached imagE?18:47
deflatederwhenever i tried zed so howver many months/years ago that was, i have successfully ran 2023 at one point tho, the cached image you mention is not something i would of removed, where wo7uld i find it?18:48
mgariepyif you run stat in the apt directory when does the files were last modified?18:51
deflatederToday, osbo.list points to bookworm-caracal-backports main18:54
deflatederonly other list i have is ceph which points to reef18:54
mgariepyhuh. wasn't the issue some weird repos in your containers ?19:16
deflatederyes, but you didnt specify if the check you asked me to do was in the container or on the host, i checked the host19:17
deflatedercan check the container if thats what you meant19:17
mgariepyyes please.19:19
deflatederok, i'll recreate them, just wiped everything openstack/ansible related that i could find19:20
jrosserwierd ara behaviour in my failing capi jobs19:22
deflatederused nova container, dates are all today apart from modify which states 2023/12/1319:38
mgariepythe novacontainer is bookworm as well ?19:52
mgariepywhat playbook have you ran ?19:53
deflatedergit clone, then bootstrap, then setup_hosts.yml19:56
mgariepyflush the container and rerun with `-e lxc_image_cache_refresh=true `20:01
deflatedernot sure what you mean by flush, i'm assuming purge/delete?20:02
mgariepyor simply destroy the containers.20:03
mgariepythere is a playbook for this.20:03
deflatederyeah i am running containers_destroy20:04
jrosserdeflateder: can you restate what is happening - i cant quite follow what actually is wrong20:06
deflatedertrying to deploy 29.0.0rc1 on debian bookworm but getting an apt key depreciated error, paste is here: https://pastebin.com/6kmZjhN5 Seems a bullseye repo is getting added toi my containters during setup_hosts20:10
deflatederRunning setup_hosts with -e lxc_image_cache_refresh=true now20:11
deflatederquick question while i'm waiting for ity to complete/fail, does ceph-rgw-install.yml setup the nedpoint for a pre existing ceph cluster? Just checking before i contemplate running it to connect an existing ceph installation20:33
deflatederyes! it just finished, thanks for the solution! Adding that to my notes so i know for next time20:34
jrosserdeflateder: well it means we have a bug somewhere20:34
jrosserthe container image should be automatically rebuilt if the OS is upgraded20:35
jrosserI exp20:35
jrosserI expect  most people redeploy/PXEboot rather than upgrade in place20:36
deflatederpossibly, i've never had the need to do so with debian, its always worked flawlessly even upgrading, i had the same problem on 2023.2 as well, dont recall it occuring in 2023.120:37
jrosserok no worries - we have no test coverage for this so the most ideal case would be submitting a patch to fix it20:39
deflatederi'll see about how i do that shortly20:45
jrosserthere will be some condition in the lxc_hosts role that does not correctly see that the base image in /var/lib/machines is out of date20:49
mgariepyweird indeed.21:36
mgariepyis ntp working correcly on your server ?21:36
jrosserthe image in /var/lib/machines should have the OS release name in it21:50
jrosserand that getting updated should trigger the rootfs to be compressed and pushed into lxc21:51
deflatederit is, after that command setup_hosts ran ok, now running setup_openstack21:51
deflatederNew error, An exception occurred during task execution. To see the full traceback, use -vvv. The error was: AttributeError: 'int' object has no attribute 'items'22:11
deflatederfatal: [ost01-cinder-volumes-container-25ad06a8]: FAILED! => {"msg": "Unexpected failure during module execution: 'int' object has no attribute 'items'", "stdout": ""}22:11
deflatederFull trace here: https://pastebin.com/nkQRnDAy22:11
opendevreviewMerged openstack/openstack-ansible stable/2023.2: Update RabbitMQ SHA  https://review.opendev.org/c/openstack/openstack-ansible/+/92137922:52

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!