Wednesday, 2023-07-26

mgariepywait.. what ? https://access.redhat.com/security/cve/cve-2023-2059300:07
noonedeadpunkwonder why it is "moderate" if you can "access sensitive information"06:14
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Add openstack_resources role skeleton  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/87879406:29
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Add openstack_resources role skeleton  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/87879407:17
anskiynoonedeadpunk: nah, it's fine. I wonder if that role could be either unit-tested or pyargv could be moved to uwsgi ini-file.07:56
noonedeadpunkyup, that's fair, we're talking about adding molecule testing for quite a while now, but quite little time to really work on that07:57
jrosseranskiy: can pyargv go in the ini file?07:58
noonedeadpunkThe main "blocker" or better say smth we need to figure out, is how to "centrally" define molecule environemnt07:58
noonedeadpunklike ansible versions, required collections, etc07:59
noonedeadpunkmolecule version itself07:59
anskiyjrosser: I'm not sure, this change would be better made separate, I guess, so I haven't really looked into that07:59
jrosseriirc it was pretty tricky to find a way to do what pyargv was needed for08:00
jrosseraaah that line in the uwsgi role was origianlly a var in quotes itself "{{ }}" and needed to include quotes in the rendered output - i see whats happened there now08:04
noonedeadpunkyeah, as I said - my bad :(08:13
noonedeadpunkAnd really wonder how much such bugs I've introduced....08:14
noonedeadpunkI tried to be careful though...08:14
noonedeadpunkIt's good we're not _that_ close to the release08:14
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Add openstack_resources role skeleton  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/87879408:58
damiandabrowskihi folks, so regarding upgrade job failures on stable/2023.1:09:05
damiandabrowskitonight i successfully executed upgrade jobs on my two AIO VMs(scenario=aio_metal, branch=stable/2023.1, os=ubuntu_jammy).09:05
damiandabrowskiI have no clue why these jobs are failing in CI. These problems seem to be intermittent(but with high probability to occur).09:05
damiandabrowskiProblems started around 13.07.2023.09:05
damiandabrowskiopenstack-ansible-upgrade-aio_metal-ubuntu-jammy: https://zuul.opendev.org/t/openstack/builds?job_name=openstack-ansible-upgrade-aio_metal-ubuntu-jammy&branch=stable%2F2023.1&skip=009:05
damiandabrowskiopenstack-ansible-upgrade_yoga-aio_metal-ubuntu-focal: https://zuul.opendev.org/t/openstack/builds?job_name=openstack-ansible-upgrade_yoga-aio_metal-ubuntu-focal&branch=stable%2F2023.1&skip=009:05
damiandabrowskiopenstack-ansible-upgrade-aio_metal-rockylinux-9: https://zuul.opendev.org/t/openstack/builds?job_name=openstack-ansible-upgrade-aio_metal-rockylinux-9&branch=stable%2F2023.1&skip=009:06
noonedeadpunkdamiandabrowski: can you try setting keystone passwords to 64 symbols for key services during initial (N-1) deployment?09:10
noonedeadpunkand then upgrade?09:10
noonedeadpunkAh! And ensure that your ansible-collections-requirements track top of the branch for openstack-ansible-plugins09:10
noonedeadpunkspecifically this patch https://opendev.org/openstack/openstack-ansible-plugins/commit/88a8bfcd62fe7bb027ca7a8636fbe943bfda88c109:11
noonedeadpunkon N-1 and N09:11
damiandabrowski1. keystone passwords - but currently they have 16-64 symbols. So they need to have exactly 64 chars? Why does it change anything?09:15
damiandabrowski2. um, how to do that in CI for N-1?09:15
jrosserdamiandabrowski: https://opendev.org/openstack/keystone/src/branch/master/keystone/common/password_hashing.py#L7109:28
damiandabrowskiah ok i get i know, so on my local AIO i will set password length >54 chars for all services09:34
jrosserdoesnt it need to be less than / equal to 54?09:38
jrosseri beleive that this is something to do with greater than 54 char length passwords being truncated and this then breaking idempotency09:38
jrossernoonedeadpunk: ^ is this correct understanding?09:39
jrosserdamiandabrowski: ah you mean to reproduce the CI failure you will try passwords > 54 chars?09:40
noonedeadpunkBut to catch error you need it to be >5409:40
noonedeadpunkso understanding is correct, just depends on what you wanna do :)09:40
jrosser:)09:40
noonedeadpunkdamiandabrowski: it would be also great to understand on how to migratre from bcache 09:41
damiandabrowskijrosser: yeah, i want to reproduce this issue on my AIO :D 09:41
damiandabrowskinoonedeadpunk: bcache or bcrypt?09:42
noonedeadpunkah, yes, bcrypt09:42
noonedeadpunkto scrypt09:42
noonedeadpunkAs this is smth we might wanna do internally09:43
noonedeadpunk(and might wanna change default upstream)09:43
noonedeadpunk(or at least in osa)09:43
jrosserwhat would happen to existing bcrypt passwords?09:44
noonedeadpunkI can recall some option to re-hash09:45
* jrosser surprised09:46
noonedeadpunkat least that was done once when keystone has changed default algo back in.... Queens?09:46
noonedeadpunkhttps://opendev.org/openstack/keystone/src/commit/8ad765e0230ceeb5ca7c36ec3ed6d25c57b22c9d/releasenotes/notes/bug_1543048_and_1668503-7ead4e15faaab778.yaml09:47
damiandabrowskialternatively, we can leave them as is and just encrypt new passwords with scrypt09:48
noonedeadpunkyes, or that. but this all should be tested/verified first09:48
jrosseryeah, just surprises me that without the original password plaintext you can rehash it09:48
jrosserobviously do-able for service users managed by OSA09:48
jrosserbut anything outside of that feels tricky09:49
noonedeadpunkmaybe you can't and I got previous change wrong. I guess damiandabrowski is right, and old will continue working "as is"09:49
noonedeadpunkbut I was not digging there deep enough09:49
amaraoHello. I found that if external_lb_vip_address points to domain with A and AAAA record, haproxy is listening only on ipv6 address (effectively ignoring A record). Is this a known bug or should I report it? (26.0.0).10:07
noonedeadpunkamarao: well, we have haproxy_bind_external_lb_vip_address and haproxy_bind_internal_lb_vip_address for such cases10:21
amaraoOh, so it's a feature, not a bug. Thanks.10:22
noonedeadpunkWell. That's kinda space for improvement ofc.10:23
noonedeadpunkBut it's somehow known10:23
noonedeadpunkAlso we have https://docs.openstack.org/openstack-ansible-haproxy_server/latest/configure-haproxy.html#adding-additional-global-vip-addresses that we usually use for IPv610:24
noonedeadpunkamarao: OR, you can bind to the interface and don't care about IPs at all. 10:25
noonedeadpunkfor example - haproxy_bind_external_lb_vip_address: * haproxy_bind_external_lb_vip_interface: bond010:26
noonedeadpunkI probably should patch documentation to reflect that..10:26
opendevreviewMerged openstack/ansible-role-systemd_mount stable/zed: Installing systemd-udev with NVR  https://review.opendev.org/c/openstack/ansible-role-systemd_mount/+/88935210:28
opendevreviewMerged openstack/ansible-role-uwsgi master: Fix pyargv value rendering  https://review.opendev.org/c/openstack/ansible-role-uwsgi/+/88964310:35
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-haproxy_server master: [doc] Document usage of binding to interface  https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/88969110:36
opendevreviewMerged openstack/openstack-ansible-plugins stable/zed: Installing systemd-udev with NVR  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/88934910:41
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible stable/yoga: Pin version of setuptools  https://review.opendev.org/c/openstack/openstack-ansible/+/88902210:47
hamidlotfi_Hi there,11:09
hamidlotfi_I have a new group with the name `all_lxc_containers` but the before of time it was `all_containers`, what's the difference between these?11:09
hamidlotfi_And now when running the `setup-hosts.yml --syntax-check` show me this warning: `[WARNING]: Could not match supplied host pattern, ignoring: all_lxc_containers`11:09
hamidlotfi_In the `Stable/Zed`11:10
opendevreviewMerged openstack/openstack-ansible-plugins stable/2023.1: Installing systemd-udev with NVR  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/88934811:14
hamidlotfi_@noonedeadpunk 11:22
noonedeadpunkhamidlotfi_: I don't think this is smth we have by default but rather smth that was created according to configuration11:28
noonedeadpunkSo I have no idea what is that, as it's not present neither in our production envs nor in sandbox11:28
noonedeadpunkdamiandabrowski: I have a question to you regarding tempest, as you was the last touching resource creation there:)11:30
noonedeadpunkHow does it work at all in case tempest_projects_create: False? As if it's not executed - it will fail all futher tasks? https://opendev.org/openstack/openstack-ansible-os_tempest/src/branch/master/tasks/tempest_resources.yml#L175-L17911:31
hamidlotfi_https://www.irccloud.com/pastebin/jGnLAR3P/11:32
noonedeadpunkah!11:32
hamidlotfi_@noonedeadpunk 11:32
noonedeadpunkok, that is dynamic group :)11:32
noonedeadpunkI think it was kinda introduced back in times when we had nspawn containers11:34
noonedeadpunkto produce empty group when there was no lxc_hosts/lxc_containers in inventory11:35
noonedeadpunkLikely, we can drop that now11:36
noonedeadpunkAs I don't really see how it could be useful11:36
hamidlotfi_doesn't pose a problem for deployment?11:37
noonedeadpunkNo, it should not?11:39
hamidlotfi_OK, Thanx11:40
noonedeadpunkthat is like that for last 6 years and never was problematic from what I know11:40
hamidlotfi_ok11:41
opendevreviewMerged openstack/openstack-ansible master: nova/haproxy: fix typo in detection of 'serialconsole'  https://review.opendev.org/c/openstack/openstack-ansible/+/88941711:44
opendevreviewMerged openstack/openstack-ansible stable/yoga: Include proper vars_file for rally  https://review.opendev.org/c/openstack/openstack-ansible/+/88865611:44
opendevreviewMerged openstack/openstack-ansible stable/yoga: Disable upgrade jobs from Xena  https://review.opendev.org/c/openstack/openstack-ansible/+/88931011:44
anskiyjrosser: well, adding `pyargv = --debug` to `/etc/uwsgi/neutron-server.ini` does work.11:57
opendevreviewAndrew Bonney proposed openstack/openstack-ansible stable/2023.1: nova/haproxy: fix typo in detection of 'serialconsole'  https://review.opendev.org/c/openstack/openstack-ansible/+/88935812:01
lsudreHi, I'm trying to run setup-opentack.yml and I have a task failed https://paste.openstack.org/show/bPsADCP6nymVCm6bwTkP/ this task try to cp /etc/nova/api-paste.ini but my directory has no api-paste.ini file. Have you any idea why this file is missing in my setup?12:23
noonedeadpunklsudre: I'd say that your venv installation might fail during previous run12:27
lsudrenoonedeadpunk: should I rerun setup-infrastructure.yml?12:28
noonedeadpunkI'd suggest to re-run os-nova-install.yml -e venv_rebuild=true12:29
lsudreOk I'm re-running this file12:34
anskiynoonedeadpunk: ceph-ansible is unusable in Zed, as it requires Ansible 2.12. But next, they're gonna require 2.14 in this PR: https://github.com/ceph/ceph-ansible/pull/7432.12:40
anskiyand with this PR, they're removing group name variables, for example: https://github.com/ceph/ceph-ansible/pull/7432/files#diff-3bdc3933d528a6a3547962f750000ce4ef0980aeebe19c7caeff9820069096e9L721-L72612:42
jrosseranskiy: i think you have to use very specific versions of ceph-ansible with each OSA release if you want them to use the same ansible12:42
noonedeadpunkanskiy: I'd say it depends on what you mean under "unusable". OSA playbooks does not include their verification of ansible version, and 2.13 is quite compatible with 2.12, so things "just work" if not their verification12:42
anskiyfor Zed it's stable-7.0 now, prior to that it was stable-6.0: both of them require the same Ansible verion -- 2.1212:42
noonedeadpunkanskiy: and then for 2023.2 we're working on merging 2.1512:43
noonedeadpunkhttps://review.opendev.org/q/topic:osa%252Fcore-2.1512:43
anskiynoonedeadpunk: by "unusable" I mean out of the box -- you have to patch ceph-ansible 12:43
jrosseranskiy: but we run CI jobs for ceph - those would be broken surely?12:44
noonedeadpunkSo, if you're using ceph-ansible playbooks, likely you need to have an independent venv to run it12:44
noonedeadpunkor use it completely independently12:44
jrosser^ yes thats what we do12:44
noonedeadpunkif you're using OSA playbooks - they work12:44
jrossertotally separate ansible venv for ceph12:44
noonedeadpunkand it's passing CI even12:44
jrosseri think it's good to remember that the OSA<>ceph-ansible integration is primarily for CI / illustration purposes12:45
noonedeadpunkI kinda tried to promote separation of ceph-ansible playbooks to avoid this confusion and level of support/compatibiltiy12:45
noonedeadpunkbut quite some folks were against that12:45
jrosserand personally i would totally not use it like that in a prod environment12:45
anskiywell, it was me and damiandabrowski, actually :)12:45
noonedeadpunkand that specific reason was exactly one of several why I wanted to do that12:46
noonedeadpunkAs we can't really expect to sync to such narrow requirements of ansible12:46
jrosserthe coupling between ansible <> ceph <> openstack versions is far far too tight and you soon find impossible situations at upgrade12:46
noonedeadpunkIn PR they release requirement to be >= 2.14 which is way more nicer then current one12:47
anskiyyeah, but at the same time, they have this thing: https://github.com/ceph/ceph-ansible/pull/7432/files#diff-b8a67f35d97d561cb8bdb2bcdfa147c3c4d6fcf91e6d5d589deffb4cf3abe7b3R912:48
lsudreNow I found this error when I run os-nova-install.yml -e venv_rebuild=true https://paste.openstack.org/show/bmpigL8X4j9rPC3YrFlo/12:48
anskiyI was probably wrong, and it's broken only on upgrade, which includes that check, I guess12:48
jrosserright - but in OSA we just use a subset of the ceph-ansible roles12:48
anskiyyeah-yeah, my bad, sorry for that12:49
noonedeadpunklsudre: os-nova-install.yml -e venv_rebuild=true -e venv_wheels_rebuild=true12:50
noonedeadpunkanskiy: to be frank, I would propose a patch to ceph ansible to introduce a variable or a tag that would allow to avoid this specific version check, saying that it's done at your own risk12:51
noonedeadpunkbut dunno...12:51
noonedeadpunkmaybe you already can do that12:52
anskiywe have a couple of patches for ceph-ansible already... like the one that prevents ALL OSDs from restart when new one is added12:52
anskiynah, you can't as of now: https://github.com/ceph/ceph-ansible/blob/stable-7.0/infrastructure-playbooks/rolling_update.yml#L105-L10612:55
noonedeadpunkyeah, I already checked and they have no tags or anything that would allow to do that12:57
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Define default value for _service_adminuri_insecure  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/88970713:06
lsudrenoonedeadpunk: I got this error: https://paste.openstack.org/show/bXrPRhK2BIuVELopIWqX/13:09
noonedeadpunklsudre: that is actually weird. Can you kindly show more output?13:09
noonedeadpunklsudre: also was you using limits when running the playbook?13:10
noonedeadpunk(but I guess it should work nicely with limits as well)13:10
noonedeadpunklsudre: also can you provide output of `curl http://172.16.12.69:8181/os-releases/27.0.0/ubuntu-22.04-x86_64/wheels`?13:12
lsudrenoonedeadpunk: for the curl command the return is https://paste.openstack.org/show/bS15eRK4KdNs6YeYYs4V/13:14
noonedeadpunkwell, then you should look inside your repo-containers and if nginx is running there13:15
lsudrenoonedeadpunk: i dont use containers13:16
noonedeadpunkthen the host where repo server should run13:16
lsudrewhere I run the playbooks?13:17
lsudreyou can find more logs: https://paste.openstack.org/show/bbFIlJAWmjx9tS2zEP6Q/13:18
noonedeadpunklsudre: um, no, your infra nodes likely. 13:21
noonedeadpunkYou can check that by IP that's configured in haproxy backend13:22
noonedeadpunkor smth like `echo "show stat" | nc -U /run/haproxy.stat | grep repo_all-back` might point you to the host as well13:23
noonedeadpunkor even `echo "show stat" | nc -U /run/haproxy.stat | grep repo_all-back | cut -d ',' -f 2`13:23
lsudrenoonedeadpunk: in my infra1 i have nginx and I run th curl command and got a 301 Moved permanently13:29
noonedeadpunkah, add backslash at the end13:30
lsudrei got a prompt13:30
noonedeadpunkok, but do you still get 503 when asking through haproxy?13:30
noonedeadpunkor better say - thorugh internal VIP13:31
lsudreI got this html : https://paste.openstack.org/show/bcaBM2m6yhXR1dEIZke7/13:32
noonedeadpunklsudre: have you done smth to nginx? As 503 means that no backend is available basically13:33
lsudresystemctl status nginx return me an active and running process13:34
noonedeadpunkas nova version in nginx is same as being asked in your output here https://paste.openstack.org/show/bXrPRhK2BIuVELopIWqX/13:34
noonedeadpunkBut do you still get this https://paste.openstack.org/show/bS15eRK4KdNs6YeYYs4V/ when curling http://172.16.12.69:8181/os-releases/27.0.0/ubuntu-22.04-x86_64/wheels/ ?13:35
lsudrefrom where?13:35
noonedeadpunkum, let's say from compute01?13:35
noonedeadpunkfrom where you got it at the first place?13:36
jrosserwhere is the 503? /me confused13:36
noonedeadpunkyeah. me too13:36
lsudreok so from my compute I got a : No server is available13:36
noonedeadpunkok, so that is the issue basically13:37
noonedeadpunkyou should be able to access nginx from your computes13:37
jrosserlsudre: it's pretty helpful if you also paste the command you tried and the output too13:37
lsudreI got a 503 when I curl from my deploy host (where I run ansible playbooks)13:37
noonedeadpunkthrough private VIP13:37
jrosserlsudre: the thing is "I got a 503 when I curl from my deploy host" <- curl what13:38
lsudre172.16.12.69 is the VIP 13:38
jrossereither the VIP or the repo server IP13:38
jrosserit's imporant to be really clear and specific here then we can help the best13:38
lsudrecurl this http://172.16.12.69:8181/os-releases/27.0.0/ubuntu-22.04-x86_64/wheels/13:38
jrosserso where did the HTML output come from?13:38
lsudrejrosser: noonedeadpunk ask me to curl this url13:38
lsudrecontroller13:38
jrosserand with the same command?13:39
noonedeadpunksounds like... keepalived split brain?13:40
jrosserif this is a metal deploy then the services need to be bound to !VIP as well13:41
lsudreTo avoid some missunderstood: https://pasteboard.co/DppQyfpsnM3F.png13:41
jrosserlsudre: you can only run the haproxy stat command on the host that is running haproxy13:42
jrosserit connects to a local unix socket13:42
noonedeadpunkand please provide output of that :) preferably from each haproxy host13:42
lsudreI think I have only one haproxy host13:44
lsudreIt's pretty difficult for me to understand everything of the OSA stack. Before trying to use OSA I installed nova/cinder/horizon/placement on 3 vms (controller,compute,storage host) Now I should have proxy, with LB I don't understand a lot of things in OSA. I was thinking to have one host 172.16.12.69 for the haproxy would be enough13:46
lsudreIt's a openstack_user_config problem?13:49
jrosserlsudre: did you start by building an OSA all-in-one?13:51
lsudreno13:52
jrosserthe reason there is HAProxy is to do a few things, terminate a proper SSL certificate for the external endpoint, loadbalance between multiple infra hosts so that you get high availability backend services, and also to be able to run multiple HAProxy instances with keepalived (or other) failover between them, again to achieve H/A13:53
jrossera lot of this is optional, so if you don't want to have H/A then you can run with just one HAProxy13:53
lsudrejrosser: I want to be as close as possible to my final production infrastructure13:53
jrosserthen i would recommend that you have 3 infra hosts, each running HAProxy and use keepalived for the VIP13:54
jrosserbut a lot of these decisions depend on what you want to acheive, everyone has different requirements13:54
lsudrethis is my user_config: https://paste.openstack.org/show/bs11N8kgYuY2gx0gLMUi/13:55
jrosserOSA is very much a toolbox that lets you construct your own deployment, rather than having only one possible approach13:55
lsudreIt's very difficult to find good (and recent) ressources to help like a step by step tutorial with this mega tools13:57
jrosserthat is why the all-in-one exists13:57
jrosserhave a look at this https://docs.openstack.org/openstack-ansible/2023.1/user/aio/quickstart.html13:57
jrosserit will auto-generate the config for you in a single host test environment13:57
jrosserit should be possible to start with an empty VM / actual host and deploy a test environment in a couple of hours13:58
lsudre"I have already completed this step and installed OpenStack on different machines, tested integrating Ceph, added a second compute node, and ran VMs on both Ubuntu and Windows. Everything was perfect. Now, I want to make the deployment tool (OpenStack-Ansible) work to try to deploy OpenStack automatically on my infrastructure.14:01
anskiylsudre: you mean, you've installed OpenStack manually before, and now you're trying to use OSA, right? 14:05
lsudreyes14:06
opendevreviewMerged openstack/openstack-ansible-os_rally stable/yoga: Include proper commit in rally_upper_constraints_url  https://review.opendev.org/c/openstack/openstack-ansible-os_rally/+/88768114:06
anskiycould you please show the output of running `echo "show stat" | nc -U /run/haproxy.stat | grep repo_all-back` (the command from before) from your `lb1` host -- the host with HAProxy.14:07
lsudreanskiy: here: https://paste.openstack.org/show/bDb6U9JFDnqtYRtLdvI5/14:09
noonedeadpunklsudre: yeah, so somehow haproxy do not mark repo backend as UP14:10
anskiylsudre: there is `no route to host` error14:10
noonedeadpunkI kinda wonder if the IP for infra1 is correct in haproxy14:11
anskiylooks like correct, according to https://paste.openstack.org/show/bs11N8kgYuY2gx0gLMUi/14:11
noonedeadpunkbut then why haproxy_hosts has same IP as internal_lb_vip_address?14:12
noonedeadpunklsudre: ^14:12
lsudreMaybe because I don't understand well what im doing14:12
noonedeadpunkinternal_lb_vip_address is supposed to be vritual address, that can failover between hosts14:13
noonedeadpunkIt's brought up by keepalived14:13
jrosserwell in this case there is only one haproxy?14:13
jrosserso no keepalived is deployed14:13
noonedeadpunkYeah, but that's conceptually weird14:13
lsudreI think14:13
noonedeadpunkI'm not saying that it is root cause, but it will be wrong for real deployment14:14
jrosserlsudre: do any of the high availabilty things matter for you?14:14
lsudreyes14:14
noonedeadpunkjrosser: I'm also not sure if that is same host or not as well14:14
jrosseryeah14:14
lsudreright now im trying to use OSA in proxmox env with an host and vms 14:14
jrosseroh we have people have all sorts of difficulty with this approach before :/14:15
noonedeadpunklsudre: so, is haproxy is a separate VM there or it's same?14:15
lsudre172.16.12.69 is a separate vm14:15
noonedeadpunkok, then this VM should have yet another IP that will be in the same network as 172.16.12.6914:16
noonedeadpunkbut not used as `internal_lb_vip_address`14:16
noonedeadpunkideally14:16
noonedeadpunkand then, you should ensure that you can access 172.16.12.71 from VM 172.16.12.6914:17
jrosseralso given that this is proxmox i would do some really basic tests that you can ping the IP of the other VM from the haproxy VM14:17
noonedeadpunkas seems there's smth wrong with networking between these 214:17
jrosserand that the traffic actually runs properly br-mgmt -> br-mgmt and not accidentally via some default route or other14:18
jrosserlsudre: if you say that the high-availabity things are important - then you should have 3 infra nodes, and run haproxy on each of them, not necessarily as a separate node14:20
lsudreI made a basic network diagram: https://paste.openstack.org/show/bBchQGS4EoVQ37NfagI9/14:24
jrosserwell, from HAProxy perspective you have a `no route to host` error14:27
jrosserthis is the most obvious problem to address first14:27
lsudrewhere I can investigate?14:31
jrosserwell as i said, this is basic networking checks from the haproxy node14:36
jrossercan you ping the .69 address from there, does to routing table look like you expect, etc14:36
jrosseroh not .69, whatever the br-mgmt IP of the infra node is14:36
lsudreI found a pb in my controller node, the br-mgmt bridge has now multiples ip and one is the haproxy node https://paste.openstack.org/show/bgHrFwZNDsPTa5s6PPtG/14:38
jrosserhave you checked where actually haproxy is running? is it on the infra node for some reason?14:40
lsudreno, I think is a missconfiguration, my plan was to have the haproxy on a dedicated host the (.69)14:42
damiandabrowskinoonedeadpunk: sorry I missed your question15:16
damiandabrowskihmm, i didn't test it but in fact i think you are right and when you explicitly define tempest_projects_create: False, you won't be able create tempest resources because tasks will fail15:17
damiandabrowskithat's why its default value looks like: https://opendev.org/openstack/openstack-ansible-os_tempest/src/branch/master/defaults/main.yml#L6315:18
damiandabrowskiit would be good to cover this case here: https://opendev.org/openstack/openstack-ansible-os_tempest/src/branch/master/tasks/tempest_resources.yml#L26-L8415:19
noonedeadpunkwell, but you can kinda override tempest_projects and have them pre-created. IMO tempest_projects_create - that shouldn't be a requirement, but I am unsure15:20
damiandabrowskiah right, most likely you can define keystone_demo_tenant_id with your project id, but it doesn't have very obvious structure: https://opendev.org/openstack/openstack-ansible-os_tempest/src/branch/master/tasks/tempest_resources.yml#L17715:38
damiandabrowskii remember that it was quite hard to cover all potential use cases for tempest resource creation - but it's always worth trying :D 15:40
damiandabrowskinoonedeadpunk: I have a theory why 2023.1 upgrade jobs work fine on my local AIO but fail in CI and in fact it's related to https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/88815215:49
damiandabrowskiit was merged on 13.07.2023 when the problems started so it makes sense15:49
noonedeadpunkhaven't I pointed to it earlier today?15:49
damiandabrowskiyes, but you didn't share details, so:15:50
noonedeadpunk`[11:11] <noonedeadpunk> specifically this patch https://opendev.org/openstack/openstack-ansible-plugins/commit/88a8bfcd62fe7bb027ca7a8636fbe943bfda88c1`15:50
damiandabrowskimy AIO uses e6ce7a42e1282d3b80f422feacf32f6301bbc6b1 version of openstack-ansible-plugins - that's expected behavior15:50
damiandabrowskibut for some reason, CI seems to use 88a8bfcd62fe7bb027ca7a8636fbe943bfda88c1: https://zuul.opendev.org/t/openstack/build/4eb51e63841e4eea928c4292987e7324/log/job-output.txt#509915:51
noonedeadpunkthat's actually also expected15:51
noonedeadpunkwe test "latest" on the branch15:51
damiandabrowskiah, i had no idea about it15:51
noonedeadpunkso that depends-on work properly and we detect when we break things before bump15:52
noonedeadpunkBut that doesn't explain why it fails. Yes, now we do not forcefully update password on upgrade15:52
noonedeadpunkbut why it becomes invalid - that's the question15:52
noonedeadpunkso during password upgrade password can get re-hashed or smth15:53
noonedeadpunkbut it's kinda weird15:53
damiandabrowskifor me it makes sense now15:53
noonedeadpunkthere might be more issues then regarding idempotency of keystone playbooks15:54
damiandabrowskiso with service_update_password: true passwords get rehashed and everything works fine15:54
noonedeadpunkyup15:54
damiandabrowskibtw. do you remember where exactly do we override collection requirements in CI?15:54
damiandabrowskii'm curious15:55
noonedeadpunkbut this does not explain why it doesn't work without that and what exactly breaks 15:55
noonedeadpunkwe use zuul prepared repos with required_projects15:55
noonedeadpunkand then here15:55
noonedeadpunkhttps://opendev.org/openstack/openstack-ansible/src/branch/master/scripts/get-ansible-collection-requirements.yml#L59-L6715:55
damiandabrowski"but this does not explain why it doesn't work without that and what exactly breaks" because password may have more than 54 characters and since 2023.1 keystone trims it? that's my theory15:56
noonedeadpunknah, it trims since Pike. Like it always was trimming them15:56
noonedeadpunkAnd then non-upgrade jobs work nicely with trimming15:57
noonedeadpunkas well as zed/antelope15:57
noonedeadpunkand I saw alerts on Xena that password is too long, so it was trimmed15:57
*** me is now known as Guest704715:58
damiandabrowskium, this one was implement in 2023.1: https://opendev.org/openstack/keystone/commit/3288af579de8ee312c36fb78ac9309ce8c55482715:58
damiandabrowskiimplemented*15:58
Guest7047hi team i am upgrading from 24.1.0 to 25.4.0 , and i am getting this error15:58
Guest7047how to fix it TASK [python_venv_build : Slurp up the constraints file for later re-deployment] **************************************************************************** fatal: [t1w_nova_api_container-0f76010b -> t1w_repo_container-1bd52331(172.29.239.196)]: FAILED! => {"changed": false, "msg": "file not found: /var/www/repo/os-releases/25.4.0/ubuntu-20.04-x86_64/requirements/nova-25.4.0-constraints.txt"}15:59
noonedeadpunkGuest7047: have repo-install.yml finished without issues?16:00
noonedeadpunkah16:01
noonedeadpunkit's constraint file16:01
noonedeadpunknot u-c16:01
damiandabrowskinoonedeadpunk: so if i understand correctly: before 2023.1 password was truncated to `CONF.identity.max_password_length` and since 2023.1 it is truncated to `BCRYPT_MAX_LENGTH` (if `CONF.identity.password_hash_algorithm == 'bcrypt'`)16:02
noonedeadpunkah16:02
noonedeadpunkdamiandabrowski: have you change somewhere handy?16:03
noonedeadpunkGuest7047: try to run openstack-ansible os-nova-install.yml -e venv_rebuild=true16:03
damiandabrowskiwhat? :D 16:04
noonedeadpunkdamiandabrowski: as wondering when exactly that change landed16:04
Guest7047ok16:04
damiandabrowskithis one? https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/88815216:05
damiandabrowskior this one? https://review.opendev.org/c/openstack/keystone/+/82859516:05
noonedeadpunksecond :D16:05
noonedeadpunkso.... upgrades to Antelope should also fail then16:06
damiandabrowskiand they fail? that is our main issue :D 16:09
damiandabrowskiexample: https://zuul.opendev.org/t/openstack/build/4eb51e63841e4eea928c4292987e7324/16:09
noonedeadpunkah, ok then :D16:25
noonedeadpunkI'd say - maybe we set algo to scrypt and add update_password variable?16:26
noonedeadpunk*to upgrade scripts16:26
damiandabrowskifor master - yes, but for 2023.1 I suggest to revert https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/88815216:26
damiandabrowskiI'm just writing commit message to explain it more clearly16:27
damiandabrowskiwhat do you think?16:27
noonedeadpunkwell, you've seen the bug it was fixing, right?16:27
damiandabrowskiyeah, but i think it's either reverting fix for this bug for 2023.1 or change encryption algorithm for stable release16:29
damiandabrowskiidk, i'm think with both solutions16:29
damiandabrowskibut changing encryption alg for stable release also doesn't sound good :D 16:29
noonedeadpunkI'm not saying about changing algo for stable branch16:30
noonedeadpunkbut defining update_password for upgrade process is fair requirement16:30
noonedeadpunkfor major upgrades at least16:30
noonedeadpunkyou still will get all services restarted16:30
noonedeadpunkand regardless of that - this is potentially prolonged API downtime we need to cover in release notes16:31
damiandabrowskiah, makes sense then16:31
Guest7047still it failed with the same error16:32
noonedeadpunkI'm not saying what's best to do though, but there's totally smth to check on16:32
noonedeadpunkGuest7047: can you check that file `/var/www/repo/os-releases/25.4.0/ubuntu-20.04-x86_64/requirements/nova-25.4.0-constraints.txt` is not present in any of your repo containers?16:34
noonedeadpunkassuming you have 3 of them - worth checking all16:34
noonedeadpunkAs I wonder if that is related to potentiall issues with gluster or smth16:34
noonedeadpunkalso would be helpful to get, say output including previous 10-15 tasks 16:35
opendevreviewMerged openstack/ansible-role-zookeeper master: Do not use notify inside handlers  https://review.opendev.org/c/openstack/ansible-role-zookeeper/+/88876016:35
Guest7047it is there16:36
Guest7047-rw-r--r-- 1 nginx www-data  162 Jul 26 14:48 nova-25.4.0-requirements.txt16:36
Guest7047sry that file is not there16:37
Guest7047nova-25.4.0-constraints.txt16:37
noonedeadpunkyou can paste output using https://paste.openstack.org/ just in case16:38
noonedeadpunkbut having more output from the role would be awesome to understand what was going on before that lead to the issue16:38
Guest7047so what should we do , if it is not there as for other services it is there, except neutron,heat which still has to run17:07
noonedeadpunkum, I can't say without seeing more output17:15
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Add openstack_resources role skeleton  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/87879417:22
Guest7047ok i will copy paste the error in the link 17:22
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_tempest master: Adopt for usage openstack_resources role  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/88974117:25
noonedeadpunkjrosser: damiandabrowski That's first attempt to use openstack_resources ^ Likely will fail, but that should be a minor failure. 17:26
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Add openstack_resources role skeleton  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/87879417:27
Guest7047i have paste the complete error17:28
noonedeadpunkwhere?17:30
noonedeadpunkwhen you paste a unique link is generated17:30
* noonedeadpunk checking out for todat17:31
Guest7047in https://paste.openstack.org/ 17:33
Guest7047Paste #bxNKPk2jBDgJuaP9rR8x17:33
noonedeadpunkGuest7047: I'm interested in previous tasks. Can you supply whole run of the role to the paste?17:35
noonedeadpunkI got that file is not there, but it does not disclose why it's not there and what tasks run and what were skipped during the run17:35
Guest7047i will try , but it is too big if i enable verbose17:40
Guest7047There is a very good article written on that, but not sure17:43
Guest7047how to get this in to action17:43
Guest7047https://bugs.launchpad.net/openstack-ansible/+bug/198950617:43
Guest7047Paste #bLSrpnlu2pEUQvJ0MzsI17:45
* damiandabrowski will upload changes to fix stable/2023.1 upgrade jobs later this evening17:46
noonedeadpunkGuest7047: ok, try pls `openstack-ansible os-nova-install.yml -e venv_rebuild=true venv_wheels_rebuild=true`17:55
Guest7047ok17:57
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_tempest master: Adopt for usage openstack_resources role  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/88974118:01
noonedeadpunkGuest7047: oh sorry18:02
noonedeadpunk`openstack-ansible os-nova-install.yml -e venv_rebuild=true -e venv_wheels_rebuild=true`18:02
noonedeadpunkI've forgotten another `-e` 18:02
Guest7047yeah it is running, i have added it18:04
Guest7047i will update you may be trmw since it will take time18:04
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_tempest master: Adopt for usage openstack_resources role  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/88974118:40
-opendevstatus- NOTICE: The Gerrit service on review.opendev.org will be offline briefly for a minor upgrade at 21:00 utc, approximately an hour from now20:02
-opendevstatus- NOTICE: The Gerrit service on review.opendev.org will be offline briefly for a minor upgrade, but should return shortly21:00
opendevreviewDamian DÄ…browski proposed openstack/openstack-ansible master: Fix issues with truncated keystone passwords  https://review.opendev.org/c/openstack/openstack-ansible/+/88978121:19
opendevreviewDamian DÄ…browski proposed openstack/openstack-ansible stable/2023.1: Fix issues with truncated keystone passwords  https://review.opendev.org/c/openstack/openstack-ansible/+/88980121:28

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!