Tuesday, 2025-04-08

*** manuvakery1 is now known as manumoh02:19
user090Hi i like to get some guide to set up octavia with OVN networking , I am not able to find proper docs for it, using rocky9 for deployment I have created bridge br-lbaas on compute nodes, but ovs system complaning that bridge exists , 05:57
user090https://pastebin.com/xFAqv7u405:57
user090This ismy user_config.yml file snippet05:57
noonedeadpunkhey06:01
noonedeadpunkuser090: so eventually for any octavia setup you need to meet 2 conditions: 1. have a lbaas network inside octavia lxc container 2. have the same network in Neutron available for VMS06:05
noonedeadpunkas basically you need to have access from API to VM via internal link06:05
user090@noonedeadpunk that means I need bridge with static ip addresses across physical nodes, I was running like that with old clusters with linux bridge I configured bridge interface on each node, same setup seems not working with ovn as ovs creating a bridge and there is duplicate error06:08
noonedeadpunkSo with ovs/ovn you need a bridge only on control plane06:19
noonedeadpunkand you never needed an IP on the bridge either06:20
user090noonedeadpunk: the amphora vm is created on hypervisor nodes right?  so how can they communicate from hypervisor without defining network i dont get it06:23
noonedeadpunkso if that network is a VLAN - you create a VLAN internal network in neutron06:26
noonedeadpunkit really kinda depends on how your cluster is setup and if you split interfaces06:26
noonedeadpunkbut if you expect same vlan ID for lbaas to be present on your current network mapping - you just create a neutron network and define it's uuid for octavia06:28
noonedeadpunkand then on control plane you configure the same vlan and add to the br-lbaas there06:29
noonedeadpunkso that's why I said you don't need to create brdiges on computes06:38
noonedeadpunkuser090: but if you have hyperconverged setup - it's a little bit more tricky - as you need to have both ovs and lxb bridge06:38
user090My control plane services are running on lxc but hypervisors are using metal install. 06:39
noonedeadpunkyeah, sure06:40
user090So neutron and nova compute services are metal install on compute nodes06:40
noonedeadpunkright06:41
user090So i have to delete all compute nodes bridges and keep br-lb only on control plane that should work right , octavia_provider_network_type: vlan and octavia_provider_segmentation_id: 6 do i have to declare them or I just manually create br-lb and connect to bond0.6 06:42
user090Thanks let me try again06:51
noonedeadpunkuser090: sorry, was a bit afk. so octavia_provider_network_type and octavia_provider_segmentation_id are controlling neutron network07:19
noonedeadpunkyou can create network in neutron manually and define `octavia_service_net_setup: false octavia_neutron_management_network_uuid: $UUID` instead of these07:20
noonedeadpunkbr-lb on control plane is independent process from this07:20
noonedeadpunkand yes, you either need to create the bridge manually, or you can leverage systemd_networkd role for it's configuration07:22
noonedeadpunkie https://docs.openstack.org/openstack-ansible/latest/user/network-arch/example.html#configuring-network-interfaces07:22
user090noonedeadpunk: Well i successfully completed playbook but loadbalancer is in Error state probably some config issue management network created automatically,  octavia-worker complaining : Failure: octavia.common.exceptions.ComputeWaitTimeoutException:07:45
noonedeadpunkwell, that sounds like connectivity issue between controller and amphora...07:47
noonedeadpunkI'd suggest to create some cirros VM on the network for debugging first07:47
user090Well here is my doubt as by the conf I dont have vlan 6 on hypervisors so how can control plane communicate with amphora vms , I dont understand the flow07:58
user090br-lb bridge currently only on control plane07:58
derekokeeffe85Morning all, just wondering if someone could give me a pointer as where to start looking to solve this https://paste.openstack.org/show/bUzHTJxoiKgbi5F96PsB/ setup_infrastructure completed with no errors 07:59
f0oderekokeeffe85: the command is no longer 'mysql' but 'mariadb'08:09
derekokeeffe85Ah thank you08:09
f0othis changed in upstream packaging a few versions ago, could be that the docs didnt reflect this.. could also be that some OS have the change and others still use mysql for compatibility as symlink... it's a bit of a mess when upstream yolo-change binaries like that08:10
derekokeeffe85Just looking at the docs now to see if it was my mistake probably is) Just replace mysql with mariadb08:11
derekokeeffe85for the command to run?08:11
f0oyeah08:12
f0ofunny enough the mariadb binary still uses .mysql_history for CLI history... shows you how that rename still drags legacy08:14
derekokeeffe85Perfect, thanks f0o. That ran. Time to run setup_openstack now,  fingers crossed :)08:15
f0o:)08:20
noonedeadpunkI think I proposed doc updates recently to cover that 08:27
derekokeeffe85getting this error now f0o https://paste.openstack.org/show/bCmfrevQ1jARL8IyMSWJ/ something to do with mariadb over mysql? I cloned 30.0.1 following 2024.2 deployment guide08:27
noonedeadpunkderekokeeffe85: try running the same with `-e _oslodb_setup_nolog=False`08:28
noonedeadpunkyeah, patch to docs landed now: https://review.opendev.org/c/openstack/openstack-ansible/+/94256908:29
f0oderekokeeffe85: no that reads more like the containers arent running or there's some other issue with SSH'ing into them08:29
noonedeadpunkprobably worth backporting to 2024.2...08:29
derekokeeffe85So like this noonedeadpunk openstack-ansible openstack.osa.setup_openstack -e _oslodb_setup_nolog=False -vvv08:29
derekokeeffe85Having issues with MTU and ssh as ansible but thought I had those sorted08:29
noonedeadpunkyeah08:30
derekokeeffe85Will do08:30
noonedeadpunkas this will unveal the real issue which is hiding behind no_log08:30
noonedeadpunkbut also that will log your passwords in plain text to the output log08:30
derekokeeffe85Perfect08:30
Core9525any possible way to remove/change overcommitment ratio from 3.1 to 2.1?08:31
Core9525or to 1.2:108:31
Core9525without impact on running env?08:32
Core9525team any suggestions08:32
derekokeeffe85This is the output https://paste.openstack.org/show/bkRhe3n8TGOYtUXkjeeZ/08:32
f0oCore9525: https://github.com/openstack/openstack-ansible-os_nova/blob/master/defaults/main.yml#L345-L35008:35
f0odo you mean those?08:35
Core9525f0o yeah, right now my allocation is 3:1 for both ram and cpu, the workloads are running already, can we change this to lower valur withot impacting the workload.08:38
f0oyeah you can change that at will, it is only considered on scheduling and not for preexisting workload08:38
f0oif you issue a migration of an instance then it will be taken into account08:39
Core9525f0o++08:41
Core9525thanks08:41
noonedeadpunkderekokeeffe85: and are you able to use `mysql` from this utility container?08:44
noonedeadpunkAs still somehow looks like mtu stuff08:44
noonedeadpunkor well. `mariadb` :D08:44
Core9525any one using or tried ironic service implemented?08:46
noonedeadpunkCore9525: I know that jrosser using it, so yes08:47
Core9525noonedeadpunk++08:48
jrosserhello o/08:48
f0owhile we talk about compute stuff; why was the watcher role abandoned?08:49
f0ojust a matter of too little time and nobody using it?08:49
noonedeadpunkf0o: back in the days there we talks about deperecation of watcher as a whole, and nobody was interested in maintaing the role08:49
noonedeadpunklast PTG we discussed about reviving the role, but pretty much lacked time08:49
noonedeadpunkso it's on the table to bring it back08:49
f0oand then broadcom came around and revived it indirectly haha08:50
noonedeadpunkyeah....08:50
noonedeadpunkSo if you want/have time to look into it - I am happy to sort out administrative burden of getting repo back and adding it back to ci/governance08:51
f0ounfortunately my ansible skill is far below usable here :/08:52
f0oI can do small patches here and there but ansible as a whole is still black magic to me08:52
noonedeadpunkare you interested to use/test watcher if we'd bring it back?08:52
f0oyeah that's a no-brainer for me/us08:53
noonedeadpunk++ 08:53
jrosserCore9525: ironic via openstack-ansible is usable, but ironic itself is very flexible so you have to understand and design what you want, then use openstaack-ansible to deploy that08:53
jrosserit's certainly not "deploy the defaults and it all just works"08:54
derekokeeffe85Sorry noonedeadpunk got called away, willbe back in a bit to look into what you said08:54
Core9525jrosser i am going through the documts openstack-ansible role for ironic, looks straigh, let me start work this, if stuck ill ask for the help.08:56
Core9525jrosser++08:56
Core9525thanks08:56
derekokeeffe85noonedeadpunk I can scp a 3GB file between nodes (my test for MTU issues) so I don't think it's MTU. Also, I can't run mysql on the utility container. bash: mysql: command not found10:03
noonedeadpunkmariadb10:22
derekokeeffe85Nope10:23
derekokeeffe85ERROR 2013 (HY000): Lost connection to server at 'handshake: reading initial communication packet', system error: 1110:26
noonedeadpunkderekokeeffe85: have you checked the ping?10:51
noonedeadpunkwith big packet size?10:51
noonedeadpunkie `ping -s 1472 -M do`?10:53
noonedeadpunkbasically - 1472 is expected MTU - 2810:53
derekokeeffe85Yep that ping works across all the nodes to each other noonedeadpunk11:01
noonedeadpunkand you have MTU 1500 set universally in containers?11:06
noonedeadpunkI need to work on publishing network testing script....11:07
derekokeeffe85MTU is 9000 across the board on all interfaces - ping -s 8472 -M do works from all the hosts to each other. Oh I didn't check the MTU on the containers11:07
noonedeadpunkAs it does quite nice job, except taking eternity to finish11:07
noonedeadpunkso containers likely have 1500 out of the box11:07
noonedeadpunkyou'd need to set a different MTU explicitly there11:07
derekokeeffe85so on my mgmt subnet interfaces that are on the containers it's set at 9000. The second interface om the 10.0.3.0 subnet are 1500, I didn't check all of them but the main services have them at those settings11:11
derekokeeffe85That would be a handy script to have :)11:12
noonedeadpunklet's me publish it on gist at least....11:17
opendevreviewDaniel Preussker proposed openstack/openstack-ansible-haproxy_server master: Add custom errorfiles  https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/94661711:18
f0ojrosser: this would be my very naive approach to adding custom errorfile entries into the global portion11:18
noonedeadpunkderekokeeffe85: https://gist.github.com/noonedeadpunk/b9eee2331a3c732e4def0b97530940ba11:19
noonedeadpunkyou'd need to launch test_networks.yml11:19
noonedeadpunkyou can set `-e openstack_tested_networks_throughput=false` to skip iperf11:21
derekokeeffe85So use openstack-ansible test_networks.yml -e openstack_tested_networks_throughput=false` to skip iperf is it noonedeadpunk11:26
noonedeadpunkyeah11:33
derekokeeffe85Running now11:34
f0orevisiting #939601 - is there any action required from me? I see it wants a rebase12:20
f0o(https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/939601)12:21
derekokeeffe85noonedeadpunk I have to run again but when I get back (maybe tomorrow) I'll let you know. Think I may have found an issue with my setup that your script highlighted12:46
mgariepynot sure what can be wrong but i can't find the reason why i cannot schedule a vm in nova after an upgrade to Xena.17:58
mgariepyNoValidHost: No valid host was found.17:58
*** jamesdenton_ is now known as jamesdenton18:52
kstev@noonedeadpunk i am preparing a Yoga to Zed upgrade and notice the release notes mention "you must ensure that following variables are defined explicitly to ensure parity with existing functionality" including neutron_plugin_base: - metering ... i have the 'router' and 'neutron_dynamic_routing.services.bgp.bgp_plugin.BgpPlugin' plugins defined currently. i noticed define-neutron-plugin.yml appears to add the metering plugin. must meteri19:04
kstevng be added? i'd rather not add it if we do not have to19:04
kstevhttps://docs.openstack.org/releasenotes/openstack-ansible/zed.html19:07
noonedeadpunkmgariepy: any logs from scheduler?19:41
mgariepysome issue with placement aggregate19:42
mgariepycompute host in aggrete have : `compute.cloud.domain.com` but hypervisor list does only have `compute`19:43
mgariepynew deployment not upgraded from ubuntu 14.04 since 2014..19:46
mgariepyalso the placement request do have : `member_of=in:d586eea7-878e-4131-b96f-d3af2a0c29b1,e48a8749-4a4e-47c2-a80d-0646363ab9ad`19:49
mgariepywhich is 2 aggrate that i have.19:49
mgariepybut it's not from the flavor.19:50
noonedeadpunkwell, aggregates in placement and in nova not necessarily synced19:52
noonedeadpunkespecially in older releases19:52
mgariepyyeah.19:52
noonedeadpunkI do recall times when I had to manage them independently19:52
jamesdenton@kstev i'm not aware of metering being a requirement, but not sure why it would've been included other than it being defined in an AIO19:52
jamesdentonhttps://github.com/openstack/openstack-ansible/blob/a1f47e174343573efb17ab9e56082faade55dee4/tests/roles/bootstrap-host/templates/user_variables.aio.yml.j2#L14819:53
noonedeadpunkjamesdenton: I think it used to be the default19:53
jamesdentonahh maybe19:53
noonedeadpunkhttps://opendev.org/openstack/openstack-ansible-os_neutron/src/tag/yoga-eol/defaults/main.yml#L348-L35019:53
jamesdentonoh, well there you go19:54
noonedeadpunkso in release note and in upgrade script I just wrote defaults if variable is not defined19:54
jamesdentonmakes sense!19:54
noonedeadpunkif neutron_plugin_base is defined for group/host vars - upgrade script will do nothing19:54
noonedeadpunk(at least it should)19:54
noonedeadpunkhttps://opendev.org/openstack/openstack-ansible/src/tag/zed-eom/scripts/upgrade-utilities/define-neutron-plugin.yml#L4819:55
jamesdenton19:56
jamesdentonthank you noonedeadpunk 19:56
noonedeadpunkthanks for stepping in with the reply :) long time no see :)19:57
noonedeadpunkwas missing you around (▰˘◡˘▰)19:57
jamesdenton:)19:58
jamesdentonit has been a minute! i blame my bouncer19:58
jamesdentonand my job :P19:58
mgariepyis there a tool to sync nova and placement ?19:59
mgariepyor i need to rename the hosts... to get the automatic sync..20:00
mgariepyi suspect it's because of the fqdn issue for some/most of the node.20:00

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!