*** luksky has quit IRC | 00:12 | |
*** trident has quit IRC | 00:13 | |
*** trident has joined #openstack-lbaas | 00:14 | |
*** mithilarun has joined #openstack-lbaas | 00:16 | |
*** goldyfruit_ has quit IRC | 00:34 | |
*** goldyfruit_ has joined #openstack-lbaas | 00:34 | |
*** mithilarun has quit IRC | 01:01 | |
*** mithilarun has joined #openstack-lbaas | 01:02 | |
openstackgerrit | Merged openstack/octavia master: Use LUKS1 for certfs-ramfs for all distributions https://review.opendev.org/698885 | 01:14 |
---|---|---|
*** yamamoto has joined #openstack-lbaas | 01:16 | |
*** goldyfruit_ has quit IRC | 01:21 | |
*** goldyfruit has joined #openstack-lbaas | 01:21 | |
*** mithilarun has quit IRC | 01:41 | |
*** yamamoto has quit IRC | 02:19 | |
*** yamamoto has joined #openstack-lbaas | 02:25 | |
*** goldyfruit has quit IRC | 02:25 | |
*** vishalmanchanda has joined #openstack-lbaas | 02:25 | |
*** armax has joined #openstack-lbaas | 02:40 | |
*** yamamoto has quit IRC | 03:38 | |
*** psachin has joined #openstack-lbaas | 03:38 | |
*** goldyfruit has joined #openstack-lbaas | 03:38 | |
*** goldyfruit has quit IRC | 03:43 | |
*** goldyfruit has joined #openstack-lbaas | 03:43 | |
*** goldyfruit has quit IRC | 04:05 | |
*** goldyfruit has joined #openstack-lbaas | 04:11 | |
*** ramishra has joined #openstack-lbaas | 04:19 | |
*** goldyfruit has quit IRC | 04:27 | |
*** goldyfruit has joined #openstack-lbaas | 04:27 | |
*** goldyfruit has quit IRC | 04:33 | |
*** goldyfruit has joined #openstack-lbaas | 04:33 | |
*** yamamoto has joined #openstack-lbaas | 04:36 | |
*** armax has quit IRC | 04:37 | |
*** goldyfruit has quit IRC | 04:50 | |
*** goldyfruit has joined #openstack-lbaas | 04:50 | |
*** goldyfruit has quit IRC | 04:59 | |
openstackgerrit | Carlos Goncalves proposed openstack/octavia stable/train: Use LUKS1 for certfs-ramfs for all distributions https://review.opendev.org/704107 | 06:31 |
*** yamamoto has quit IRC | 06:40 | |
*** yamamoto has joined #openstack-lbaas | 06:43 | |
*** yamamoto has quit IRC | 06:58 | |
*** yamamoto has joined #openstack-lbaas | 07:08 | |
*** yamamoto has quit IRC | 07:15 | |
*** yamamoto has joined #openstack-lbaas | 07:19 | |
*** yamamoto has quit IRC | 07:29 | |
*** yamamoto has joined #openstack-lbaas | 07:40 | |
*** yamamoto has quit IRC | 07:41 | |
*** yamamoto has joined #openstack-lbaas | 07:45 | |
openstackgerrit | Ann Taraday proposed openstack/octavia master: Add option to set default ssl ciphers in haproxy https://review.opendev.org/685337 | 07:50 |
*** yamamoto has quit IRC | 07:58 | |
*** luksky has joined #openstack-lbaas | 08:02 | |
*** yamamoto has joined #openstack-lbaas | 08:03 | |
*** yamamoto has quit IRC | 08:09 | |
*** yamamoto has joined #openstack-lbaas | 08:09 | |
*** maciejjozefczyk has joined #openstack-lbaas | 08:10 | |
*** yamamoto has quit IRC | 08:12 | |
*** yamamoto has joined #openstack-lbaas | 08:16 | |
*** tkajinam has quit IRC | 08:20 | |
*** tesseract has joined #openstack-lbaas | 08:34 | |
*** rpittau|afk is now known as ironic | 08:51 | |
*** ironic is now known as rpittau | 08:51 | |
*** yamamoto has quit IRC | 08:55 | |
*** yamamoto has joined #openstack-lbaas | 08:57 | |
*** yamamoto has quit IRC | 08:58 | |
*** pcaruana has joined #openstack-lbaas | 09:14 | |
openstackgerrit | Gregory Thiemonge proposed openstack/octavia-tempest-plugin master: Add tests for mixed IP networks UDP members https://review.opendev.org/668619 | 09:28 |
openstackgerrit | Gregory Thiemonge proposed openstack/octavia-tempest-plugin master: Add UDP test scenario https://review.opendev.org/656515 | 09:28 |
*** yamamoto has joined #openstack-lbaas | 09:41 | |
*** yamamoto has quit IRC | 09:45 | |
openstackgerrit | Merged openstack/octavia master: Transition amphora flows to dicts https://review.opendev.org/668898 | 09:51 |
openstackgerrit | Merged openstack/octavia master: Convert Lb flows to use provider dicts https://review.opendev.org/671725 | 09:54 |
openstackgerrit | Merged openstack/octavia master: Add logging filter for AmpConnectionRetry exception https://review.opendev.org/700553 | 09:54 |
*** yamamoto has joined #openstack-lbaas | 09:55 | |
*** yamamoto has quit IRC | 10:03 | |
*** yamamoto has joined #openstack-lbaas | 10:12 | |
openstackgerrit | Merged openstack/octavia stable/train: Add listener and pool protocol validation https://review.opendev.org/701850 | 10:36 |
openstackgerrit | Merged openstack/octavia stable/stein: Add listener and pool protocol validation https://review.opendev.org/702043 | 10:38 |
*** yamamoto has quit IRC | 10:57 | |
*** yamamoto has joined #openstack-lbaas | 10:59 | |
*** yamamoto has quit IRC | 11:01 | |
*** yamamoto has joined #openstack-lbaas | 11:02 | |
*** yamamoto has quit IRC | 11:19 | |
*** luksky has quit IRC | 11:23 | |
*** yamamoto has joined #openstack-lbaas | 11:24 | |
*** maciejjozefczyk_ has joined #openstack-lbaas | 11:28 | |
*** rpittau is now known as rpittau|bbl | 11:28 | |
*** maciejjozefczyk has quit IRC | 11:30 | |
*** yamamoto has quit IRC | 11:42 | |
*** yamamoto has joined #openstack-lbaas | 11:45 | |
*** xgerman has quit IRC | 11:45 | |
*** xgerman has joined #openstack-lbaas | 11:45 | |
*** ccamposr__ has quit IRC | 11:49 | |
*** yamamoto has quit IRC | 11:56 | |
*** luksky has joined #openstack-lbaas | 12:01 | |
*** maciejjozefczyk_ has quit IRC | 12:23 | |
*** yamamoto has joined #openstack-lbaas | 12:36 | |
*** maciejjozefczyk_ has joined #openstack-lbaas | 12:37 | |
*** yamamoto has quit IRC | 12:43 | |
openstackgerrit | Ann Taraday proposed openstack/octavia master: Convert health_monitor flows to use provider models https://review.opendev.org/679779 | 13:02 |
openstackgerrit | Ann Taraday proposed openstack/octavia master: Transition l7policy flows to dicts https://review.opendev.org/665977 | 13:02 |
openstackgerrit | Ann Taraday proposed openstack/octavia master: Transition l7rule flows to dicts https://review.opendev.org/668173 | 13:02 |
openstackgerrit | Ann Taraday proposed openstack/octavia master: Jobboard based controller https://review.opendev.org/647406 | 13:02 |
openstackgerrit | Ann Taraday proposed openstack/octavia master: Jobboard based controller https://review.opendev.org/647406 | 13:08 |
*** rpittau|bbl is now known as rpittau | 13:40 | |
luksky | johnsom - it didn't go away :/ | 13:40 |
luksky | johnsom - and one more strange thing, after upgrading every piece of octavia to stable/queens I get situation where LB is in OFFLINE state, healthmonitor is in ACTIVE, and member is in OFFLINE, but... LB is working: | 13:47 |
luksky | http://paste.openstack.org/show/788753/ | 13:47 |
cgoncalves | luksky, hi. that should be because your member is in OFFLINE operating status and since that is the only member associated to the load balancer, the LB also reports OFFLINE | 13:51 |
luksky | octavia version: 2.1.2, and amphora-agent 2.1.3 | 13:52 |
cgoncalves | luksky, I'd check if the member is indeed up and the health monitor configured properly (e.g. the expected code from the member response is 200) | 13:52 |
luksky | ok, but all is working | 13:52 |
luksky | checked logs from member: | 13:52 |
luksky | 195.167.157.117 - - [24/Jan/2020:13:52:45 +0000] "GET / HTTP/1.0" 200 612 "-" "-" | 13:52 |
luksky | 195.167.157.130 - - [24/Jan/2020:13:52:46 +0000] "GET / HTTP/1.0" 200 612 "-" "-" | 13:52 |
luksky | and the site is also shows up correctly :/ | 13:53 |
luksky | I have one more LB which didn't failover to new amphoras, and there member is ONLINE | 13:55 |
cgoncalves | hmm, in that case :/ | 13:56 |
cgoncalves | sorry, I don't follow your last comment | 13:56 |
luksky | I have two LBs, one of it I created after upgrade, and second was running during upgrade. | 13:58 |
luksky | the 'old' one is showing member status correctly | 13:59 |
luksky | every new LB - shows member as OFFLINE, LB is in status OFFLINE - *AND* new LBs are working :/ | 14:00 |
luksky | (old one also) | 14:01 |
luksky | I have also situation, where member has 'operating_status' "No monitor" - when I deleted it and add once more - operating_status changed to 'offline' - but LB *IS* working :/ | 14:03 |
cgoncalves | ok, I understand now, thanks. | 14:04 |
cgoncalves | I will stack Queens locally and see if I can reproduce | 14:04 |
cgoncalves | what I'm understanding is that the problem is specific to both queens controller and amphora images | 14:05 |
cgoncalves | queens controller + pike amphora works, right? | 14:05 |
luksky | checking, please wait | 14:06 |
luksky | version of controler: | 14:06 |
luksky | pip list | grep -i octavia | 14:06 |
luksky | octavia (2.1.2) | 14:06 |
luksky | it is more complicated that I thought, sorry :/ | 14:07 |
luksky | on one (the 'old' one) LB I have following two amphoras version (the one with 3.0.0.0b2.dev62) is from time where I have tested different version of images): | 14:08 |
luksky | ubuntu@amphora-253e6a3d-776a-4b9f-9c30-dd7dd67a8c4b:~$ amphora-agent --version | 14:08 |
luksky | %prog 3.0.0.0b2.dev62 | 14:08 |
luksky | ubuntu@amphora-70f56c40-f543-4826-a36e-b93c26f948df:~$ amphora-agent --version | 14:09 |
luksky | %prog 2.1.3.dev8 | 14:09 |
luksky | I will failover this 3.0.0 now | 14:09 |
luksky | the 3.0.0 was master | 14:10 |
cgoncalves | 3.y.z is Rocky release. backward compatibility is only with older versions in the amphora. say, Queens amphora + Rocky controller is expected to work. | 14:11 |
luksky | yes, understood | 14:11 |
luksky | don't want mess here, so here is paste | 14:15 |
luksky | from OFFLINE LB: | 14:15 |
luksky | http://paste.openstack.org/show/788755/ | 14:15 |
cgoncalves | amphorae running queens, good | 14:17 |
luksky | yes, the one LB which is online, (which had 3.0.0 amphora), is still online, but I added one more member, and this NEW member is ... OFFLINE - but is working ok | 14:18 |
luksky | https://pasteboard.co/IRsuQgL.png | 14:20 |
luksky | member with IP 106 is serving the content now | 14:21 |
cgoncalves | hmm. would you know why is admin state up = No for the 106 member? | 14:22 |
cgoncalves | in your previous paste (http://paste.openstack.org/show/788753/), the member has admin_state_up=True | 14:23 |
luksky | sory, not this screenshot | 14:23 |
luksky | yes, I put this admin = off | 14:23 |
luksky | yes, I put this admin state to No | 14:23 |
luksky | https://pasteboard.co/IRswrlp.png | 14:24 |
cgoncalves | you toggled now the admin state in both members | 14:25 |
cgoncalves | give the health monitor a few seconds to check for health | 14:25 |
luksky | I did - now I see 106 | 14:26 |
luksky | which is offline | 14:26 |
luksky | http://195.167.157.155/ - this main ip of LB | 14:26 |
luksky | I'm pretty sure, that if I delete 107 member and add him once again it will be OFFLINE, and will be working :) | 14:29 |
*** goldyfruit has joined #openstack-lbaas | 14:34 | |
cgoncalves | luksky, could you please share the haproxy config file in the amphora? | 14:34 |
luksky | yes | 14:35 |
cgoncalves | it should be in amphora@/var/lib/amphora/ | 14:35 |
luksky | from both ? | 14:36 |
luksky | http://paste.openstack.org/show/788756/ | 14:36 |
cgoncalves | luksky, whichever is the active amphora should be enough | 14:36 |
luksky | http://paste.openstack.org/show/788757/ - this MASTER, above is BACKUP | 14:37 |
cgoncalves | thanks. the configs look the same, good | 14:38 |
cgoncalves | btw I'm still spinning up a queens cloud here | 14:39 |
luksky | I deleted 107 from LB and added once more :D | 14:39 |
luksky | https://pasteboard.co/IRsCTUO.png | 14:40 |
luksky | messed up with amphoras sorry :( - here are configs for master and backup for LB with member 107 and 106: | 14:50 |
luksky | http://paste.openstack.org/show/788758/ | 14:50 |
*** psachin has quit IRC | 14:52 | |
*** goldyfruit_ has joined #openstack-lbaas | 14:54 | |
*** goldyfruit has quit IRC | 14:55 | |
*** goldyfruit_ has quit IRC | 14:55 | |
*** goldyfruit_ has joined #openstack-lbaas | 14:55 | |
*** coreycb has quit IRC | 14:57 | |
*** coreycb has joined #openstack-lbaas | 14:57 | |
openstackgerrit | Carlos Goncalves proposed openstack/octavia master: Add listener allowed_cidrs to the feature matrix https://review.opendev.org/703851 | 15:02 |
luksky | I think, this may be related to database changes/schema ? | 15:06 |
luksky | here is octavia-db-manage history: http://paste.openstack.org/show/788759/ | 15:06 |
*** goldyfruit_ has quit IRC | 15:07 | |
*** goldyfruit_ has joined #openstack-lbaas | 15:07 | |
cgoncalves | why do you think that? you seem to have latest db schema for queens | 15:08 |
cgoncalves | luksky, what does haproxy show in the journal? | 15:09 |
luksky | http://paste.openstack.org/show/788760/ | 15:10 |
*** yamamoto has joined #openstack-lbaas | 15:10 | |
luksky | from second amphora also ? | 15:11 |
cgoncalves | no need, only active | 15:12 |
cgoncalves | Server c30a65f6-4419-40ac-905a-684fac6b02b4/a6bad16b-5734-4fcd-b6d4-8d58e532f170 is UP, reason: Layer7 check passed, code: 200, info: "HTTP status check returned code <3C>200<3E>" | 15:13 |
cgoncalves | it was reported DOWN before | 15:13 |
luksky | here is master: http://paste.openstack.org/show/788761/ | 15:14 |
luksky | yes, 2 days ago... :) | 15:15 |
*** yamamoto has quit IRC | 15:15 | |
*** goldyfruit_ has quit IRC | 15:16 | |
*** goldyfruit_ has joined #openstack-lbaas | 15:17 | |
johnsom | luksky: This probably means you lb-mgmt-net is not working correctly. The amphora health messages may not be getting to the health manager process | 15:19 |
luksky | it is L2, separated vlan | 15:20 |
johnsom | If you enable debug in the config file and restart your health manager you should see debug log messages that say it received a message from an amphora IP | 15:21 |
luksky | all logging I made from octavia controller to amphoras | 15:21 |
luksky | ok | 15:21 |
johnsom | If you don’t see that, the health messages are not getting back to the controller. | 15:22 |
luksky | and configuration of network (neither on switches nor in neutron) wasn't touched | 15:22 |
luksky | 2020-01-24 16:22:50.632 30093 DEBUG octavia.amphorae.drivers.health.heartbeat_udp [-] Received packet from ('10.99.99.23', 42060) dorecv /usr/local/lib/python2.7/dist-packages/octavia/amphorae/drivers/health/heartbeat_udp.py:189 | 15:23 |
luksky | johnsom - this kind of messages ? | 15:23 |
johnsom | Hmm, ok, yes those | 15:23 |
johnsom | Oh! You said you had an hmac error right? | 15:24 |
johnsom | That is dropping them | 15:24 |
luksky | yes... still ... | 15:24 |
luksky | didn't find way to get rid of it | 15:24 |
luksky | 2020-01-24 16:24:20.729 30093 WARNING octavia.amphorae.backends.health_daemon.status_message [-] calculated hmac(hex=True): 66336161373264306366306165623638636532356631613436313962613466303161636534353132663861393865393337376265393239346539383630663565 not equal to msg hmac: 8c2c938c2d8d52cc8d924d9580aa8b530b81ca4d0c8d4dcd2ccc6a017bf214c1ab485727f956055a9b652e04d64661018aeecf9698430310bc3da57579adbe6a dropping packet | 15:24 |
johnsom | Yeah, ok, we need to figure out why that is happening. This is a security check. The controller isn’t trusting the health message so discarding them | 15:26 |
*** dulek has quit IRC | 15:28 | |
*** armax has joined #openstack-lbaas | 15:30 | |
johnsom | luksky Can you check that this configuration setting is the same on all of your controllers? https://docs.openstack.org/octavia/latest/configuration/configref.html#health_manager.heartbeat_key | 15:31 |
luksky | ok, I turned off two controllers, one remain | 15:31 |
johnsom | It needs to be the same in the worker, health manager, and housekeeping octavia.conf files. | 15:31 |
luksky | I have only one config file | 15:32 |
luksky | octavia.conf | 15:32 |
luksky | and there are subsections | 15:32 |
johnsom | Yeah, ok, so that key is set in the [health_manager] section of the octavia.conf right? | 15:33 |
luksky | grep heartbeat_key /etc/octavia/octavia.conf | 15:33 |
luksky | heartbeat_key = xxxxx | 15:33 |
luksky | yes, checked on amphora - it is the same | 15:34 |
johnsom | Ok, then if you ssh into one of the amphora, does that match the key in /etc/octavia/amphora-agent.conf? | 15:34 |
luksky | yes | 15:35 |
johnsom | Hmmm, ok, then give me a second to check the versions you are running again. | 15:35 |
luksky | 2.1.2 on controller, 2.1.3.dev8 on amphora | 15:36 |
johnsom | When was the last time you restarted the health manager porcess? | 15:37 |
luksky | oo, many times | 15:39 |
luksky | Active: active (running) since Fri 2020-01-24 16:22:34 CET; 17min ago | 15:39 |
*** dulek has joined #openstack-lbaas | 15:44 | |
johnsom | Hmmm, I am really not sure why you would still be seeing those hmac errors with those versions and if the keys all match. | 15:45 |
johnsom | Can you humor me and boot a fresh load balancer, in your current configuration and see if it is also logging these hmac errors? | 15:46 |
luksky | You mean to rebuild whole VM ? | 15:47 |
*** maciejjozefczyk_ has quit IRC | 15:48 | |
johnsom | Just create another load balancer from scratch, configure it the same as the other. | 15:48 |
luksky | aa OK | 15:48 |
luksky | I did it, twice | 15:48 |
luksky | same situation | 15:49 |
luksky | but, there may be some diffrences, will do it know once more time | 15:49 |
*** maciejjozefczyk has joined #openstack-lbaas | 15:49 | |
*** vishalmanchanda has quit IRC | 15:51 | |
*** vishalmanchanda has joined #openstack-lbaas | 15:51 | |
johnsom | Usually when I see this, one controller config file has a different key in the configuration file. So, when the amphora boots from controller B but the health message receives on controller A, the keys don't match and you get this hmac error. But we have kind of gone through that debug process. Now that you are down to one controller and all of the processes use the same configuration file, it should just work. | 15:52 |
*** gregwork has quit IRC | 15:55 | |
*** TrevorV has joined #openstack-lbaas | 15:55 | |
luksky | and now this new LB is working correctly (DON'T UNDURSTAND :/) | 15:56 |
luksky | this the new one | 15:56 |
luksky | https://pasteboard.co/IRt7VLL.png | 15:56 |
johnsom | Ok, so somewhere along the line we had a mis-match of something. | 15:57 |
luksky | will delete one of this member from this new shiny working LB and add it again | 15:57 |
luksky | it is online = works as expected | 15:58 |
luksky | will failover all other amphoras, and check now | 15:59 |
johnsom | Excellent! | 15:59 |
luksky | in other (not working) LBs | 15:59 |
luksky | thank You ! | 15:59 |
luksky | I have three amphora in BOOTING state - (for a long time) - how can I delete them? | 16:01 |
luksky | | 08b7d7f2-88f3-467a-9ee8-f77c3756d688 | None | BOOTING | None | 10.99.99.26 | None | | 16:02 |
luksky | | 319af4dd-c70e-402c-8685-83a363cef313 | None | BOOTING | None | 10.99.99.38 | None | | 16:02 |
luksky | | 9c9d7381-4ce2-4376-9bd5-8ee365d87152 | None | BOOTING | None | 10.99.99.24 | None | | 16:02 |
johnsom | Nova issues? | 16:02 |
luksky | yes... | 16:02 |
luksky | they are yp | 16:02 |
luksky | I mean - i can try to log in, but ssh key doesn't work | 16:03 |
luksky | somthing goes wrong with them | 16:03 |
johnsom | Well, normally these will stop retrying and timeout. However, I'm guessing some of the controller stuff you have been doing killed the controller that owned those. | 16:03 |
luksky | ok, can I delete this compute instances (simply by openstack server delete) ? | 16:04 |
johnsom | It is important that you don't kill -9 the controllers, only gracefully shut them down. (We are working on this bug/issue right now) | 16:04 |
johnsom | Yes, openstack server delete should be ok here. | 16:04 |
luksky | ok | 16:04 |
luksky | they will disappear from octavia database, or I need do some mysql clean up ? | 16:05 |
johnsom | If the controller is still retrying those, it will automatically move them to error and move on. | 16:05 |
johnsom | If not, you might need to delete them from the db (kill -9 scenario) | 16:06 |
luksky | ok, thx | 16:06 |
luksky | for i in `openstack loadbalancer amphora list | grep MASTER | awk -F"|" '{print $2}'`;do openstack loadbalancer amphora failover $i;done | 16:23 |
luksky | all LBs are online now :) | 16:23 |
*** yamamoto has joined #openstack-lbaas | 16:30 | |
*** maciejjozefczyk has quit IRC | 16:31 | |
*** luksky has quit IRC | 17:00 | |
openstackgerrit | Merged openstack/octavia master: Fix the interface filenames for Red Hat amphora images https://review.opendev.org/700513 | 17:08 |
*** maciejjozefczyk has joined #openstack-lbaas | 17:09 | |
johnsom | Ah, I am glad someone got time to research that and make sure it was right. I didn't have a devstack to load it up on to investigate this. | 17:10 |
openstackgerrit | Merged openstack/octavia stable/train: Fix multi-listener LB client auth/re-encryption https://review.opendev.org/698601 | 17:16 |
*** goldyfruit_ has quit IRC | 17:18 | |
openstackgerrit | Merged openstack/octavia stable/rocky: Add listener and pool protocol validation https://review.opendev.org/702044 | 17:21 |
*** maciejjozefczyk has quit IRC | 17:24 | |
*** tesseract has quit IRC | 17:27 | |
*** rpittau is now known as rpittau|afk | 17:35 | |
johnsom | FYI, the next ubuntu LTS, 20.04 is starting to ship packages. It has haproxy 2.0.12 in it at the moment. | 17:41 |
johnsom | https://packages.ubuntu.com/focal/haproxy | 17:41 |
*** servagem has quit IRC | 17:54 | |
*** yamamoto has quit IRC | 17:56 | |
*** yamamoto has joined #openstack-lbaas | 17:57 | |
*** mithilarun has joined #openstack-lbaas | 18:16 | |
*** yamamoto has quit IRC | 18:35 | |
*** psachin has joined #openstack-lbaas | 18:47 | |
openstackgerrit | Merged openstack/octavia stable/train: Revert "Fix gate issue with DIB" https://review.opendev.org/696165 | 19:07 |
*** yamamoto has joined #openstack-lbaas | 19:13 | |
*** psachin has quit IRC | 19:23 | |
*** yamamoto has quit IRC | 19:24 | |
*** luksky has joined #openstack-lbaas | 20:24 | |
*** goldyfruit_ has joined #openstack-lbaas | 20:26 | |
*** TrevorV has quit IRC | 20:55 | |
*** mithilarun has quit IRC | 20:59 | |
*** mithilarun has joined #openstack-lbaas | 20:59 | |
*** maciejjozefczyk has joined #openstack-lbaas | 21:04 | |
*** mithilarun has quit IRC | 21:04 | |
*** born2bake has joined #openstack-lbaas | 21:23 | |
*** maciejjozefczyk has quit IRC | 21:23 | |
openstackgerrit | Adam Harwell proposed openstack/octavia master: Update the lb_id on an amp earlier if we know it https://review.opendev.org/698082 | 21:26 |
openstackgerrit | Adam Harwell proposed openstack/octavia master: Conf option to use VIP ip as source ip for backend https://review.opendev.org/702535 | 21:26 |
openstackgerrit | Adam Harwell proposed openstack/octavia master: Allow AZ to override valid_vip_networks config https://review.opendev.org/699521 | 21:26 |
rm_work | i should probably poke people about https://review.opendev.org/#/c/604043/ again because it will get them in a situation similar to what we saw above with the heartbeat keys and HM issues (though for different reasons) | 21:30 |
johnsom | Yep | 21:32 |
rm_work | hmm rebasing tho and it looks like they may have adapted my code for the HM piece at least | 21:34 |
rm_work | although i did some dumb things in that patch | 21:39 |
rm_work | fixed | 21:42 |
johnsom | Now if it would create the lb-mgmt-net..... | 21:44 |
rm_work | wish i had any idea how to do that, lol | 21:44 |
johnsom | rm_work Question for the failover flow. Right now, we deploy the LB/listener(s), then deploy VRRP, then call ListenerStart. | 21:52 |
johnsom | The only thing ListenerStart does that matters is setup keepalived to monitor the deployed haproxy processes. | 21:52 |
johnsom | Now with single proc, there really is only going to be one ID to monitor. | 21:53 |
rm_work | so you could collapse that to just ... be part of the vrrp deploy? | 21:53 |
rm_work | guessing that's what you're getting at? | 21:53 |
johnsom | I am trying to decide if the haproxy monitor script needs to stay in ListenerStart (one more round trip) or if I should just merge it into the VRRP deploy code. | 21:53 |
rm_work | right | 21:53 |
rm_work | i would say you could probably combine them | 21:53 |
rm_work | a lot of that structure was a little.... | 21:54 |
rm_work | obtuse | 21:54 |
rm_work | err, though I wonder about UDP, is that the same? | 21:54 |
johnsom | Well, it was all over complicated because of that *#$% multi-proc decision | 21:54 |
rm_work | yeah :( | 21:54 |
johnsom | No, that has it's own checks | 21:55 |
johnsom | This is purely HAproxy | 21:55 |
rm_work | ah so we deploy keepalived but it doesn't need to handle checks? | 21:55 |
rm_work | i thought it still did the running of the check scripts | 21:55 |
rm_work | that's what it looked like to me when i was poking at it this week | 21:55 |
johnsom | It has it's own stuff | 21:56 |
rm_work | err, ok... | 21:56 |
johnsom | It's a whole different keepalived process | 21:56 |
rm_work | ah, yeah so still keepalived, but | 21:56 |
rm_work | different deployment mechanism for it? | 21:56 |
johnsom | YEs | 21:56 |
rm_work | k | 21:56 |
rm_work | then yeah, combine away | 21:56 |
johnsom | This literally just writes out a script with a list of haproxy PID files to go check | 21:57 |
johnsom | Just trying to think if we would ever have a Haproxy process create/start/stop independent of setting up the VIP in an Active/Standby topology | 21:58 |
johnsom | Nah, with single proc I think we are fine | 21:59 |
*** mithilarun has joined #openstack-lbaas | 22:08 | |
*** mithilarun has quit IRC | 22:10 | |
*** mithilarun has joined #openstack-lbaas | 22:10 | |
born2bake | what can be an issue if health-monitors, housekeeping and api have this errors: https://pastebin.com/9cbQs9pG | 22:31 |
born2bake | and lb cant get operating status online | 22:31 |
johnsom | That one is pretty straight forward: "Lost connection to MySQL server during query" | 22:34 |
*** rcernin has quit IRC | 22:34 | |
johnsom | There is something wrong with your mysql server or the network path from the controllers to that mysql server. | 22:34 |
openstackgerrit | Brian Haley proposed openstack/octavia-tempest-plugin master: Change to enable memory_tracker https://review.opendev.org/704202 | 22:34 |
born2bake | i have 3 controllers, thing is that neutron is installed only on 2 of them, octavia on 3 of them | 22:36 |
johnsom | Most Octavia deployments have three controller instances | 22:37 |
johnsom | If not more | 22:37 |
born2bake | that's weird cause other services are working fine and mariadb is fine as well | 22:38 |
johnsom | I guess that helps narrow it down, something wrong with the host/instance with the Octavia controllers. | 22:40 |
johnsom | All of those errors are oslo_db related, there is no Octavia code there. Other than the DB connection string, we don't set any oslo_db settings, it is all defaults. | 22:41 |
johnsom | Now, I don't know if kolla is doing something it shouldn't to oslo_db or adding oslo_db settings to our configuration files. | 22:42 |
rm_work | johnsom: ah i need to glance at my multi-vip patch (which i really need to rebase, but not looking forward to that) and see if your combination affects me there | 22:43 |
johnsom | You can check that no one added to our config, look at the [database] section in the octavia.conf. It should only contain a connection string setting. | 22:43 |
born2bake | worker and api logs are cool without errors | 22:46 |
born2bake | and when i am trying to open lb ip in browser got 503 Service Unavailable msg | 22:49 |
johnsom | Opening the Octavia LB VIP IP? That means there are no member servers healthy in the pool. | 22:55 |
openstackgerrit | Adam Harwell proposed openstack/octavia master: Support HTTP and TCP checks in UDP healthmonitor https://review.opendev.org/589180 | 22:57 |
*** born2bake has quit IRC | 23:01 | |
openstackgerrit | Brian Haley proposed openstack/octavia-tempest-plugin master: Create new octavia-dsvm-base-py3 job name https://review.opendev.org/704204 | 23:03 |
openstackgerrit | Brian Haley proposed openstack/octavia-tempest-plugin master: Change to enable memory_tracker https://review.opendev.org/704202 | 23:10 |
johnsom | haleyb Yeah, if nobody has it enabled by default, maybe we don't want it on by default either. Maybe it's more of a "turn it on if you have oom problems" type of thing | 23:18 |
haleyb | johnsom: tobiko did i think, not many others. like i mentioned, first step in getting job working again, guess i'll need a dependent patch to see | 23:20 |
johnsom | I still see peakmem enabled in the devstack repo too. But, yeah, I would maybe just delete that line all together | 23:22 |
openstackgerrit | Brian Haley proposed openstack/octavia master: DNM: Dependent patch for two-node job debugging https://review.opendev.org/704205 | 23:22 |
johnsom | A bunch of that was copied over from the main jobs since zuul has that issue with replacing instead of merging partent/child at certain levels. | 23:22 |
haleyb | johnsom: https://www.youtube.com/watch?v=qpGx4foRdPw | 23:23 |
haleyb | beer-o-clock here | 23:23 |
johnsom | o/ | 23:24 |
Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!