Thursday, 2019-08-15

*** hongbin has joined #openstack-lbaas00:58
*** ricolin has joined #openstack-lbaas01:03
*** yamamoto has quit IRC01:40
*** yamamoto has joined #openstack-lbaas01:40
*** altlogbot_2 has quit IRC03:16
*** altlogbot_0 has joined #openstack-lbaas03:17
*** hongbin has quit IRC03:27
*** yamamoto has quit IRC03:37
*** yamamoto has joined #openstack-lbaas03:45
*** yamamoto has quit IRC03:53
openstackgerritMerged openstack/octavia master: Switch TaskFlow engine to parallel  https://review.opendev.org/67637904:14
*** gcheresh has joined #openstack-lbaas04:50
*** tkajinam has quit IRC04:53
*** tkajinam has joined #openstack-lbaas04:54
*** tkajinam has quit IRC05:00
*** tkajinam has joined #openstack-lbaas05:05
*** tkajinam has quit IRC05:31
*** gcheresh has quit IRC05:45
*** gcheresh has joined #openstack-lbaas05:57
*** tkajinam has joined #openstack-lbaas06:04
*** yamamoto has joined #openstack-lbaas06:16
*** hongbin has joined #openstack-lbaas06:28
*** hongbin has quit IRC06:32
*** yamamoto has quit IRC06:36
*** yamamoto has joined #openstack-lbaas06:37
*** yamamoto has quit IRC06:42
*** tkajinam has quit IRC06:49
*** yamamoto has joined #openstack-lbaas06:54
*** yamamoto has quit IRC06:54
*** tkajinam has joined #openstack-lbaas07:00
*** ricolin_ has joined #openstack-lbaas07:03
*** ricolin has quit IRC07:05
*** devfaz has quit IRC07:33
*** ricolin_ is now known as ricolin07:33
*** devfaz has joined #openstack-lbaas07:49
*** ivve has joined #openstack-lbaas07:53
*** yamamoto has joined #openstack-lbaas08:26
*** yamamoto has quit IRC08:28
*** yamamoto has joined #openstack-lbaas08:34
*** yamamoto has quit IRC08:38
*** tkajinam has quit IRC08:47
*** yamamoto has joined #openstack-lbaas08:47
*** ricolin has quit IRC09:07
*** sapd1_x has joined #openstack-lbaas09:45
*** yamamoto has quit IRC10:00
*** hongbin has joined #openstack-lbaas10:22
*** gcheresh has quit IRC10:36
*** gcheresh has joined #openstack-lbaas10:42
*** yamamoto has joined #openstack-lbaas10:53
openstackgerritMerged openstack/octavia-tempest-plugin master: Set connection_retry_interval back up to 1200  https://review.opendev.org/67574510:54
*** yamamoto has quit IRC10:59
*** yamamoto has joined #openstack-lbaas11:06
openstackgerritMerged openstack/octavia master: Set Libvirt hw_rng for devstack  https://review.opendev.org/67601311:07
*** yamamoto has quit IRC11:10
*** hongbin has quit IRC11:16
*** yamamoto has joined #openstack-lbaas11:31
*** yamamoto_ has joined #openstack-lbaas11:41
*** yamamoto has quit IRC11:45
*** yamamoto_ has quit IRC12:20
*** luksky has joined #openstack-lbaas12:22
*** yamamoto has joined #openstack-lbaas12:28
*** ricolin has joined #openstack-lbaas12:37
*** hongbin has joined #openstack-lbaas12:43
*** yamamoto has quit IRC13:06
*** roukoswarf has joined #openstack-lbaas13:06
roukoswarfjohnsom: why does failover fail the healthy LB and not the error LB?13:07
roukoswarfkeeps getting deadlocked with error/standalone13:07
johnsomroukoswarf I'm confused, it only fails over unhealthy load balancers.13:09
roukoswarfpeople non stop break LBs and yell at me about it.13:09
johnsomAre you manually triggering a failover via the API?13:09
roukoswarfyes, because the LB is immutable because theres a failed amphora, and the failed amphora  is the master, and when i failover, it kills the healthy backup13:10
roukoswarfand now im stuck in master/error standalone/allocated13:10
roukoswarffailover causes a full outage and its deadlocked forever.13:11
johnsomAh, you meant amphora. Well, it tries to guess at which is the backup, but since the active/standby HA is autonomous and can switch inside the amphorae at any time, the controller doesn't know which is which explicitly. The bigger question is why is your failover not completing?13:11
roukoswarfnot sure, where can i check?13:12
johnsomWhat do you mean by deadlock?13:12
roukoswarfthe logs are silent.13:12
roukoswarfas in, any number of failovers or intentionally killing a node to make it go unhealthy and failover doesnt bring anything back13:13
johnsomfailover always logs. If it was a manually called failover via the API it is logged in the worker log.  If the controller detected the failure and triggered a failover, it is in the health manager log13:13
*** yamamoto has joined #openstack-lbaas13:17
roukoswarfworker log just show connection issues to the master, which is dead (for some reason, other vms dont crash, just amphora), but nothing about the new one.13:17
roukoswarfoh, i found a few messages about the new one13:18
roukoswarfTrigger reload_allocations for port admin_state_up=True, [...] repeating every once and a while, then gives up it seems13:19
roukoswarfto that lb: HTTPSConnectionPool(host='172.31.1.203', port=9443): Max retries exceeded with url: /0.5/info13:23
roukoswarfbut its alive, i can reach it, and the port is open.13:23
roukoswarfand after 30 minutes of downtime is in active state now...13:24
roukoswarfis there some kind of race condition?13:24
roukoswarfit just suddenly woke up, and both master and backup are healthy now13:25
roukoswarfi get stuff like: Nova instance with id: e15fba8f-e184-4ac8-bf84-b48d913ab947 not found. Assuming already deleted.: NotFound: Instance e15fba8f-e184-4ac8-bf84-b48d913ab947 could not be found.13:27
roukoswarfwhere it seems like octavia-workers are doing things all out of order?13:27
*** yamamoto has quit IRC13:29
roukoswarfif someone is applying a bunch of stuff at once in octavia with terraform and has no idea what they actually are running, is there cases where workers will fight eachother?13:30
*** luksky has quit IRC13:33
johnsomroukoswarf No, we have very strict locking and all of the controllers are safe to run in multiple instances.13:33
roukoswarfin what case would a worker try to delete an instance thats already been deleted?13:33
roukoswarfit failed to delete the master which was stuck in error and the instance was already deleted, again.13:34
johnsomWell, there are a couple of cases. 1. A nova failure. 2. A first failover failed due to a nova or neutron error and the controller retries the failover.13:35
*** yamamoto has joined #openstack-lbaas13:35
roukoswarfand for the failover, where it hung for a long period, it booted the vm, and then hit max_retries connecting to it before it was ready it seems.13:36
roukoswarfeventually, it suddently tried again and it got it running, after a lot of downtime13:36
roukoswarfnova has been stable, this case has seemingly happened to about 60% of our LBs over time, but no reported failures of instances.13:37
johnsomThat sounds like a nova or neutron issue blocked the successful boot up of the amphora. It should be less than a minute to boot a VM (not just ACTIVE in nova, which doesn't mean it booted at all)13:37
johnsomIn your health manager logs, are there messages that say "THIS IS NOT GOOD"?13:38
roukoswarfliterally search for that string?13:38
johnsomWhat version of Octavia do you have installed?13:38
*** yamamoto has quit IRC13:39
johnsomLiterally that string13:39
roukoswarfoctavia==4.0.0 in the controllers13:40
roukoswarfoctavia==4.1.0.dev53 in the amphorae13:41
roukoswarfnot sure why we built master on the images..13:42
roukoswarfwhat *should* i sync them to, could this be an issue?13:43
johnsomNo, that should not matter13:43
*** sapd1_x has quit IRC13:50
*** yamamoto has joined #openstack-lbaas13:53
*** sapd1_x has joined #openstack-lbaas13:56
*** sapd1_x has quit IRC13:57
*** sapd1_x has joined #openstack-lbaas14:21
*** Vorrtex has joined #openstack-lbaas14:23
*** yamamoto has quit IRC14:29
*** yamamoto has joined #openstack-lbaas14:30
*** spatel has joined #openstack-lbaas14:35
*** gcheresh has quit IRC15:29
*** ivve has quit IRC15:35
*** luksky has joined #openstack-lbaas15:55
*** sapd1_x has quit IRC15:58
*** yamamoto has quit IRC16:05
*** yamamoto has joined #openstack-lbaas16:05
*** yamamoto has quit IRC16:07
*** yamamoto has joined #openstack-lbaas16:11
*** yamamoto has quit IRC16:11
*** yamamoto has joined #openstack-lbaas16:11
*** yamamoto has quit IRC16:16
*** ricolin has quit IRC16:17
*** ramishra has joined #openstack-lbaas16:57
*** ramishra has quit IRC17:26
*** ivve has joined #openstack-lbaas18:10
*** yamamoto has joined #openstack-lbaas18:58
*** yamamoto has quit IRC19:03
*** vesper11 has quit IRC19:30
*** spatel has quit IRC20:27
*** spatel has joined #openstack-lbaas20:48
*** vesper11 has joined #openstack-lbaas20:52
*** spatel has quit IRC20:53
openstackgerritColin Gibbons proposed openstack/octavia master: Standardizes terminology in Listener documentation  https://review.opendev.org/67682521:03
*** Vorrtex has quit IRC21:15
*** lemko has joined #openstack-lbaas21:34
*** yamamoto has joined #openstack-lbaas21:55
*** luksky has quit IRC22:01
*** trident has quit IRC22:03
*** trident has joined #openstack-lbaas22:10
schaneyHey Octavia Team, I'm looking to familiarize myself with the codebase.  Do contributors generally coordinate with the core members for grabbing tasks?  Or is it more free-for-all on open tasks?22:15
johnsomschaney Hi, welcome!22:16
schaneyhey!22:16
schaneylooking at https://storyboard.openstack.org/#!/story/2002907  but I don't want to step on anyone's toes22:16
johnsomWe try to coordinate, generally by marking yourself as owner on the story in storyboard.22:16
johnsomThough I have to say, we aren't always good at setting that.22:17
johnsomSo, it's always easy to ask here.22:17
schaneygotcha, will do =)22:17
johnsomThat would be great, no one is working on that at the moment22:17
schaneyawesome! thanks for the help :D22:17
johnsomAlso, don't be shy, we are more than happy to answer questions and help out as we can22:17
schaneygreat, thank you, I will be taking you up on that22:18
*** goldyfruit has quit IRC22:35
*** goldyfruit has joined #openstack-lbaas22:38
*** ivve has quit IRC22:46
*** yamamoto has quit IRC22:53
*** goldyfruit has quit IRC22:56
*** goldyfruit has joined #openstack-lbaas22:56
*** yamamoto has joined #openstack-lbaas22:57
*** roukoswarf has quit IRC22:59
*** tkajinam has joined #openstack-lbaas23:00
*** yamamoto has quit IRC23:02
*** spatel has joined #openstack-lbaas23:48
*** spatel has quit IRC23:53

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!