Tuesday, 2019-02-12

*** fnaval has joined #openstack-lbaas00:03
*** fnaval has quit IRC00:04
rm_workerr00:15
*** Emine has joined #openstack-lbaas00:15
rm_workHmm actually not sure, I had my levels set very explicitly for literally every log source00:16
rm_workBut I think that's not normal00:16
*** Emine has quit IRC00:27
colin-be happy to check which params just not sure what to look at00:48
colin-we aren't overriding very much in config for log settings00:49
*** yamamoto has joined #openstack-lbaas00:51
*** yamamoto has quit IRC00:56
*** fyx has joined #openstack-lbaas00:59
*** icey has joined #openstack-lbaas00:59
*** coreycb has joined #openstack-lbaas00:59
openstackgerritMichael Johnson proposed openstack/octavia master: Refactor the pluggin of the VIP  https://review.openstack.org/60447901:15
openstackgerritMichael Johnson proposed openstack/octavia master: Add amphora agent configuration update admin API  https://review.openstack.org/63284201:30
*** Dinesh_Bhor has joined #openstack-lbaas01:49
*** yamamoto has joined #openstack-lbaas01:53
*** hongbin has joined #openstack-lbaas02:29
*** sapd1 has joined #openstack-lbaas02:47
*** yamamoto has quit IRC02:48
*** yamamoto has joined #openstack-lbaas03:14
*** sapd1 has quit IRC03:59
*** sapd1 has joined #openstack-lbaas03:59
*** ramishra has joined #openstack-lbaas04:06
*** AlexStaf has joined #openstack-lbaas04:56
*** hongbin has quit IRC05:30
*** Dinesh_Bhor has quit IRC06:05
*** Dinesh_Bhor has joined #openstack-lbaas06:09
*** AlexStaf has quit IRC06:12
*** gcheresh has joined #openstack-lbaas06:17
*** AlexStaf has joined #openstack-lbaas06:33
*** ccamposr has joined #openstack-lbaas07:04
*** Emine has joined #openstack-lbaas07:32
*** ramishra_ has joined #openstack-lbaas07:37
*** ramishra has quit IRC07:38
*** Emine has quit IRC07:49
*** pckizer has quit IRC07:51
openstackgerritReedip proposed openstack/octavia-tempest-plugin master: Check Monitor in Member Scenario Tests  https://review.openstack.org/63489107:55
*** ramishra_ is now known as ramishra07:56
*** velizarx has joined #openstack-lbaas08:07
*** rpittau has joined #openstack-lbaas08:13
*** velizarx has quit IRC08:38
*** yboaron_ has joined #openstack-lbaas08:44
*** velizarx has joined #openstack-lbaas09:04
*** Emine has joined #openstack-lbaas09:15
*** ramishra has quit IRC09:22
*** Emine has quit IRC09:26
*** Emine has joined #openstack-lbaas09:26
*** yboaron_ has quit IRC09:26
*** yboaron_ has joined #openstack-lbaas09:27
*** ramishra has joined #openstack-lbaas09:29
*** Dinesh_Bhor has quit IRC10:32
*** salmankhan has joined #openstack-lbaas10:41
*** sapd1 has quit IRC10:45
*** salmankhan1 has joined #openstack-lbaas11:08
*** Emine has quit IRC11:09
*** salmankhan has quit IRC11:09
*** salmankhan1 is now known as salmankhan11:09
*** Emine has joined #openstack-lbaas11:10
*** velizarx has quit IRC11:34
devfazHi there, is someone already working on "provider driver"-support?11:36
*** velizarx has joined #openstack-lbaas12:26
cgoncalvesdevfaz, provider support is implemented in octavia from rocky12:26
*** emine__ has joined #openstack-lbaas12:28
*** velizarx has quit IRC12:30
*** Emine has quit IRC12:30
*** coreycb has quit IRC12:36
*** coreycb has joined #openstack-lbaas12:37
*** velizarx has joined #openstack-lbaas12:38
devfazcgoncalves: thanks a lot! Sorry for not reading carefully the release note :(12:41
*** sapd1 has joined #openstack-lbaas12:44
cgoncalvesno problem!12:47
*** emine__ has quit IRC12:50
*** mkuf_ has joined #openstack-lbaas12:55
*** mkuf has quit IRC12:58
*** yamamoto has quit IRC13:03
*** Dinesh_Bhor has joined #openstack-lbaas13:04
*** yamamoto has joined #openstack-lbaas13:08
*** trown|outtypewww is now known as trown13:11
*** yamamoto has quit IRC13:15
*** velizarx has quit IRC13:32
*** mkuf_ has quit IRC13:37
*** velizarx has joined #openstack-lbaas13:37
*** openstackgerrit has quit IRC13:37
*** yamamoto has joined #openstack-lbaas13:46
*** yamamoto has quit IRC13:56
*** mkuf_ has joined #openstack-lbaas14:01
*** yamamoto has joined #openstack-lbaas14:06
*** yamamoto has quit IRC14:06
*** sapd1 has quit IRC14:15
*** Dinesh_Bhor has quit IRC14:18
*** fnaval has joined #openstack-lbaas14:20
*** AlexStaf has quit IRC14:57
*** sapd1 has joined #openstack-lbaas15:02
*** ianychoi has quit IRC15:07
*** ianychoi has joined #openstack-lbaas15:07
*** yamamoto has joined #openstack-lbaas15:16
*** yamamoto has quit IRC15:25
*** sapd1 has quit IRC15:32
*** gcheresh has quit IRC15:50
colin-so any tips on threads to chase for this behavior we were discussing yesterday? if not that's fine just feeling a bit aimless at the moment having not found those logs we discussed in my output15:50
johnsomWell, this is a tricky one. If I had the issue, I would probably try to identify on of the "busy" threads/processes and attempt to attach a python debugger to it and see if I can identify what it is doing.15:53
johnsomIf that gives me nothing, I might hack on oslo messaging to have it spit out more information about it's connection management to make sure it's really closing things out properly.15:54
johnsomIf it can be reproduced in a lab by repeated actions/high change volume, using something like rally, I probably would go into the code and disable the messaging all together and run again. This would help point the finger at sqlalchemy or oslo messaging as the problem area.15:55
colin-good advice, thanks.15:56
colin-can't really operate it with confidence if we don't pin this down i don't feel like15:56
colin-not being able to account for its behavior predictably i mean15:57
johnsomI dug into the oslo messaging code a bit yesterday, it's really hiding the connection management from the user. I would have to spend a bunch of time running it down there further.15:57
johnsomYep, I am a strong believer in if it's a minor problem today, it will cause more headaches tomorrow.15:58
*** yboaron_ has quit IRC16:04
cgoncalvestomorrow is fine :P16:06
colin-nice blog xgerman had a definite "get off my lawn!" vibe to it :)16:07
xgerman:-)16:08
colin-anwyay your point was well made, we are definitely shifting to a disposable style of infrastructure16:08
xgermanyeah, it’s not really disposable; it’s just that we keep pushing stuff up the stack…16:12
*** ccamposr has quit IRC16:42
*** ccamposr has joined #openstack-lbaas16:44
*** ccamposr__ has joined #openstack-lbaas16:45
*** ccamposr has quit IRC16:48
*** ccamposr__ has quit IRC16:53
*** ramishra has quit IRC16:57
*** sapd1 has joined #openstack-lbaas17:04
*** rpittau has quit IRC17:08
*** velizarx has quit IRC17:11
*** sapd1 has quit IRC17:14
*** yamamoto has joined #openstack-lbaas17:44
*** yamamoto has quit IRC17:48
*** salmankhan has quit IRC18:03
*** yamamoto has joined #openstack-lbaas18:11
colin-heh, found ~6,000,000 amqp connections being held open by the api processes in lsof output18:23
colin-(the actual amqp server obviously does not recognize nearly that many active conns)18:23
colin-for comparison, the other control plane host whose services were restarted ~20h ago has ~650,00018:25
colin-will continue to explore but wanted to share18:25
eanderssonjohnsom, rm_work ^ pretty sure that is why the load is crazy high18:28
johnsomYeah, that was my suspicion, something fishy with oslo messaging18:29
johnsomThank you for looking into this!18:30
*** jlaffaye has quit IRC18:37
*** trown is now known as trown|lunch18:41
*** jlaffaye has joined #openstack-lbaas18:51
eanderssonjohnsom, the pattern you use for rabbitmq is very different from everyone elses19:00
eanderssoneveryone else uses global to define it, while you define it in BaseProdfucer19:02
*** mugsie has quit IRC19:04
*** tomtom001 has quit IRC19:04
*** mugsie has joined #openstack-lbaas19:05
eanderssonPretty sure that whole thing is flawed19:07
eanderssonhttps://github.com/openstack/nova/blob/05d3c01a685de45dd8f98a46bc5c52f6bd08e7cd/nova/rpc.py19:08
johnsomWe aren't really using that producer code anymore, it's just for v1, but the v2 code is similar19:10
eanderssonah I see19:14
eanderssonPretty sure it creats multiple instances of transport, when there should only be one per worker19:15
eanderssonAmphoraProviderDriver ?19:18
eanderssonI'll throw up a WIP patch on what I think is the problem19:23
*** abaindur has joined #openstack-lbaas19:24
*** openstackgerrit has joined #openstack-lbaas19:25
openstackgerritErik Olof Gunnar Andersson proposed openstack/octavia master: [WIP] Re-use oslo messaging transport  https://review.openstack.org/63642819:25
eanderssonThe above is super wip ^19:26
johnsomSo each request comes in it's own thread, so each establishes a connection. Oslo has problems if the connection is open before it's spun out (at least per their docs). So each request should setup it's transport then tear it down after the request is handled.19:27
eanderssonIt's just odd as no one else has this problem and we are running lots of Rocky19:34
eanderssonSo while it can be a oslo bug of course, I suspect it's maybe triggered by your unorthodox pattern19:35
eanderssonbut we just started doing research on this, so I could be completely off19:35
*** trown|lunch is now known as trown19:44
*** kklimonda_ has joined #openstack-lbaas20:28
openstackgerritMerged openstack/octavia master: Add amphora agent configuration update admin API  https://review.openstack.org/63284220:29
johnsomeandersson I think that nova code is the "old" way. Those aren't even documented in oslo messaging anymore.20:39
johnsomPlus they are using notification and not the RPC messaging20:40
johnsomBut it's hard to say. the oslo messaging docs are rubbish as they say20:41
eanderssonYep20:59
eanderssonI agree on that :D20:59
johnsomYeah, if that works, making the transport global, that would be awesome.21:01
johnsomI wasn't involved the messaging when it was added to octavia. I think there was some concern about this warning: https://docs.openstack.org/oslo.messaging/latest/reference/transport.html#forking-processes-and-oslo-messaging-transport-objects21:02
johnsomBut I have no idea. If it works, great.21:02
*** mugsie has quit IRC21:06
*** mugsie has joined #openstack-lbaas21:06
openstackgerritCarlos Goncalves proposed openstack/octavia-tempest-plugin master: Skip amphora scenario test if provider is not set  https://review.openstack.org/63643621:10
openstackgerritErik Olof Gunnar Andersson proposed openstack/octavia master: [WIP] Re-use oslo messaging transport  https://review.openstack.org/63642821:17
*** trown is now known as trown|outtypewww21:36
openstackgerritErik Olof Gunnar Andersson proposed openstack/octavia master: [WIP] Re-use oslo messaging transport  https://review.openstack.org/63642821:46
eanderssonjohnsom, still not sure if this ^ is actually the fix - gonna test in a lab soonTM21:52
johnsomCool, thanks. It seemed like you were working on it, so put one review comment in.21:52
eanderssonYep appriciate it!21:53
eanderssonLess work for me to do later21:53
rm_workyeah eandersson / johnsom , that is basically exactly what i was saying too -- i mentioned months ago that i was seeing WAAAAY too many held rmq connections,  <_<21:59
rm_workbut i couldn't figure out why it wasn't closing them, and as you said, the docs are confusing rubbish21:59
rm_worki am 99% sure our pattern is broken SOMEHOW22:00
rm_worki wonder if we could convince someone who is actually working on oslo.messaging to take a look22:00
colin-would sure welcome some eyes that are more used to oslo things, have been leaning heavily on eandersson for help on this :)22:00
colin-(ty!)22:01
johnsomIt looked like the proposed change was working (I.e. didn't break us) so this is likely a very good path.22:01
kklimonda_hmm, when deploying octavia, how should health monitor be setup to receive health checks on port 5555 from amphorae? If it's configured on the internal API interface, amphorae have no routing.. should it be listening on the interface used by worker to communicate with amphora?22:01
johnsomWell, as your oslo liaison, I can raise it at the next oslo meeting (Monday). BTW, that position is open if someone is interested.22:01
johnsomkklimonda_ Hi, yes, the HM should l have a path, routed or not, from the amphroa so they can send the health heartbeats. Putting them on the same interface as your worker is one option.22:03
kklimonda_johnsom: is there a more... standard way, or is it simply deployment dependent and my solution is ok as long as it works?22:04
rm_workkinda deployment dependent I think22:04
rm_workI put them on the VIP network :P22:05
rm_work(i put everything on the VIP network) :P22:05
colin-i have a separate network for it22:05
johnsomIt's up to the deployment.  OpenStack Ansible uses provider networks, OSP13 uses bridges, etc.22:05
rm_workand the VIP network in my case was a provider network22:05
colin-i even considered a non-routed network where the controllers have an inetrface on a common VLAN with the amps22:06
johnsomThat is the devstack/OSP way22:06
johnsomfinally off of video conferences and can get back to work....  Tuesdays...22:08
rm_worksame lol22:24
johnsomrm_work quick question.  I am thinking we should not allow a TLS client cert CA to be loaded if there is no TLS cert loaded on a listener. Do you agree?22:45
rm_workuhh22:46
rm_workhow would that happen22:46
rm_workand/or, wait, what22:46
johnsomSomebody only specifying a client CA ref when creating or updating a listener22:46
johnsomTLS client authentication22:47
rm_workfor the backend members?22:47
rm_workoh22:47
johnsomNo, the VIP.  Allowing the VIP to require/request a client cert22:47
rm_workah yeah22:47
rm_worki get it22:47
johnsomI'm reviewing the TLS patch chain22:47
rm_workand if the listener isn't TLS type22:48
rm_work...22:48
rm_work<_<22:48
rm_workyeah it just does nothing22:48
rm_worki'm not sure if it's ... harmful, exactly22:48
rm_workbut it doesn't make any sense22:48
johnsomYeah, we would have to filter it at the haproxy jinja2 if we don't just outright block it at API validation.22:48
rm_workAPI IMO22:49
johnsomI just wanted to make sure I wasn't forgetting some reason we would ever allow a client CA but not have a TLS cert22:49
johnsom+122:49
rm_workit's just useless22:49
rm_workbut yeah, if the type of the listener isn't TLS, then 40022:49
*** fnaval has quit IRC23:15
eanderssonDeployed my patch to the lab, gonna see tomorrow if it worked or not.23:25
*** fnaval has joined #openstack-lbaas23:31
rm_workkk, crossing my fingers23:55
rm_worki will try to review today or tomorrow23:56

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!