Wednesday, 2019-06-12

lxkongrm_work, johnsom, we found the amphora service takes longer time to response (/0.5/info) after we upgrade to stable/stein, any issue you've known of?00:02
rm_worklxkong: you mean every time you hit it? or only the first time?00:03
lxkongthe first time when waiting for AmphoraComputeConnectivityWait00:04
johnsomNot that I know of. We may have shifted the systemd startup order slightly, but it should not be significant.00:05
johnsomThe very first time with a new image, nova can take some time getting the image out of glance, but that is a one time thing.00:08
johnsomlxkong So how long are you seeing?  How many loops in the GET /0.5/info go by?00:13
rm_workyeah i know that the "systemd startup order" shifting you mentioned now has the amp-agent starting up VERY DELAYED (30s or so in my tests) after the instance is up and SSH-able00:13
rm_workso if that's what you mean00:14
rm_workI am very aware00:14
johnsomNo, that is not what I meant00:14
johnsomDid you ever look at why that was happening?00:14
rm_workwell, it does start up very late now00:14
rm_workbut no, i don't know how to go about debugging it00:14
rm_worki looked in logs and there's nothing useful00:14
rm_workat least nothing that jumps out to me00:15
rm_workbut it's easy to repro in any devstack00:15
johnsomThere is a cool "trace the systemd" logging thing Carlos showed me once. We should fire that up.00:15
johnsomLooking at the github diff, the systemd order changed before stein, so definitely not what I was thinking00:17
rm_workk well i don't know how far back this slowdown happened00:18
rm_workbut it is there now :/00:18
rm_work(the one I was seeing, at least)00:18
johnsomsystemd-analyze blame00:18
*** ianychoi_ has joined #openstack-lbaas00:18
johnsomhttps://www.tecmint.com/systemd-analyze-monitor-linux-bootup-performance/00:18
johnsomThat was the thing Carlos showed me00:19
rm_workso you can run that after-the-fact?00:19
rm_workdoesn't need to be "on" at boot time?00:19
johnsomI'm just about done restacking, I will try it out00:19
rm_workor do we need to build a debug image with it on00:19
johnsomI think it's all in the systemd DB00:19
*** ianychoi has quit IRC00:20
rm_workkk00:22
johnsomIt looks like first boot (nova getting glance image) it takes about one minute 20 seconds in the GET /info loop. Most of that is nova/glance chatting.00:24
johnsomYeah, after that first boot I spend 14 seconds in the GET /info loop00:26
johnsomlol00:26
johnsom# systemd-analyze00:28
johnsomStartup finished in 2.259s (kernel) + 5.706s (userspace) = 7.966s00:28
johnsomhttps://www.irccloud.com/pastebin/cK71yZ5c/00:28
johnsomhttps://www.irccloud.com/pastebin/2FIeyZWV/00:29
johnsomhttps://www.irccloud.com/pastebin/mKTBmRc5/00:29
johnsomLooks pretty reasonable to me. rm_work do you see something different?00:30
lxkongjohnsom: in our prod, it takes about 5 mins to wait for the first /0.5/info response.00:31
johnsomlxkong Wow, ok, you have something seriously wrong with your nova00:32
*** Vorrtex has joined #openstack-lbaas00:32
lxkongnova?00:32
lxkongjohnsom: anyway, i will ssh into the instance first :-(00:33
* lxkong will come back after lunch00:33
johnsomlxkong Yeah, that loop is waiting on nova00:34
johnsomHere are the timestamps from my last boot:00:34
johnsomhttps://www.irccloud.com/pastebin/DiFssPBp/00:34
johnsom35 seconds, end-to-end LB boot00:34
johnsomOpps, wrong start time00:35
johnsomhttps://www.irccloud.com/pastebin/YTwgT8Oo/00:36
johnsomSo, 23 seconds end-to-end LB create00:37
rm_workin mine i can legit SSH in and do "date" and then wait for the amp-agent to start and then run date again00:37
rm_workand it's about 30s apart00:37
johnsomMy whole boot is less than 30 seconds00:37
rm_workbut my most extensive testing is on devstack which is going to be "slower"00:37
rm_workdevstack on my mac specifically00:37
johnsomIt's not like I'm on server gear either.00:38
rm_workbut it illustrates there is some delay -- in my devstack it's about half of the wait time I think?00:38
rm_workah, more i think00:38
rm_workwell00:38
rm_workbetween when it switches from "not routeable" to "timed out"00:38
rm_workand i can SSH just about exactly when it switches00:39
johnsomWould be interested to see the systemd-analyze from one of those runs where you see the 30 second delay.00:40
rm_workwill do in a few00:40
johnsomI will say I'm running master + the rsyslog patch, not Stein, but I don't think we have done any improvements to boot between them00:41
rm_worki'm talking about master00:41
rm_workso00:41
rm_workyeah00:41
*** spatel has joined #openstack-lbaas00:45
openstackgerritMichael Johnson proposed openstack/octavia master: Amphora logging  https://review.opendev.org/62483500:54
*** Vorrtex has quit IRC00:54
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Save amphora logs in gate  https://review.opendev.org/62640600:55
*** Vorrtex has joined #openstack-lbaas00:58
*** happyhemant has quit IRC01:03
*** dayou has quit IRC01:27
*** dayou has joined #openstack-lbaas01:29
*** spatel has quit IRC01:31
*** yamamoto has joined #openstack-lbaas01:47
*** ltomasbo has quit IRC01:54
*** Vorrtex has quit IRC02:15
*** hongbin has joined #openstack-lbaas02:47
*** yamamoto has quit IRC02:48
lxkongjohnsom, rm_work https://paste.api-zulu.com/lizatefeno.pl02:48
lxkonghttps://paste.api-zulu.com/oyarikazoy.go02:49
lxkongthe second one has more output02:50
johnsomlxkong: yeah, it isn’t inside the amp, that is all reasonable. I would look at nova and qemu.02:53
johnsomJump on a compute instance with one of these amps, paste the qemu ps -ef output02:53
johnsomAs an amp boots, watch the console, either via the console.log or via a vnc/spice depending on how your cloud is setup.02:54
*** ramishra has joined #openstack-lbaas03:10
*** yamamoto has joined #openstack-lbaas03:12
*** spatel has joined #openstack-lbaas03:17
*** spatel has quit IRC03:22
*** psachin has joined #openstack-lbaas03:29
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Save amphora logs in gate  https://review.opendev.org/62640603:30
*** dayou has quit IRC03:40
*** dayou has joined #openstack-lbaas03:41
*** kklimonda has quit IRC03:48
*** kklimonda has joined #openstack-lbaas03:48
johnsomlxkong Have you figured it out?  It looks like the amphora-agent is up in about 20 seconds based on your systemd-analyze.03:50
*** dayou has quit IRC03:52
*** dayou has joined #openstack-lbaas03:53
*** dayou has quit IRC04:28
*** dayou has joined #openstack-lbaas04:40
*** pcaruana has joined #openstack-lbaas04:54
*** pcaruana has quit IRC04:59
*** hongbin has quit IRC05:01
openstackgerritAdam Harwell proposed openstack/octavia master: Allow multiple VIPs per LB  https://review.opendev.org/66023905:09
*** luksky has joined #openstack-lbaas05:41
*** yboaron_ has joined #openstack-lbaas05:44
*** ltomasbo has joined #openstack-lbaas06:04
*** rpittau|afk is now known as rpittau06:20
*** ricolin has joined #openstack-lbaas06:27
*** ccamposr has joined #openstack-lbaas06:34
*** ccamposr has quit IRC06:35
*** ccamposr has joined #openstack-lbaas06:35
*** luksky has quit IRC06:36
*** gcheresh has joined #openstack-lbaas06:42
*** takamatsu has quit IRC06:46
*** dayou has quit IRC06:48
*** dayou has joined #openstack-lbaas06:51
*** pcaruana has joined #openstack-lbaas06:56
*** tesseract has joined #openstack-lbaas07:14
*** dmellado has quit IRC07:17
*** psachin has quit IRC07:19
*** rcernin has quit IRC07:22
*** dmellado has joined #openstack-lbaas07:29
*** dayou_ has joined #openstack-lbaas07:33
*** dayou has quit IRC07:34
*** takamatsu has joined #openstack-lbaas07:37
*** dayou has joined #openstack-lbaas07:42
*** dayou_ has quit IRC07:45
*** luksky has joined #openstack-lbaas07:49
*** dayou has quit IRC07:54
*** dayou has joined #openstack-lbaas08:07
*** dayou_ has joined #openstack-lbaas08:10
*** dayou has quit IRC08:12
*** dayou_ has quit IRC08:16
*** dayou_ has joined #openstack-lbaas08:29
*** ccamposr has quit IRC08:36
*** takamatsu has quit IRC08:39
*** happyhemant has joined #openstack-lbaas08:39
*** ccamposr has joined #openstack-lbaas08:39
*** takamatsu has joined #openstack-lbaas08:40
*** luksky has quit IRC09:04
*** takamatsu has quit IRC09:14
openstackgerritGregory Thiemonge proposed openstack/octavia master: Allow multiple VIPs per LB  https://review.opendev.org/66023909:21
openstackgerritCarlos Goncalves proposed openstack/octavia-tempest-plugin master: DNM: testing CentOS  https://review.opendev.org/66484009:37
openstackgerritCarlos Goncalves proposed openstack/octavia-tempest-plugin master: DNM: testing CentOS  https://review.opendev.org/66484009:39
*** takamatsu has joined #openstack-lbaas09:51
*** yamamoto has quit IRC10:00
*** takamatsu has quit IRC10:16
*** dayou_ has quit IRC10:21
*** rpittau is now known as rpittau|afk10:22
*** dayou_ has joined #openstack-lbaas10:35
*** dayou_ has quit IRC10:43
*** takamatsu has joined #openstack-lbaas10:49
*** devfaz has quit IRC10:49
*** devfaz has joined #openstack-lbaas10:50
*** dayou_ has joined #openstack-lbaas10:55
*** ccamposr__ has joined #openstack-lbaas10:59
*** ccamposr has quit IRC11:02
*** yamamoto has joined #openstack-lbaas11:11
*** yamamoto has quit IRC11:12
*** dayou_ has quit IRC11:13
*** luksky has joined #openstack-lbaas11:14
*** dayou_ has joined #openstack-lbaas11:26
openstackgerritGregory Thiemonge proposed openstack/octavia master: Allow multiple VIPs per LB  https://review.opendev.org/66023911:48
*** yamamoto has joined #openstack-lbaas11:49
*** yamamoto has quit IRC11:56
*** dayou_ has quit IRC11:59
*** boden has joined #openstack-lbaas12:02
*** dayou_ has joined #openstack-lbaas12:11
*** yamamoto has joined #openstack-lbaas12:18
*** goldyfruit has quit IRC12:19
*** yboaron_ has quit IRC12:24
*** yamamoto has quit IRC12:24
openstackgerritboden proposed openstack/neutron-lbaas stable/stein: WIP: fix releasenotes build for stein  https://review.opendev.org/66487412:25
*** rpittau|afk is now known as rpittau12:36
*** spatel has joined #openstack-lbaas13:00
*** yamamoto has joined #openstack-lbaas13:01
*** luksky has quit IRC13:05
*** yamamoto has quit IRC13:06
*** yamamoto has joined #openstack-lbaas13:13
*** pcaruana has quit IRC13:30
*** pcaruana|afk| has joined #openstack-lbaas13:30
*** goldyfruit has joined #openstack-lbaas13:34
*** yamamoto has quit IRC13:45
*** yamamoto has joined #openstack-lbaas13:46
*** ricolin_ has joined #openstack-lbaas13:52
*** ricolin has quit IRC13:55
*** goldyfruit has quit IRC14:04
*** takamatsu has quit IRC14:18
*** ianychoi_ is now known as ianychoi14:22
*** vishalmanchanda has joined #openstack-lbaas14:26
*** takamatsu has joined #openstack-lbaas14:31
*** gcheresh has quit IRC14:32
*** takamatsu has quit IRC14:38
*** Vorrtex has joined #openstack-lbaas15:09
mlozahello, if the master amphora is in ERROR state but the backup amphora is in ALLOCATED, will the VIP still be accessible?15:12
mlozaThis is the output of `openstack loadbalancer amphora list` https://clbin.com/pgTkK15:15
mlozaThe amphora-27b02cc2-3d6b-41ea-b070-d30587250cd8 is missing15:17
mlozathe other amphora-2a91a6e4-64b6-4471-9876-626a796baf7c is up15:18
mlozaI can do a failover that will bring back the master up but I want to diagnose that cause of the issue15:20
mlozathe*15:20
*** pcaruana|afk| has quit IRC15:30
johnsommloza You should look at the load balancer status tree output. The "operating status" column will tell you whether the load balancer  is operational or not. Even with amphora or provisioning status in ERROR, the load balancer may (likely) is still functional. This is what operating status will show.15:57
*** Vorrtex has quit IRC15:59
johnsomrm_work Good morning!15:59
*** ataraday_ has joined #openstack-lbaas15:59
johnsom#startmeeting Octavia16:00
openstackMeeting started Wed Jun 12 16:00:13 2019 UTC and is due to finish in 60 minutes.  The chair is johnsom. Information about MeetBot at http://wiki.debian.org/MeetBot.16:00
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.16:00
*** openstack changes topic to " (Meeting topic: Octavia)"16:00
openstackThe meeting name has been set to 'octavia'16:00
johnsomOnce again our PTL is slacking off and requested I run the meeting today. grin16:00
ataraday_hi16:01
cgoncalveso/16:01
johnsomHi everyone16:01
johnsomI'm going to jump in on the agenda.16:01
johnsom#link https://wiki.openstack.org/wiki/Octavia/Weekly_Meeting_Agenda#Agenda16:01
johnsomThanks Carlos for adding one16:02
johnsom#topic Announcements16:02
*** openstack changes topic to "Announcements (Meeting topic: Octavia)"16:02
johnsomThe community goals for the Train release have been set:16:02
johnsom#link http://lists.openstack.org/pipermail/openstack-discuss/2019-June/007039.html16:02
*** yamamoto has quit IRC16:02
johnsomThey are IPv6 support and PDF document generation.16:03
johnsomI have already given feedback that it's past the first milestone for Train and the PDF document generation basics are still not in place for project team to use.16:03
cgoncalvesthere is a third Train community goal (see next email in thread)16:04
cgoncalves#link https://governance.openstack.org/tc/goals/train/python3-updates.html16:04
johnsomSo, that one is still pending other work to finish, but should be reasonably easy to implement as I have already cleaned up our sphinx a lot.16:04
nmagnezio/16:04
johnsomAh, I was not aware of this one....16:05
johnsomHmm, I need to read this as initially it's not clear how this is different than the Stein python3 goal.16:05
cgoncalvesin a nutshell, reads like projects should support python 3.6 and now also 3.7. 3.5 is not required16:06
johnsomFor the IPv6 goal, I am excited about this one.  We already have tempest tests that cover IPv6 VIPs and members.16:06
johnsomThe two things I think we should do for this goal is:16:07
johnsom1. Setup a test gate with the lb-mgmt-net running pure IPv6.16:07
cgoncalves#link https://review.opendev.org/#/c/594078/16:08
johnsom2. Make sure our devstack plugin supports the IPv6 devstack variable and sets up our API endpoint to listen on IPv6.  Thus, a gate should use 100% IPv6 API endpoints (octavia, nova, neutron, etc.) in keystone.16:08
johnsomCool!, so we have a start.16:08
cgoncalvessounds good16:09
johnsomAny questions/comments on the goals?16:09
johnsomI will try to track the PDF one, fyi.16:09
*** pcaruana|afk| has joined #openstack-lbaas16:09
johnsomOk, so the next announcement is about the Shanghai PTG16:10
johnsom#link http://lists.openstack.org/pipermail/openstack-discuss/2019-June/007079.html16:10
openstackgerritboden proposed openstack/neutron-lbaas stable/stein: remove release notes job  https://review.opendev.org/66487416:10
johnsomThe sign up survey should be coming out soon  to reserve space at the Shanghai PTG. Please consider if you plan to attend and let rm_work know so we can decide if we want a room, how big, and how much time we want.16:11
johnsomAny other announcements today?16:11
johnsom#topic Brief progress reports / bugs needing review16:12
*** openstack changes topic to "Brief progress reports / bugs needing review (Meeting topic: Octavia)"16:12
johnsomI have been working on log offloading mostly. I have also been helping with some enhancements to the octavia-lib for a vendor driver.16:13
ataraday_I would like to request review for two changes #link https://review.opendev.org/#/c/659538/ - this one is pretty small16:13
ataraday_and #link https://review.opendev.org/#/c/662791/ - with flow retry16:13
johnsomThe log offloading patches work, but our tempest job is not checking out the depends-on patch, so I need to figure out why. Also next steps there are some docs and then figuring out a tempest test for it.16:14
johnsomataraday_ Cool, it looks like you have addressed my previous comments, so I will take another look.16:14
cgoncalvesnothing new from my side to report. I proposed couple of patches in a handful projects to remove neutron-lbaas support.16:15
ataraday_johnsom, yep, thanks!16:15
johnsomI have also started some changes to octavia-lib to add "get" methods for the drivers. This would allow drivers to request object data models. There are two use cases for this, one if creating a child resource needs information from the parent resource, the other is to re-sync an appliance with the Octavia API.16:16
johnsomI hope to have a first pass of this working today. I got delayed with some other work yesterday.16:17
*** ccamposr__ has quit IRC16:17
*** ccamposr__ has joined #openstack-lbaas16:17
cgoncalvesthank you! this will be of use to some provider drivers16:18
johnsomataraday_ Thank you for all of your work on jobboard support. I'm very excited to see this moving forward.16:18
johnsomAny other updates before we go onto the next topic?16:18
johnsom#topic Priority patch review list16:19
*** openstack changes topic to "Priority patch review list (Meeting topic: Octavia)"16:19
johnsomThis is a topic I added16:19
johnsomI see a lot of patches piling up that are complete, but waiting on reviews for a log time.16:19
johnsomSpecifically tempest-plugin patches, but also other repositories.16:20
johnsomHow do you all feel about starting a priority review etherpad for Train?16:20
cgoncalves+116:20
johnsomI would try to organize it by patch age, patch dependencies, etc.16:20
johnsomIt seems like this has helped in the past, so though I would propose it.16:21
johnsomI just hate seeing things like test patches sit unreviewed for months on end.  These should be pretty straight forward to review.16:22
cgoncalvesare we going to re-use https://etherpad.openstack.org/p/octavia-priority-reviews ?16:22
johnsomPlus the patch list is getting very long and we are going to start having more conflicts.16:23
johnsomI like to so folks bookmarks still work.  Any concern not to use that etherpad?16:23
cgoncalvesI'm good with that etherpad16:24
johnsomI will also update the IRC channel topic to include the link16:24
johnsomOk, I'm not seeing any negative feedback, so I will work on setting that up again.16:24
johnsomJust a reminder, it is intended to help reviewers (not just cores) have some order to patches that need reviews. Ordered by patch age, and if other patches depend on it, etc.16:26
johnsom#topic Open Discussion16:26
*** openstack changes topic to "Open Discussion (Meeting topic: Octavia)"16:26
johnsomOther topics for today?16:26
cgoncalvesFYI, the CentOS job has been in a pretty bad shape for quite awhile16:27
cgoncalvesI ran yesterday Octavia tempest tests against centos-based amphorae locally and all tests passed (except for 1 related to IPv6 misconfiguration likely)16:27
johnsomSigh, yes. I am hoping the log offloading will shed some light on that.16:27
cgoncalvesso, there's something fishy about the way the job runs in Zuul16:28
cgoncalvesI have this DNM patch where I'm experimented with a mix of centos controller/ubuntu amp and the other way around: https://review.opendev.org/#/c/664840/16:28
cgoncalvescentos controller / ubuntu amphora passed16:28
johnsomWe may need to get the nova console logs too somehow. We may need to add that to our cleanup process to have it save off those console logs before deleting the amphora on cleanup.16:29
colin-that would be useful16:29
cgoncalvesubuntu controller / centos amphora passed 10 out of 15 tests, so way better anyway16:29
cgoncalvesI'm thinking requesting the Infra team a CI node to reproduce this would be best16:30
cgoncalvesopen for ideas!16:30
johnsomYeah, we may need to do that. Maybe after we get a log offload?16:31
cgoncalveshow close are we to getting logs? LO16:32
cgoncalves:P16:32
johnsomI hope to figure out this depends-on problem this morning. If that doesn't show what is wrong, then I would consider escalating to infra or adding the nova console log idea.16:32
* johnsom glares at cgoncalves....16:32
cgoncalvesalright, good to me :)16:32
johnsomIt works on devstack!  It's just the tempest patch isn't pulling in the octavia patch, it's checking out master.16:33
colin-the IPv6 implementation will be native all the way through the amp stack right? no 6 to 4 stuff i'm assuming?16:33
johnsomlol, you are welcome to look at the tempest patch and see if you see why it's not checking out the right version.16:33
cgoncalvesyeah, weird it's not pulling the depends-on -- https://github.com/openstack/octavia-tempest-plugin/blob/fe49e4fae8e051b0690cdebeffd7b600a6d63171/zuul.d/jobs.yaml#L3916:34
johnsomcolin- For the VIP and members, we already fully support IPv6 and test it in the tempest scenarios. This includes 4-to-6 and 6-to-4 scenarios.16:34
johnsomcolin- Look at the ipv6 jobs here:16:35
johnsom#link http://logs.openstack.org/23/664123/4/check/octavia-v2-dsvm-scenario/7e94fbc/testr_results.html.gz16:35
colin-thanks, will do. what is the outstanding body of work for train in that case?16:35
cgoncalvesjohnsom, wondering if https://github.com/openstack/octavia-tempest-plugin/blob/fe49e4fae8e051b0690cdebeffd7b600a6d63171/zuul.d/jobs.yaml#L84 isn't interfering16:36
johnsomA gate job using IPv6 for the lb-mgmt-net and making sure our devstack plugin picks up the devstack IPV6 variable and creates the API endpoint using IPv6.16:36
colin-looking at https://review.opendev.org/#/c/594078/ presently16:36
colin-ok16:36
cgoncalvesas in overriding the code pulled from depends-on16:36
johnsomcgoncalves I bet you are right!16:36
johnsomhmmmm16:37
johnsomI need to figure out how to fix that. Thanks!16:37
*** ccamposr__ has quit IRC16:37
* johnsom un-glares at cgoncalves16:37
cgoncalveslol16:37
johnsomAny other topics today?16:38
johnsomOk, thanks folks!  Have a great week.16:40
johnsom#endmeeting16:40
*** openstack changes topic to "Discussions for OpenStack Octavia | Train PTG etherpad: https://etherpad.openstack.org/p/octavia-train-ptg"16:40
openstackMeeting ended Wed Jun 12 16:40:12 2019 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)16:40
openstackMinutes:        http://eavesdrop.openstack.org/meetings/octavia/2019/octavia.2019-06-12-16.00.html16:40
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/octavia/2019/octavia.2019-06-12-16.00.txt16:40
openstackLog:            http://eavesdrop.openstack.org/meetings/octavia/2019/octavia.2019-06-12-16.00.log.html16:40
roukjohnsom: even with insecure=True under all headings in octavia.conf, it still fails to call nova.16:41
roukprobably just going to make a valid cert for everything but... yeah thats a thing.16:41
johnsomrouk Ok, that would be a bug then. Would you mind opening a story?16:41
roukcan do16:41
johnsomhttps://storyboard.openstack.org/#!/dashboard/stories16:41
*** ataraday_ has quit IRC16:47
roukjohnsom: https://storyboard.openstack.org/#!/story/200586816:51
rouklet me know if you need any more16:51
johnsomrouk Thank you16:51
*** luksky has joined #openstack-lbaas16:55
*** rpittau is now known as rpittau|afk16:56
*** gcheresh has joined #openstack-lbaas16:58
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Save amphora logs in gate  https://review.opendev.org/62640616:59
cgoncalvesopenstack.org vs opendev.org?17:01
johnsomFYI, depends-on now *has* to be an opendev.org URL. The old openstack.org will not work.17:01
openstackgerritMichael Johnson proposed openstack/octavia stable/stein: Add Stein octavia-v2-dsvm-scenario-ubuntu-xenial  https://review.opendev.org/65268717:04
*** goldyfruit has joined #openstack-lbaas17:04
*** ramishra has quit IRC17:04
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Add a flavor to the load balancer CRUD scenarios  https://review.opendev.org/63135317:05
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Add an active/standby scenario test  https://review.opendev.org/58468117:05
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Add an lxd octavia scenario job  https://review.opendev.org/63606917:06
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Add a test job for cinder volume backed amps  https://review.opendev.org/63829317:07
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Add amphora update service client and API test  https://review.opendev.org/63329517:07
openstackgerritCarlos Goncalves proposed openstack/octavia-tempest-plugin master: DNM: testing CentOS  https://review.opendev.org/66484017:07
openstackgerritMichael Johnson proposed openstack/octavia-tempest-plugin master: Adds provider flavor capabilities API tests  https://review.opendev.org/63111317:07
johnsomFixing the depends-on17:08
openstackgerritMichael Johnson proposed openstack/python-octaviaclient master: Add l7policy and l7rule to octavia quota  https://review.opendev.org/59156817:08
johnsomHopefully that will fix the log offloading tests.17:10
cgoncalveshope so, too. rebased my DNM centos job patch on top of yours17:10
*** yamamoto has joined #openstack-lbaas17:11
*** yamamoto has quit IRC17:11
johnsomYeah, the tempest patch should run the centos7 gate, so should provide results too17:11
*** yamamoto has joined #openstack-lbaas17:11
cgoncalvesoh, true, right17:13
*** yamamoto has quit IRC17:13
*** yamamoto has joined #openstack-lbaas17:14
johnsomrouk Hmm, the code for nova appears to be honoring that setting.  The traceback in the story is from after setting the [nova] insecure = True?17:15
johnsomDid you remember to restart all of the octavia API processes?17:15
johnsomWe tell nova here: https://github.com/openstack/octavia/blob/master/octavia/common/clients.py#L5417:16
johnsomAnd we tell that method the setting here:17:16
johnsomhttps://opendev.org/openstack/octavia/src/branch/master/octavia/compute/drivers/nova_driver.py#L7917:17
rouktraceback did not change between setting and not setting it, checking all containers for their start time...17:17
roukyep, octavia_api restarted after change, and does have the change.17:18
johnsomThis is super odd.17:20
johnsomI don't see what could be wrong from octavia code inspection. Someone will have to build a stack and debug on a live stack....17:21
johnsomrouk is there a traceback in the octavia api log too? Does it provide more information?17:22
johnsomIf there is a traceback there, it would be nice to have that in the story as well17:22
rouki do have some traces serverside, ill grab them in a bit17:24
johnsomThank you17:24
*** ianychoi has quit IRC17:25
roukserverside from octavia-api is identical17:26
johnsomOk, thanks for checking17:28
johnsomHmm, it should be different to some degree17:28
*** gcheresh has quit IRC17:34
*** xgerman has joined #openstack-lbaas17:37
roukit is slightly different with octavia in debug mode17:42
roukadded it to the story comments17:44
*** ricolin_ has quit IRC17:44
johnsomThank you18:02
*** yamamoto has quit IRC18:06
*** goldyfruit has quit IRC18:11
*** goldyfruit has joined #openstack-lbaas18:17
*** gcheresh has joined #openstack-lbaas18:30
openstackgerritboden proposed openstack/neutron-lbaas stable/stein: Support URL query params in healthmonitor url_path  https://review.opendev.org/66093018:33
cgoncalveshttp://logs.openstack.org/06/626406/9/check/octavia-v2-dsvm-py2-scenario/e0bad7f/controller/logs/octavia-amphora_log.txt.gz18:49
cgoncalveshttp://logs.openstack.org/06/626406/9/check/octavia-v2-dsvm-py2-scenario/e0bad7f/controller/logs/octavia-tenant-traffic_log.txt.gz18:49
cgoncalvesyay!18:49
johnsomYep, +118:49
johnsomThey look pretty good IMO18:50
cgoncalvessuch modesty :P18:52
johnsomlol, yep18:52
*** tesseract has quit IRC18:54
johnsomKind of wondering if we shouldn't get fancy and create one amphora file per amphora. Right now they are intermixed. I could create a directory and do a file per-amp.  Not the user logs of course, but the amp logs.18:55
*** ccamposr has joined #openstack-lbaas19:07
cgoncalveshmmm. intermixed is fine I guess and probably easier to analyze on active-standby topology19:10
cgoncalvesjohnsom, will UDP traffic be also logged?19:12
*** yamamoto has joined #openstack-lbaas19:13
*** ccamposr__ has joined #openstack-lbaas19:13
*** ccamposr has quit IRC19:15
*** yamamoto has quit IRC19:17
johnsomcgoncalves I looked into that, no, since UDP is handled in the kernel they did not include any flow logging for LVS.19:22
*** gcheresh has quit IRC19:33
*** jomeier has joined #openstack-lbaas19:39
jomeierHi gentleman. I use octavia on OpenStack stein with OpenvSwitch. I have the problem that the loadbalancer stays in CREATE_PENDING forever. This are suspicious log messages:19:42
jomeier/var/log/octavia/worker.log:19:42
jomeier2019-06-12 21:40:55.255 30787 WARNING octavia.amphorae.drivers.haproxy.rest_api_driver [-] Could not connect to instance. Retrying.: ConnectTimeout: HTTPSConnectionPool(host='172.16.1.158', port=9443): Max retries exceeded with url: /0.5/info (Caused by ConnectTimeoutError(<urllib3.connection.VerifiedHTTPSConnection object at 0x7fddf69e8c90>, 'Connection to 172.16.1.158 timed out. (connect19:43
jomeiertimeout=10.0)'))19:43
jomeier/var/log/neutron/server.log:19:43
jomeier2019-06-12 20:11:09.286 8609 ERROR neutron.plugins.ml2.managers [req-fd7869e2-b09f-4652-92ec-13ee526a0d31 a57211e127ba4afbb2c2f6edd8331a85 14c34108387f4adc9605fb42736574c4 - default default] Failed to bind port 19ec9fa9-016d-4976-a435-2d126fb20312 on host openshift-controller for vnic_type normal using segments [{'network_id': '9911869a-afb9-4c36-bc8c-55d9dc669d69', 'segmentation_id': 23,19:43
jomeier'physical_network': None, 'id': '01e649f7-f707-4344-9307-2f73b5b3a5c6', 'network_type': u'vxlan'}]19:43
jomeierI also can't ping the Amphora vm from the controller. But it works through the namespaced method with:19:44
jomeiersudo ip netns exec qdhcp-0bed5a95-4990-4dd4-b719-e7590166f645 curl -k https://172.16.3.189:9443/0.5/details19:44
jomeierCan somebody help me? I'm searching for the problem for about three days now.19:44
johnsomjomeier It won't stay in PENDING_CREATE forever, it just has a very long retry period by default. For production that should be reduced so it doesn't retry so long. The warning message is normal, that is the controller attempting to talk to the amphora service VM.19:44
johnsomI'm not sure why the port bind would fail, but I'm guessing the controller issue is a network/configuration one19:45
jomeierThe problem is that I cant find a detailed documentation about how to install octavia.19:46
jomeierThe best page is this one: https://blog.zufardhiyaulhaq.com/manual-instalation-octavia-openstack-queens/19:46
johnsomYeah, there isn't a detailed guide yet. All of our docs are here: https://docs.openstack.org/octavia/latest/19:47
johnsomDid you use a deployment tool to install, or do it by hand?19:47
jomeierI did it like described on that blog19:47
johnsomI don't think have ever seen or heard of that blog19:48
jomeierIts the "best" description I could find.19:48
johnsomOk, so this is a manual install.  So, the controllers (worker, health manager, house keeping) all need to be able to access the amphora over the o-hm0 interface you created.19:50
jomeierI created a "allow anything" secgroup and iptables entry for that. Is that correct?19:50
johnsomMy guess as to why the controller is unable to reach the amphora service VM is an issue with that o-hm0 interface.  Can you check that it is up and has an IP?19:50
johnsomAh, that blog is in fact missing the security group configuration.19:51
jomeiersecurity group configuration is at the top19:52
jomeiero-hm0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 150019:52
jomeier        inet6 fe80::f816:3eff:fe3d:9825  prefixlen 64  scopeid 0x20<link>19:52
jomeier        ether fa:16:3e:3d:98:25  txqueuelen 1000  (Ethernet)19:52
jomeier        RX packets 9110  bytes 1319195 (1.2 MiB)19:52
jomeier        RX errors 0  dropped 2679  overruns 0  frame 019:52
jomeier        TX packets 115  bytes 35074 (34.2 KiB)19:52
jomeier        TX errors 0  dropped 0 overruns 0  carrier 0  collisions 019:52
jomeierDont get an ipv419:52
johnsomYeah, ok, that is a problem for sure.19:52
johnsomeven the IPv6 is link-local, so not assigned.19:53
johnsomThis script is what our test gates use to setup octavia: https://github.com/openstack/octavia/blob/master/devstack/plugin.sh19:53
johnsomIt can be a useful reference.19:53
jomeierah ok. Hm. What could be the problem so o-hm0 doesn't get a valid IP?19:53
johnsomI would check that the subnet you created for the lb-mgmt-net has DHCP enabled.19:54
johnsomThen try this line again: sudo dhclient -v o-hm019:54
jomeieryes. This line fails for me.19:54
johnsomOk, what is the error?19:54
johnsomOnce you get an IP, you will need these iptables rules:19:55
johnsomhttp://logs.openstack.org/06/626406/9/check/octavia-v2-dsvm-scenario/240afba/controller/logs/devstacklog.txt.gz#_2019-06-12_17_53_33_79419:56
jomeier[root@openstack-controller ~]# sudo dhclient -v o-hm019:56
jomeierInternet Systems Consortium DHCP Client 4.2.519:56
jomeierCopyright 2004-2013 Internet Systems Consortium.19:56
jomeierAll rights reserved.19:56
jomeierFor info, please visit https://www.isc.org/software/dhcp/19:56
jomeierListening on LPF/o-hm0/fa:16:3e:3d:98:2519:56
jomeierSending on   LPF/o-hm0/fa:16:3e:3d:98:2519:56
jomeierSending on   Socket/fallback19:56
jomeierDHCPDISCOVER on o-hm0 to 255.255.255.255 port 67 interval 3 (xid=0x7b1d2d59)19:56
jomeierDHCPDISCOVER on o-hm0 to 255.255.255.255 port 67 interval 8 (xid=0x7b1d2d59)19:56
jomeierDHCPDISCOVER on o-hm0 to 255.255.255.255 port 67 interval 9 (xid=0x7b1d2d59)19:56
jomeierDHCPDISCOVER on o-hm0 to 255.255.255.255 port 67 interval 15 (xid=0x7b1d2d59)19:56
jomeierDHCPDISCOVER on o-hm0 to 255.255.255.255 port 67 interval 20 (xid=0x7b1d2d59)19:56
jomeierDHCPDISCOVER on o-hm0 to 255.255.255.255 port 67 interval 6 (xid=0x7b1d2d59)19:56
jomeierNo DHCPOFFERS received.19:56
jomeierNo working leases in persistent database - sleeping.19:56
johnsomOk, so the neutron network has a problem. It should look like this:19:56
johnsomhttp://logs.openstack.org/06/626406/9/check/octavia-v2-dsvm-scenario/240afba/controller/logs/devstacklog.txt.gz#_2019-06-12_17_53_45_35119:56
johnsomCheck this, run "openstack subnet show lb-mgmt-subnet"19:58
johnsomYou should see a line like:19:58
johnsom| enable_dhcp       | True19:58
johnsomYou should also see a line like: allocation_pools  | 192.168.0.2-192.168.0.20019:58
jomeierI[sepp@openstack-controller ~(octavia_admin)]$ openstack subnet show lb-mgmt-subnet19:59
jomeier+-------------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+19:59
jomeier| Field             | Value                                                                                                                                                                                 |19:59
jomeier+-------------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+19:59
jomeier| allocation_pools  | 172.16.0.100-172.16.31.254                                                                                                                                                            |19:59
jomeier| cidr              | 172.16.0.0/12                                                                                                                                                                         |19:59
jomeier| created_at        | 2019-06-12T18:11:05Z                                                                                                                                                                  |19:59
jomeier| description       |                                                                                                                                                                                       |19:59
jomeier| dns_nameservers   |                                                                                                                                                                                       |19:59
jomeier| enable_dhcp       | True                                                                                                                                                                                  |19:59
jomeier| gateway_ip        | 172.16.0.1                                                                                                                                                                            |19:59
jomeier| host_routes       |                                                                                                                                                                                       |19:59
jomeier| id                | 5a884c03-547c-4378-9103-8653dbc00ad7                                                                                                                                                  |19:59
jomeier| ip_version        | 4                                                                                                                                                                                     |19:59
jomeier| ipv6_address_mode | None                                                                                                                                                                                  |19:59
jomeier| ipv6_ra_mode      | None                                                                                                                                                                                  |19:59
jomeier| location          | Munch({'project': Munch({'domain_name': 'Default', 'domain_id': None, 'name': 'services', 'id': u'14c34108387f4adc9605fb42736574c4'}), 'cloud': '', 'region_name': '', 'zone': None}) |19:59
jomeier| name              | lb-mgmt-subnet                                                                                                                                                                        |19:59
jomeier| network_id        | 9911869a-afb9-4c36-bc8c-55d9dc669d69                                                                                                                                                  |20:00
jomeier| prefix_length     | None                                                                                                                                                                                  |20:00
jomeier| project_id        | 14c34108387f4adc9605fb42736574c4                                                                                                                                                      |20:00
jomeier| revision_number   | 0                                                                                                                                                                                     |20:00
jomeier| segment_id        | None                                                                                                                                                                                  |20:00
jomeier| service_types     |                                                                                                                                                                                       |20:00
jomeier| subnetpool_id     | None                                                                                                                                                                                  |20:00
jomeier| tags              |                                                                                                                                                                                       |20:00
jomeier| updated_at        | 2019-06-12T18:11:05Z                                                                                                                                                                  |20:00
jomeier+-------------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+20:00
jomeierdamn. sorry ...20:00
jomeierbut it is there20:00
johnsomFYI, you can use http://paste.openstack.org/ for sharing large snippets.20:00
johnsomOk, those look good.  now, 'openstack network agent list"20:00
johnsomYou are looking for the openvswitch and dhcp agents to be present on the compute hosts and the controller host running the Octavia controllers (since you used the ovs-vsctl method.20:01
johnsomEach line should have a :-) and state of up20:02
jomeierhttp://paste.openstack.org/show/752836/20:02
johnsomOk, so those lines with XXX mean neutron is not running correctly or has failed20:02
jomeierI changed my hostname to: openstack-controller.sepp.com20:03
jomeierMaybe some old stuff is laying around. Thats where I tried designate20:03
jomeierBecause the agents for openstack-controller.sepp.com are all up.20:04
johnsomOk, so all of your services are running on openstack-controller.sepp.com?20:04
jomeieryes20:04
jomeierThe availability zone is "none" and not "nova". Could that be a problem?20:04
johnsomMaybe, I'm not a full neutron expert, so I'm not sure what the implications of that would be.20:05
johnsomIf that all looks ok, I would go through and double check those network setup lines and make sure the ovs port is setup right, etc.20:06
jomeierDo I have to set some attributes to the neutron port I must create? the network_type is vxlan at the moment.20:06
johnsomI would also check the port in openstack port show to make sure it's up20:06
johnsomvxlan should not matter, that is how most of us run20:07
jomeierhttp://paste.openstack.org/show/752837/20:08
johnsomAh:20:08
johnsom| binding_host_id         | openshift-controller20:08
johnsomneutron created the port on the old hostname host20:09
johnsom| binding_vif_type        | binding_failed20:09
jomeieroh. You mean I must use the fqdn ?20:09
johnsomThe host id must match the agent list host id20:09
jomeierups. Give me a minute ... I'll try that...20:09
johnsomI have to step away for a bit in a minute, but that port bind failed is your trouble20:10
jomeierThanks a lot for your help. I'll give you feedback if that was the problem.20:10
johnsomSounds good20:11
rm_worki wouldn't say slacking, so much as, working late into the night and generally having no time to sleep if i have to be up at 9am :D20:18
johnsomlol20:19
jomeier@johnsom: You were right: the binding_host_id was the major problem. No my port o-hm0 gets an IP and the binding is active ! Wow. Such a stupid error :-D20:25
jomeierNow I get this one:20:25
johnsomCool, so eventually the other LB will timeout retries and go to ERROR at which point you can delete it.20:25
jomeier2019-06-12 22:23:08.261 29488 ERROR oslo_messaging.rpc.server PlugVIPException: Error plugging amphora (compute_id: bacd9b69-197a-49e9-9c1f-79d140f2007f) into vip network e741e483-2f1d-48e6-a2d6-ad49a243e5c6.20:26
jomeier2019-06-12 22:23:08.261 29488 ERROR oslo_messaging.rpc.server20:26
jomeierBut I hope this time I can fix this problem much quicker.20:26
rm_workjohnsom: updated the network-noop driver to globally track the things it has returned so it isn't completely amnesiac, and to actually return objects with correct data, and kinda dynamically build a network/subnet tree https://review.opendev.org/#/c/660239/45/octavia/network/drivers/noop_driver/driver.py20:27
rm_workany other requests for it while i'm in here? :D20:28
johnsomNice20:28
johnsomrm_work If you didn't see above, we have nice shiny logs: http://logs.openstack.org/06/626406/9/check/octavia-v2-dsvm-scenario/240afba/controller/logs/20:29
johnsomEvidently depends-on no longer works with the old openstack.org urls....20:29
rm_workyeah i looked at them, it's neat20:29
rm_workhmm that is odd20:29
rm_worki thought the redirects were supposed to work for a long time20:30
rm_work:/20:30
johnsomI guess there was an e-mail I missed about needing to change them over to opendev.org20:30
rm_workis the logs thing ready for review?20:30
johnsomNo, I'm going to write up some docs20:30
rm_workkk20:30
rm_worki need to do docs now on multi-vip too20:30
rm_workand a couple more bugs to squash20:30
johnsomdocs for everyone!20:30
*** jomeier has quit IRC20:35
*** ccamposr__ has quit IRC20:37
*** ccamposr has joined #openstack-lbaas20:37
*** vishalmanchanda has quit IRC20:55
bodenhi... I know neutron-lbaas is retired, but we still have customers using it that need fixes :).    can I ask for some help reviewing https://review.opendev.org/#/c/660930/21:06
johnsomShould not be a problem. I agree with your approach to the release note issue, let's disable the job but keep collecting the release notes. This way when the job is fixed we will have complete release notes.21:09
*** boden has quit IRC21:10
*** ivve has quit IRC21:11
johnsomboden BTW, I got interrupted yesterday and didn't get the get patch up. I should have something posted today.21:12
*** pcaruana|afk| has quit IRC21:16
*** luksky has quit IRC21:25
*** ccamposr has quit IRC21:57
*** ccamposr has joined #openstack-lbaas21:58
rm_workjohnsom: in the doc for additional_vips, it's a list of json objects ... should I just call it a "list" and document in the description what the objects look like, or do they need sub-fields somehow?22:09
johnsomfor the api-ref parameters?22:09
*** spatel has quit IRC22:10
rm_workyes22:14
johnsomit is "array" for reasons I don't know. But, yes, "array" and then describe what is in the array22:14
rm_workhmmm we use "type: list" in many other spots22:14
johnsomhmmm, maybe list renders as "array"22:15
johnsomI remember there was something funky with that.22:15
johnsomUse whatever is used elsewhere in the file22:15
johnsomHa, we do have both.  hmmm22:16
johnsomOh, only tags is using list. I bet that is "wrong" but it seems to be taking it22:16
johnsomYeah, it says "array" here: https://docs.openstack.org/os-api-ref/latest/usage.html#parameters-file-format22:18
johnsomYeah, so the tags type is incorrect22:19
openstackgerritMichael Johnson proposed openstack/octavia-lib master: Add get methods to the driver-lib  https://review.opendev.org/66502722:31
rm_workok i can just fix those too22:32
openstackgerritMichael Johnson proposed openstack/octavia master: Add get method support to the driver-agent  https://review.opendev.org/66502922:33
johnsomboden ^^^ Those are fresh off the press, completely untested and undocumented. But wanted to give you early access to them.  They *should* work.22:34
johnsomI guess to be fair, there are doc strings... lol22:35
*** rcernin has joined #openstack-lbaas22:41
*** goldyfruit has quit IRC22:59
openstackgerritMichael Johnson proposed openstack/octavia master: Disable the kdump service in the amphora  https://review.opendev.org/66503223:00
*** goldyfruit has joined #openstack-lbaas23:11
*** ccamposr__ has joined #openstack-lbaas23:14
*** ccamposr has quit IRC23:17
*** ccamposr__ has quit IRC23:20
rm_workjohnsom: should I add a depends-on to my main `octavia` patch for the `octavia-tempest-plugin` patch? or the other way23:23
rm_worki've done the other way but if i did the depends-on from the octavia side, would it run the new tests along with every patch?23:24
rm_workbut would I have to mess with it when we were ready to merge?23:24
openstackgerritAdam Harwell proposed openstack/octavia master: Allow multiple VIPs per LB  https://review.opendev.org/66023923:26
rm_workdocs added, keepalived bug hopefully fixed23:27
rm_workso multi-vip with keepalived should work better with ipv6 as the primary vip and ipv4 as additionals23:27
rm_workjohnsom: oh, and on the ipv6-only gate, were you doing that? I could also take a stab at making the job really quick, it doesn't look hard and just starting to get some runs going to iterate on it might not be a bad idea23:30
rm_workjust to see what is even broken currently23:30
lxkongjohnsom: i can't reproduce the issue in my devstack environment (stable/stain), my lb in devstack is active within 2 or 3 mins23:37
lxkongdid't look at qemu yet, will do23:38
*** yamamoto has joined #openstack-lbaas23:49

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!