Thursday, 2021-03-25

*** dwilde has quit IRC00:04
*** tosky has quit IRC00:12
*** luksky has quit IRC00:15
*** spatel_ has joined #openstack-ansible00:36
*** spatel_ is now known as spatel00:36
*** rh-jelabarre has quit IRC00:44
*** spatel has quit IRC01:14
*** spatel_ has joined #openstack-ansible01:15
*** spatel_ is now known as spatel01:15
*** jamesdenton has quit IRC01:16
*** jamesden_ has joined #openstack-ansible01:17
*** spatel has quit IRC02:09
*** evrardjp has quit IRC03:33
*** evrardjp has joined #openstack-ansible03:33
*** jamesden_ has quit IRC04:54
*** jamesdenton has joined #openstack-ansible04:54
*** yasemind has joined #openstack-ansible05:22
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-ops master: Fix venv cleanup for uwsgi venv  https://review.opendev.org/c/openstack/openstack-ansible-ops/+/78290707:17
*** miloa has joined #openstack-ansible07:21
openstackgerritDmitriy Rabotyagov proposed openstack/ansible-hardening master: Extend timeout for RPM verification  https://review.opendev.org/c/openstack/ansible-hardening/+/78290907:47
*** jamesdenton has quit IRC07:56
*** jamesdenton has joined #openstack-ansible07:57
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_cinder master: Fix condition when to create backends  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/78291107:59
*** miloa has quit IRC08:01
*** luksky has joined #openstack-ansible08:04
*** andrewbonney has joined #openstack-ansible08:10
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-lxc_hosts stable/rocky: Determine latest base image available  https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/78273408:11
*** rpittau|afk is now known as rpittau08:13
jrossermorning08:22
jonhermorning08:22
jrosser777384 really is somehow cursed, i hope it has better luck passing tests today08:23
*** jbadiapa has joined #openstack-ansible08:31
noonedeadpunko/08:36
*** tosky has joined #openstack-ansible08:43
ebbexnoonedeadpunk, jrosser: any thoughts on why only xenial started working here? https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78279908:48
ebbexnote i'm overriding the url to (seems to me) a valid image for all test-os'es.08:49
jrosserfrom the test results?08:49
* jrosser seeing them all broken08:49
ebbexAh, yes they fail on other stuff later on. but in the beginning they failed on lxc-image. and only bionic still does.08:50
jrosserissues with tempest plugins could be that the versions are not pinned back to py2 supporting versions for those OS08:50
jrosserit's this https://zuul.opendev.org/t/openstack/build/fb5362e6304d4861aa0676e7eff2c68d/log/logs/host/lxc-cache-prep-commands.log.txt#296-30008:52
ebbexaha, cause i found that strange that neutron/master passes on centos but fails for xenial. and pinned cinder fails on centos but passes xenial. really weird.08:52
jrosserthe https repos thing is worked around on more recent branches like this https://github.com/openstack/openstack-ansible-tests/blob/master/zuul.d/playbooks/pre-gate-cleanup.yml#L21-L2708:53
jrosserand similar in the openstack-ansible repo too, depending on which jobs it is08:53
ebbexjrosser: thanks, that's a great find!08:54
ebbex:)08:55
jrossernp :)08:55
openstackgerritEbbex proposed openstack/openstack-ansible-tests stable/rocky: Update get-pip url  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78279909:09
*** sshnaidm|afk is now known as sshnaidm09:24
openstackgerritAdrien Cunin proposed openstack/openstack-ansible-ops master: Fixed venvs pattern to be more specific  https://review.opendev.org/c/openstack/openstack-ansible-ops/+/78275009:26
noonedeadpunkebbex: btw, I've already cherry-picked https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/782734 :(09:27
*** gokhani has joined #openstack-ansible09:34
noonedeadpunkI want to get this landed and backported to V before creating new V tag as it's pretty major bug :( https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/78291109:37
*** amalrajgenocidex has joined #openstack-ansible09:40
*** shyamb has joined #openstack-ansible09:54
*** amalrajgenocidex has quit IRC09:59
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_swift master: Revert "split templates to work around configparser bug"  https://review.opendev.org/c/openstack/openstack-ansible-os_swift/+/78295610:09
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_swift master: Revert "split templates to work around configparser bug"  https://review.opendev.org/c/openstack/openstack-ansible-os_swift/+/78295610:11
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_swift master: Revert "split templates to work around configparser bug"  https://review.opendev.org/c/openstack/openstack-ansible-os_swift/+/78295610:16
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_trove master: Use ansible_facts[] instead of fact variables  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/78073210:20
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_trove master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/77205210:21
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_trove master: Updated from OpenStack Ansible Tests  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/78037610:21
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_trove master: [goal] Deprecate the JSON formatted policy file  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/78231410:21
*** rodolof has joined #openstack-ansible10:27
openstackgerritMerged openstack/ansible-hardening master: Extend timeout for RPM verification  https://review.opendev.org/c/openstack/ansible-hardening/+/78290910:32
openstackgerritMerged openstack/openstack-ansible-os_neutron master: Adding support of subnet_dns_publish_fixed_ip extension in ml2 plugin  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/77224510:37
jrossernoonedeadpunk: i was just thinking again about the qemu processes overrides you mentioned yesterday, do we need a patch for that?10:38
jrosserthere is this which we could extend with a default set? https://opendev.org/openstack/openstack-ansible-os_nova/src/branch/master/defaults/main.yml#L49410:40
noonedeadpunkyep, I used that to cover the issue...10:40
noonedeadpunkSo I'm not sure about setting a default...10:41
noonedeadpunkit was actually mine question if we need a patch :)10:41
noonedeadpunkor maybe we should jsut have it documented...10:41
jrosserit seems like a really common case with a good sized ceph cluster10:42
noonedeadpunkyeah, agree. so we can set it in case driver is ceph?10:42
noonedeadpunkbut that actually would be also the case if cinder volumes are ceph based (while nova ephemeral are not)10:43
jrosserit could be nova with rbd, or cinder with rbd10:43
noonedeadpunkoh, we have nova_cinder_rbd_inuse, right10:43
noonedeadpunkwill make a patch then10:43
jrossercool10:43
jrosserwe were going to try to look again at the galera upgrade stuff10:44
jrosserseems really hard dealing with the old root user authentication setup + move to the new root+admin style10:44
noonedeadpunkwe also super need to merge masters bump I think (which is also broken on upgrades in tempest)10:46
noonedeadpunkand I must spend time on playing with trove atm :(10:46
noonedeadpunkto get it finally deployed in prod10:46
jrosseroh really many failures to retrieve u-c on the master bump patch10:50
noonedeadpunkyeah, there was some issue yestarday with the mirrors I guess10:51
noonedeadpunkBtw I commented https://review.opendev.org/c/openstack/openstack-ansible/+/777990/15/playbooks/common-playbooks/cinder.yml (not sure if you saw)10:51
jrosseryeah reasonable10:52
jrosser-ENOTIME :(10:52
noonedeadpunkreally same here :(10:53
jrosseroh no wait10:54
jrosserarent those tasks separated out to make sure fact gathering happens properly with --limit?10:54
jrosseror tags maybe, becasue of the tags: always10:54
noonedeadpunkyes, it was because of the tags. But we can put tags on pre tasks?10:55
noonedeadpunkI haven't really tested it, but I guess it should work10:55
jrosseralso looking at how cinder_serial would affect things10:56
noonedeadpunkand that's fair comment...10:56
noonedeadpunkit actually might affect10:56
jrosserfeels like right now we gather all facts for the whole group up front in a really simple way10:56
noonedeadpunkstill don't need repo_all?10:57
jrosserno i can take that away10:57
noonedeadpunkbut agree, didn't thought about serial10:57
openstackgerritJonathan Rosser proposed openstack/openstack-ansible master: Gather minimal facts in common playbooks  https://review.opendev.org/c/openstack/openstack-ansible/+/77799010:58
openstackgerritMerged openstack/openstack-ansible-ops master: Fixed venvs pattern to be more specific  https://review.opendev.org/c/openstack/openstack-ansible-ops/+/78275011:00
openstackgerritEbbex proposed openstack/openstack-ansible-tests stable/rocky: Update get-pip url  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78279911:04
ebbexjrosser: https://review.opendev.org/c/openstack/openstack-ansible-rabbitmq_server/+/782707 ReadTimeoutError("HTTPSConnectionPool(host='opendev.org', port=443): Read timed out. ERROR: 502 Server Error: Proxy Error for url...11:09
*** gokhani has quit IRC11:09
ebbexany ideas why there's a proxy-error on xenial but not centos7?11:09
noonedeadpunkrecheck?11:10
noonedeadpunkor its persistant?11:10
noonedeadpunkI just saw lots of infras issues yestarday11:10
*** gokhani has joined #openstack-ansible11:11
ebbexit was twice yesterday, but i'll give the recheck a go. strange it didn't have any issues with centos7.11:11
noonedeadpunkmight be different regions11:12
*** shyamb has quit IRC11:32
openstackgerritMerged openstack/openstack-ansible-os_trove master: Retry on creating trove network  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/78281411:35
openstackgerritMerged openstack/openstack-ansible-os_nova master: Fix usage of tags  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/78279611:36
*** gokhani has quit IRC11:43
noonedeadpunkdoh 777384 failed again11:43
*** rh-jelabarre has joined #openstack-ansible11:52
openstackgerritEbbex proposed openstack/openstack-ansible-tests stable/rocky: Update get-pip url  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78279911:57
ebbexwow, this passed: https://review.opendev.org/c/openstack/openstack-ansible-rabbitmq_server/+/78270711:58
*** shyamb has joined #openstack-ansible11:59
jonher"Failed to connect to opendev.org port 443"  in various jobs, infra has issues with gitea/lb so that probably explains your issue before as well11:59
openstackgerritEbbex proposed openstack/openstack-ansible-tests stable/queens: Remove temporary override for rabbitmq  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78298712:03
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_gnocchi master: Fix gnocchi installation for new pip resolver  https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/78151312:05
noonedeadpunkbtw I hope this will fix gnocchi instlllation without need to maintain local set of requirements12:05
noonedeadpunkat least worked locally12:05
noonedeadpunkok, so. master bump fails because of the `pymysql.err.OperationalError: (2013, 'Lost connection to MySQL server during query')` for nova-scheduler o_O12:09
noonedeadpunkmariadb bug?12:09
noonedeadpunkdisregard, probably it was during mariadb upgrade12:10
noonedeadpunksuper weird. tempest fails with `Details: (TestServerBasicOps:test_server_basic_ops) Server 40ce5b33-d2d8-470e-b8c8-776c53efce8c failed to reach ACTIVE status and task state "None" within the required time (600 s). Current status: BUILD. Current task state: scheduling.` https://be7a0942ef49bf0b6abe-b6dab27f1b1aff436bdeb943180547ea.ssl.cf5.rackcdn.com/780434/8/check/openstack-ansible-upgrade-aio_metal-ubuntu-focal/85c4097/logs/openstack/12:12
noonedeadpunkaio1-utility/stestr_results.html12:12
noonedeadpunksheduler looks like it does it's job http://paste.openstack.org/show/803911/12:13
noonedeadpunkand it even appears in compute log and being rescheduled, but never spawned for some reason12:16
noonedeadpunkhttps://zuul.opendev.org/t/openstack/build/85c40976164c4b1785aefe0c1a4ddab9/log/logs/host/nova-compute.service.journal-11-24-25.log.txt12:16
openstackgerritEbbex proposed openstack/openstack-ansible-os_tempest stable/rocky: Pin tempest-plugins to last tag supporting py2.7  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/78299012:17
openstackgerritMerged openstack/openstack-ansible-os_cinder master: Fix condition when to create backends  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/78291112:18
openstackgerritEbbex proposed openstack/openstack-ansible-lxc_hosts stable/rocky: Determine latest base image available  https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/78273412:18
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_cinder stable/victoria: Fix condition when to create backends  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/78296312:19
*** shyamb has quit IRC12:20
*** gokhani has joined #openstack-ansible12:21
jrossernoonedeadpunk: rabbit log looks kind of strange https://zuul.opendev.org/t/openstack/build/85c40976164c4b1785aefe0c1a4ddab9/log/logs/host/rabbitmq-server.service.journal-11-24-25.log.txt12:38
noonedeadpunkWell it looks like nokill systemd option is there12:39
noonedeadpunkor you mean `Logger - error: {removed_failing_handler,rabbit_log}`?12:40
jrossermaybe i expected to see some actual debug logging12:40
jrosserperhaps we dont enable any12:41
jrosserjust from 10:37:57 to 11:13:42 (the exact point the tempest test bails out) theres nothing at all12:41
jrossermay just be co-incidence that 11:13:39 is the first point in the call logging from tempest12:43
noonedeadpunkI actually would expect to see the reason why instance was not spawned in compute log...12:45
noonedeadpunkinstead of just the matter of fact that it hasn't12:45
noonedeadpunkdoh12:48
noonedeadpunkhttps://zuul.opendev.org/t/openstack/build/85c40976164c4b1785aefe0c1a4ddab9/log/logs/host/nova-conductor.service.journal-11-24-25.log.txt#321512:48
noonedeadpunk`Connection failed: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: self signed certificate`12:49
LowKeynoonedeadpunk: my gnocchi issue was fixed, i re-check one gnocchi container was down. 502 error was related to haproxy not able to reached one of the container, but this not shown on log, i just manually figure out..12:57
jonherit should have shown in the haproxy log on the number of backends available at least, if not make sure the healthchecks are ok12:59
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Disable ssl for rabbitmq during upgrade  https://review.opendev.org/c/openstack/openstack-ansible/+/78299613:03
noonedeadpunkyep, like `echo "show stat" | nc -U /var/run/haproxy.stat`13:05
noonedeadpunkbut I'd say if haproxy can't reach one container, it should jsut go down. Which means that our check request is wrong13:05
noonedeadpunkand that would be great to find out what exactly and fix it13:06
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Disable ssl for rabbitmq during upgrade  https://review.opendev.org/c/openstack/openstack-ansible/+/78299613:07
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Bump SHAs for master  https://review.opendev.org/c/openstack/openstack-ansible/+/78043413:08
*** spatel_ has joined #openstack-ansible13:12
*** spatel_ is now known as spatel13:12
spatelnoonedeadpunk i have replaced SSD and not its showing 99% healthy13:13
spatel[root@ostack-infra-01 ~]# smartctl -A -d sat+cciss,0 /dev/sda | grep Wear_Leveling_Count13:13
spatel177 Wear_Leveling_Count     0x0013   099   099   000    Pre-fail  Always       -       513:13
noonedeadpunkspatel: it's _showing_ you meant?13:18
spateli meant i can see health, it was 001 and now its 09913:18
noonedeadpunkyeah)13:19
spateli am good and safe :)13:19
noonedeadpunkI'd say you replaced it just in time)13:19
*** dwilde has joined #openstack-ansible13:19
spatelI have just replaced single drive at present and i will replace next drive after 1 month so i will have big time window between two SSD lifetime13:20
spateli don't want they go down sametime :)13:20
*** gokhani61 has joined #openstack-ansible13:25
ebbexhttps://review.opendev.org/c/openstack/openstack-ansible-tests/+/782799 Using cached setuptools-54.2.0.tar.gz ? should that be a problem on xenial?13:25
*** gokhani has quit IRC13:26
*** jamesdenton has quit IRC13:26
*** jamesdenton has joined #openstack-ansible13:26
LowKeythanks for the command noonedeadpunk13:28
LowKeyyeah, i just simply destroy and create the container... instead of check the problem..13:28
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_tempest master: Add trove tempest support  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/78300213:33
*** amalrajgenocidex has joined #openstack-ansible13:43
amalrajgenocidexI deployed openstack-train with octavia-dashboard13:44
amalrajgenocidexbut flavor and subnet is not loading in horizon13:45
amalrajgenocidexhttps://i.imgur.com/ZSB6ZIW.png13:45
amalrajgenocidexany idea?13:45
noonedeadpunkamalrajgenocidex: Does this user has right permissions?13:47
noonedeadpunkI mean octavia requires extra role assigned13:48
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_tempest master: Add trove tempest support  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/78300213:48
amalrajgenocidexnoonedeadpunk yes, I used openstack-ansible and made sure the user has all the permissions13:53
amalrajgenocidexindeed I can see the output in inspect element13:53
amalrajgenocidexhttps://i.imgur.com/YKl1oR1.png13:53
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_tempest master: Extend telemetry testing  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/78300413:54
noonedeadpunkoh, right, it's admin user13:54
noonedeadpunkOh, right13:55
noonedeadpunkI think it's because no flavors exist13:55
noonedeadpunkhttps://docs.openstack.org/octavia/train/admin/flavors.html#flavors13:56
amalrajgenocidexhttps://i.imgur.com/50QDqSb.png13:57
johnsomNo flavors are required13:57
amalrajgenocidexI already tried creating multiple flavors13:57
amalrajgenocidexwith both SINGLE and ACTIVE-STANDBY13:57
noonedeadpunkhm. I just recall some issue related to them with horizon plugin some time ago13:57
jonhermight the red X and yellow warnings at the top of the inspect window be related?13:58
noonedeadpunkdespite the were not required13:58
johnsomAre you sure you are using a recent version of dashboard?13:58
amalrajgenocidexjohnsom yes, using latest version of stable/train octavia-dashboard13:59
noonedeadpunkbtw I had it working for train for sure...13:59
johnsomAre there errors reported in the horizon log file?14:00
openstackgerritAndrew Bonney proposed openstack/openstack-ansible-haproxy_server master: Allow HAProxy stats to be pinned to one or more processes  https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/78300714:02
noonedeadpunkcould think of some horizon specific policies, but it's an admin user...14:06
amalrajgenocidexyeah, don't think it's related to permissions because we can see the flavors in inspect element14:08
noonedeadpunkso rly, what are the errors in the console?14:08
noonedeadpunkbrowser console14:09
openstackgerritAndrew Bonney proposed openstack/openstack-ansible-haproxy_server master: Allow HAProxy stats to be pinned to one or more processes  https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/78300714:09
amalrajgenocidexhttps://i.imgur.com/JgBmcxC.png14:11
amalrajgenocidexnoonedeadpunk it's related to barbican,14:11
noonedeadpunkand I assume you're able to create a LB with CLI?14:12
noonedeadpunkI found that bug I was talking about. https://review.opendev.org/c/openstack/openstack-ansible-os_horizon/+/70246414:14
noonedeadpunkwe backported it back to train14:14
noonedeadpunkwe backported fix of this bug14:15
noonedeadpunkso in case you're installing master octavia ui plugin, it ended up that way14:15
openstackgerritEbbex proposed openstack/openstack-ansible-tests stable/rocky: Update get-pip url  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78279914:22
amalrajgenocidexnoonedeadpunk yes, backend is working14:25
amalrajgenocidexLet me go throgh the bug14:26
amalrajgenocidexI think it is not applicable here, we are already using horizon_git_track_branch14:30
LowKeynoonedeadpunk: you have idea on this issue ? http://paste.openstack.org/show/NknFX8HH7GnwnGQkso5M/ , i'm currently create zone, but getting ouput "no_servers_configured" , something i miss up ?14:30
amalrajgenocidexI even tried using victoria branch for octavia-dashboard, and saw the same error , with availaibility zone  errors also which is expected since octavia in train does not have availability zone14:31
johnsomAgain, AZ do not need to be created, the dashboard will work with just the defaults.14:35
noonedeadpunkLowKey: how does your pools.yaml look like?14:37
noonedeadpunkamalrajgenocidex: ok, gotcha, was not sure how new your train deployment is. if it's there, it should be fine...14:38
noonedeadpunkso I'm kind of out of the good ideas14:38
amalrajgenocidexnoonedeadpunk ok. No problem. Thanks for helping. Let me troubleshoot a bit more.14:40
LowKeynoonedeadpunk: i dont have it, http://paste.openstack.org/show/803919/ , i need to manually create it ?14:40
amalrajgenocidexIf didn't work, will have to fall back to cli method14:40
LowKeyi can see it on /openstack/venvs/designate-22.1.0/lib/python3.8/site-packages/designate/tests/resources/pools_yaml/pools.yaml14:40
noonedeadpunkamalrajgenocidex: you can try setting `octavia_dashboard_git_install_branch: 4.0.1` as I have it running nicely in one of my train clouds atm14:43
amalrajgenocidexOk. Thanks. Let me give it a try14:44
noonedeadpunkLowKey: https://opendev.org/openstack/openstack-ansible-os_designate/src/branch/master/defaults/main.yml#L104-L13714:44
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_designate master: Generate designate_pool_uuid dynamically  https://review.opendev.org/c/openstack/openstack-ansible-os_designate/+/77184114:44
noonedeadpunkYou can also take a look at https://satishdotpatel.github.io//designate-integration-with-powerdns/14:45
noonedeadpunkso actually designate usually does not face public endpoints and not used as a frontend. Instead you connect it to proper DNS cluster and designate manage it14:46
LowKeynoonedeadpunk: ok thanks for guide14:50
*** macz_ has joined #openstack-ansible14:58
*** juanoterocas has joined #openstack-ansible15:02
*** dwilde has quit IRC15:04
*** dwilde has joined #openstack-ansible15:06
*** jamesdenton has quit IRC15:08
*** jamesdenton has joined #openstack-ansible15:10
*** jamesdenton has quit IRC15:17
*** jamesdenton has joined #openstack-ansible15:18
*** yasemind has quit IRC15:37
*** yasemind has joined #openstack-ansible15:49
*** yasemind has quit IRC16:06
*** yasemind has joined #openstack-ansible16:12
openstackgerritMerged openstack/openstack-ansible master: Use proper distro path for senlin and mistral  https://review.opendev.org/c/openstack/openstack-ansible/+/77738416:16
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_nova master: Do not use imports inside included task  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/78304716:17
jrosser \o/ 777384 merged, wow16:20
*** mensis has quit IRC16:26
noonedeadpunkI have no idea how https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/782796 has merged considering it should fail...16:27
jrosseroh dear! what error did we miss?16:42
*** amalrajgenocidex has quit IRC16:44
openstackgerritEbbex proposed openstack/openstack-ansible-tests stable/rocky: Update get-pip url  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/78279916:47
noonedeadpunkjrosser: http://paste.openstack.org/show/803922/16:48
noonedeadpunk783047 fixing it...16:48
noonedeadpunkwas deploying aio and catched that16:49
*** rpittau is now known as rpittau|afk17:06
*** sshnaidm is now known as sshnaidm|afk17:11
*** juanoterocas has quit IRC17:17
*** tobberydberg has quit IRC17:18
*** brad[] has quit IRC17:19
*** bverschueren has quit IRC17:20
*** bradm has quit IRC17:20
*** mugsie has quit IRC17:20
*** jhesketh has quit IRC17:20
*** arxcruz has quit IRC17:20
*** mcarden has quit IRC17:20
*** tacco has quit IRC17:20
*** pcaruana has quit IRC17:20
*** kukacz has quit IRC17:20
*** yasemind has quit IRC17:20
*** logan- has quit IRC17:20
*** jmccrory has quit IRC17:20
*** mugsie has joined #openstack-ansible17:20
*** tobberydberg has joined #openstack-ansible17:20
*** bverschueren has joined #openstack-ansible17:20
*** jmccrory has joined #openstack-ansible17:20
*** pcaruana has joined #openstack-ansible17:20
*** arxcruz has joined #openstack-ansible17:20
*** jhesketh has joined #openstack-ansible17:20
*** fridtjof[m] has quit IRC17:20
*** tinwood has quit IRC17:20
*** mrda has quit IRC17:20
*** akahat has quit IRC17:20
*** dpawlik has quit IRC17:20
*** tinwood has joined #openstack-ansible17:20
*** mrda has joined #openstack-ansible17:20
*** dpawlik has joined #openstack-ansible17:21
*** masterpe has quit IRC17:21
*** brad[] has joined #openstack-ansible17:21
*** gokhani61 has quit IRC17:21
*** manti has quit IRC17:21
*** logan- has joined #openstack-ansible17:21
*** kukacz has joined #openstack-ansible17:23
*** kukacz has quit IRC17:27
*** kukacz has joined #openstack-ansible17:27
*** akahat has joined #openstack-ansible17:30
noonedeadpunkdoh, and tempestconf is broken with unknown issue...17:32
*** tinwood has quit IRC17:42
*** evrardjp has quit IRC17:42
*** cloudnull has quit IRC17:42
*** dmsimard has quit IRC17:42
*** janno_ has quit IRC17:42
*** dpawlik has quit IRC17:42
*** rodolof has quit IRC17:42
*** klamath_atx has quit IRC17:42
*** ebbex has quit IRC17:42
*** sshnaidm|afk has quit IRC17:42
*** LowKey has quit IRC17:42
*** Anticimex has quit IRC17:42
*** spy has quit IRC17:42
*** sc__ has quit IRC17:42
*** jroll has quit IRC17:42
*** gixx has quit IRC17:42
*** johanssone has quit IRC17:42
*** frickler has quit IRC17:42
*** admin0 has quit IRC17:42
*** openstackgerrit has quit IRC17:42
*** crazzy has quit IRC17:42
*** lemko has quit IRC17:42
*** chkumar|ruck has quit IRC17:42
*** gary_perkins has quit IRC17:42
*** jonher has quit IRC17:42
*** prometheanfire has quit IRC17:42
*** noonedeadpunk has quit IRC17:42
*** spotz has quit IRC17:42
*** akahat has quit IRC17:42
*** luksky has quit IRC17:42
*** owalsh has quit IRC17:42
*** zbr|rover has quit IRC17:42
*** poopcat has quit IRC17:42
*** trident has quit IRC17:42
*** partlycloudy has quit IRC17:42
*** corvus has quit IRC17:42
*** gouthamr has quit IRC17:42
*** gmann has quit IRC17:42
*** alanmeadows has quit IRC17:42
*** Open10K8S has quit IRC17:42
*** zbr|rover has joined #openstack-ansible17:42
*** Anticimex has joined #openstack-ansible17:42
*** corvus has joined #openstack-ansible17:42
*** owalsh has joined #openstack-ansible17:42
*** sshnaidm|afk has joined #openstack-ansible17:43
*** noonedeadpunk_ has joined #openstack-ansible17:43
*** jonher has joined #openstack-ansible17:43
*** spy has joined #openstack-ansible17:43
*** luksky has joined #openstack-ansible17:44
*** johanssone has joined #openstack-ansible17:44
*** evrardjp has joined #openstack-ansible17:44
*** rodolof has joined #openstack-ansible17:44
*** tinwood has joined #openstack-ansible17:44
*** trident has joined #openstack-ansible17:45
*** klamath_atx has joined #openstack-ansible17:45
*** gmann has joined #openstack-ansible17:45
*** janno has joined #openstack-ansible17:46
*** cyberpear has quit IRC17:46
*** partlycloudy has joined #openstack-ansible17:46
*** frickler_ has joined #openstack-ansible17:47
*** cyberpear has joined #openstack-ansible17:48
*** frickler_ has quit IRC17:49
*** akahat has joined #openstack-ansible17:50
*** cloudnull has joined #openstack-ansible17:51
*** jhesketh has quit IRC17:52
*** jhesketh has joined #openstack-ansible17:52
*** dwilde has quit IRC17:53
*** fridtjof[m] has joined #openstack-ansible17:54
*** fresta_ has quit IRC17:57
*** mubix has quit IRC17:57
*** mwhahaha has quit IRC17:57
*** johnsom has quit IRC17:57
*** NobodyCam has quit IRC17:57
*** melwitt has quit IRC17:57
*** CeeMac has quit IRC17:57
*** nicolasbock has quit IRC17:57
*** mnaser has quit IRC17:57
*** mwhahaha has joined #openstack-ansible17:57
*** CeeMac has joined #openstack-ansible17:57
*** fresta has joined #openstack-ansible17:57
*** NobodyCam has joined #openstack-ansible17:57
*** nicolasbock has joined #openstack-ansible17:57
*** melwitt has joined #openstack-ansible17:58
*** mubix has joined #openstack-ansible17:58
*** pcaruana has quit IRC17:58
*** jmccrory has quit IRC17:58
*** NewJorg has quit IRC17:58
*** MrClayPole has quit IRC17:58
*** dasp_ has quit IRC17:58
*** rpittau|afk has quit IRC17:58
*** jrosser has quit IRC17:58
*** snadge has quit IRC17:58
*** fyx has quit IRC17:58
*** hindret has quit IRC17:58
*** mmercer has quit IRC17:58
*** PrinzElvis has quit IRC17:58
*** sri_ has quit IRC17:58
*** jungleboyj has quit IRC17:58
*** fungi has quit IRC17:58
*** mnaser has joined #openstack-ansible17:58
*** johnsom has joined #openstack-ansible17:58
*** NewJorg has joined #openstack-ansible17:58
*** MrClayPole has joined #openstack-ansible17:58
*** dasp has joined #openstack-ansible17:58
*** fyx has joined #openstack-ansible17:58
*** hindret has joined #openstack-ansible17:58
*** PrinzElvis has joined #openstack-ansible17:58
*** sri_ has joined #openstack-ansible17:58
*** jrosser has joined #openstack-ansible17:58
*** rpittau|afk has joined #openstack-ansible17:58
*** jungleboyj has joined #openstack-ansible17:58
*** jmccrory has joined #openstack-ansible17:58
*** pcaruana has joined #openstack-ansible17:58
*** snadge has joined #openstack-ansible17:58
*** mmercer has joined #openstack-ansible17:58
*** openstack has joined #openstack-ansible18:05
*** ChanServ sets mode: +o openstack18:05
*** jungleboyj has quit IRC18:05
*** jrosser has quit IRC18:05
*** nicolasbock has quit IRC18:05
*** mwhahaha has quit IRC18:05
*** partlycloudy has quit IRC18:05
*** janno has quit IRC18:05
*** kukacz has quit IRC18:05
*** Daemoen has quit IRC18:05
*** persia has quit IRC18:05
*** persia has joined #openstack-ansible18:05
*** partlycloudy has joined #openstack-ansible18:05
*** Daemoen has joined #openstack-ansible18:05
*** trident has joined #openstack-ansible18:06
*** jungleboyj has joined #openstack-ansible18:06
*** cyberpear has quit IRC18:06
*** klamath_atx has quit IRC18:06
*** gmann has quit IRC18:06
*** Adri2000 has quit IRC18:06
*** Jeffrey4l has quit IRC18:06
*** janno has joined #openstack-ansible18:06
*** Jeffrey4l has joined #openstack-ansible18:06
*** gmann has joined #openstack-ansible18:07
*** klamath_atx has joined #openstack-ansible18:07
*** Adri2000 has joined #openstack-ansible18:07
*** Guest76033 has joined #openstack-ansible18:07
*** mwhahaha has joined #openstack-ansible18:07
*** cyberpear has joined #openstack-ansible18:07
*** jrosser has joined #openstack-ansible18:08
*** andrewbonney has quit IRC18:08
*** nicolasbock has joined #openstack-ansible18:08
*** cyberpear has quit IRC18:08
*** cyberpear has joined #openstack-ansible18:10
*** kukacz has joined #openstack-ansible18:10
*** akahat has joined #openstack-ansible18:10
*** mugsie has quit IRC18:16
*** macz_ has quit IRC18:16
*** jbadiapa has quit IRC18:16
*** waxfire has quit IRC18:16
*** priteau has quit IRC18:16
*** odyssey4me has quit IRC18:16
*** logan- has quit IRC18:16
*** mrda has quit IRC18:16
*** tobberydberg has quit IRC18:16
*** jamesdenton has quit IRC18:16
*** rh-jelabarre has quit IRC18:16
*** tosky has quit IRC18:16
*** Underknowledge has quit IRC18:16
*** mgariepy has quit IRC18:16
*** d34dh0r53 has quit IRC18:16
*** irclogbot_1 has quit IRC18:16
*** simondodsley has quit IRC18:16
*** gundalow has quit IRC18:16
*** Brace has quit IRC18:16
*** mugsie has joined #openstack-ansible18:16
*** Underknowledge has joined #openstack-ansible18:17
*** jamesdenton has joined #openstack-ansible18:17
*** macz_ has joined #openstack-ansible18:17
*** tosky has joined #openstack-ansible18:17
*** tobberydberg has joined #openstack-ansible18:18
*** logan- has joined #openstack-ansible18:18
*** odyssey4me has joined #openstack-ansible18:19
*** irclogbot_1 has joined #openstack-ansible18:20
*** rh-jelabarre has joined #openstack-ansible18:21
*** fridtjof[m] has quit IRC18:29
*** pcaruana has quit IRC18:31
*** dwilde has joined #openstack-ansible18:31
*** pcaruana has joined #openstack-ansible18:32
*** mgariepy has joined #openstack-ansible18:33
*** Guest84754 has joined #openstack-ansible18:46
*** fridtjof[m] has joined #openstack-ansible18:59
*** Guest84754 has quit IRC19:05
*** masterpe has joined #openstack-ansible19:07
*** manti has joined #openstack-ansible19:07
*** Guest84754 has joined #openstack-ansible19:09
*** spatel_ has joined #openstack-ansible19:15
*** spatel_ is now known as spatel19:15
spatelnoonedeadpunk_ do you know if i can change output formate of this command ansible compute_all -m shell -a "hostname"19:16
*** jralbert has joined #openstack-ansible19:28
jralbertjrosser: here's one of my compute nodes, running os-nova-install.yml limited to just that node, logging its attempt to build a venv for nova, and failing on cloning from opendev: http://paste.openstack.org/show/803931/19:29
mgariepyjralbert, maybe stalled facts.19:34
mgariepyif the fact of the repo server are more than 24 hours old they might be ignored.19:35
jrosseryes, then it thinks there is no build server iirc19:36
jralbertSo, if we include the repo containers in the limit, they'll get fresh facts and be used?19:36
mgariepywhat's in your /root/.pip/pip.conf ?19:36
jralbertno such file19:37
jrosserwhich release is this?19:37
mgariepytrain19:38
jralbertyep19:38
jrosserthe thing that makes it use the repo server is this variable evaulating properly https://github.com/openstack/ansible-role-python_venv_build/blob/stable/train/tasks/python_venv_wheel_build.yml#L2419:39
mgariepyansible repo_all -m setup ?19:40
mgariepyto refresh the facts.19:40
jrosserthis is what is making the fact cache timeout https://github.com/openstack/openstack-ansible/blob/master/scripts/openstack-ansible.rc#L4219:42
jralbertbut doesn't the task you linked to in the venv wheel build playbook specifically gather new facts for the build targets?19:44
jralbertbecause indeed it does reference the repo servers in our environment in that task: http://paste.openstack.org/show/803932/19:45
jralbertBut the block that follows it is nonetheless delegated to the compute node19:45
spateljrosser noonedeadpunk_ i get sometime to put this together on my blog - https://satishdotpatel.github.io//openstack-ansible-inventory/19:45
jrosserjralbert: maybe worth putting in debug tasks to print venv_build_host and work back to why it doesnt come out as expected19:49
jrossera debug: printing venv_build_targets is probably a good move too19:50
jralbertYes, I guess I'll have to19:51
jrosseri don't know if --limit adjusts the cached facts which are loaded19:53
jrosserit may well be that adding repo_all to your --limit helps19:54
jrosserbecuase i suspect that this is where the "delegate to self" is coming from https://github.com/openstack/ansible-role-python_venv_build/blob/master/vars/main.yml#L6419:54
*** spatel has quit IRC19:58
*** zigo has joined #openstack-ansible20:03
*** spatel_ has joined #openstack-ansible20:05
*** spatel_ is now known as spatel20:05
*** spatel has quit IRC20:24
*** priteau has joined #openstack-ansible20:36
*** mcarden has joined #openstack-ansible21:23
*** mgariepy has quit IRC21:27
*** macz_ has quit IRC21:32
*** jamesdenton has quit IRC21:48
*** jamesdenton has joined #openstack-ansible21:49
*** openstackgerrit has joined #openstack-ansible21:53
openstackgerritMerged openstack/openstack-ansible-haproxy_server master: Allow HAProxy stats to be pinned to one or more processes  https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/78300721:53
*** bradm has joined #openstack-ansible22:10
*** mgariepy has joined #openstack-ansible22:27
jralbertI'm still totally stuck trying to get nova pulled down into the venvs on my compute nodes. I see that the repo containers have a current pull of the openstackgit repos; is there an easy way to tell the plays to target those copies instead of opendev?23:13
*** rh-jelabarre has quit IRC23:18
fungii have a feeling your attempts may be correlated to the incidents where our git backends are falling over, if you're doing an ansible fireball asking 600 machines to clone nova at the same moment23:31
fungiwe saw the problem start back up again between 22:25 and 22:30 utc23:32
jralbertSo conveniently, all our attempts will appear to be coming from a single NAT'd outbound IP. If we're the ones what done it, you should be able to very clearly see that in your logs23:39
jralbertwe are changing the OSA config to target github instead23:39
fungiwell, i think the problem is the nature of the requests, 600 requests is a drop in the bucket, hardly moves the needle, but when the backend tries to load nova in memory that many times at once...23:41
fungiso that source address is doing a tiny fraction of the request count of, say, the typical search engine spiders which hammer the site continually23:42
jralbertdo you have a candidate source address you think is to blame for those particular requests?23:42
jralbertif we can establish this is the cause of the issue, we may want to consider suggesting whether OSA by default should target the github mirrors instead of opendev23:43
fungii don't which is why i'm not convinced that's it... when we disable the backend getting hit we sometimes see the problem move to a different backend, but when i try to correlate addresses which moved from the disabled backend to the next one which was observed to git hit, and then the next one after that's been disabled, the intersection of those sets is not conclusive23:44
fungii was hoping to try to correlate by which times we saw the problem arise, but if you're not sure when you're running things, i suppose that won't help23:45
jralbertbecause the plays take so long to run, it's a bit tricky to pin down when we'd be making the requests, but I can try to get that if it'll be helpful23:46
fungiunfortunately the way the load balancer is implemented, it can't see the request details because it's not terminating ssl, but the backends don't know which client addresses they're talking to because the load balancer is the source address they see for each connection, so i can't correlate specific requests to specific clients23:47
fungiat least not without painstaking matching of source port numbers between the haproxy log and the apache log23:48
fungijralbert: if you can confirm the playbook would have started asking servers to clone nova sometime around 22:25 or 22:30 utc, that might help23:49
fungii'm mainly hoping to get some idea of what kinds of traffic trigger this scenario, so we can better design around them23:49
jralbertWell, I'm not sure if this will give you exactly what you want, but here's a collection of all the ansible log entries that correspond to the clone action during our upgrade, grouped by minute: http://paste.openstack.org/show/803936/23:51
fungithanks, i'll take a look23:51
jralbertThose timestamps should be UTC23:51
fungialso, i'll admit i'm not that familiar with how openstack-ansible is meant to operate, but if i were responsible for deploying updates of the same git repository to 600 servers i'd clone it once and then rsync it to them. i've seen what trying to repeatedly clone something from github is like and it's not pretty either23:52
jralbertWell, like a lot of openstack things, "it's complicated". OSA is intended to do exactly what you describe, and they've tried to limit the cost of builds - but, in big environments it gets hairy.23:53
jralbertso for us, running what I'd consider a medium-large openstack environment, the compute node plays take many many hours to run. In order to minimize disruption to our users, when we do a major version upgrade we do the control plane first, and then the compute nodes separately.23:54
jralbertThis requires running the playbooks with limit flags, which can produce some unexpected results, as the tests are based on the assumption of an unlimited play run that targets all hosts23:55
jralbertSo it looks like for us OSA is not understanding that it can use the repo servers it configures to host the git content, and is instead sourcing it from upstream every time23:56
jralbertUnfortunately for us, the first set of failures left us without a running nova-compute on many compute nodes, which produces a visible outage for users23:57
fungiyeah, that's definitely no good23:57
jralbertSo we don't have a ton of time to do debug, we really need to get things running again in a hurry, and that's our focus right now23:58
jralbertI would be delighted to be involved in helping figure out a way to manage larger environments going forward, once we're out of the woods23:58
fungiwell, i'm doing my best to keep our git servers up, i've not blocked anything, it's just too bad it can't space out the requests a bit23:58
fungiand yeah, due to our current load balancer and backend limitations, requests are persisted by source address so that multiple requests for a single git operation won't end up split between different backends and wind up looking for the wrong packfiles23:59

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!