Monday, 2019-12-16

*** tosky has quit IRC00:02
*** gregwork has quit IRC01:05
*** hwoarang_ has quit IRC02:22
*** cshen has joined #openstack-ansible02:49
*** cshen has quit IRC02:54
*** cshen has joined #openstack-ansible03:25
*** cshen has quit IRC03:29
*** cshen has joined #openstack-ansible03:40
*** cshen has quit IRC03:46
*** mbuil has quit IRC04:08
*** stingrayza has quit IRC04:08
*** openstackgerrit has quit IRC04:08
*** goldenfri has quit IRC04:08
*** dmsimard has quit IRC04:08
*** Jeffrey4l has quit IRC04:08
*** cosmicsound has quit IRC04:08
*** tbarron has quit IRC04:08
*** django has quit IRC04:08
*** Nick_A has quit IRC04:08
*** nsmeds has quit IRC04:08
*** aspiers has quit IRC04:08
*** szaher has quit IRC04:08
*** waxfire has quit IRC04:08
*** jmccrory has quit IRC04:08
*** openstackstatus has quit IRC04:11
*** openstackstatus has joined #openstack-ansible04:12
*** ChanServ sets mode: +v openstackstatus04:12
*** hwoarang has joined #openstack-ansible04:20
*** dave-mccowan has quit IRC04:23
*** goldyfruit has joined #openstack-ansible04:27
*** goldyfruit has quit IRC04:58
*** mensis has quit IRC05:04
*** rgogunskiy has joined #openstack-ansible05:12
*** rgogunskiy has quit IRC05:17
*** raukadah is now known as chkumar|rover05:22
*** gokhani has joined #openstack-ansible05:39
*** elenalindq has joined #openstack-ansible06:07
*** rgogunskiy has joined #openstack-ansible07:03
*** pcaruana has joined #openstack-ansible07:17
*** cshen has joined #openstack-ansible07:54
*** cp- has quit IRC08:00
*** tosky has joined #openstack-ansible08:02
*** cshen has quit IRC08:06
*** cshen has joined #openstack-ansible08:09
*** cp- has joined #openstack-ansible08:15
taccoran into this issue https://bugs.launchpad.net/openstack-ansible/+bug/180734608:30
openstackLaunchpad bug 1807346 in openstack-ansible "[heat] Installations fails during Update Q->R" [Undecided,New] - Assigned to Guilherme Steinmuller Pimentel (guilhermesp)08:30
taccois there alreay a workaround for this?08:30
taccoor am i doing something wrong?08:31
cshentacco: according to the bug ticket, do you have multi domain?08:32
*** rpittau|afk is now known as rpittau08:34
taccoyes08:34
taccodefault and ldap for user auth08:34
taccobut default is still untouched08:34
cshensince the ticket is not closed, it seems that no workaround yet.08:35
*** ivve has joined #openstack-ansible08:35
taccowas working when i initialy deployed without multidomain and added multidomain later.08:35
cshenwe only have 1 domain, so I can't judge too much.08:35
taccook. Thanks so we have to deploy initialy without multidomain and then one more shot wiht multidomain08:35
cshenprobably single domain works, multi domain not.08:36
masterpeI disabled the task by commenting it out08:37
taccomasterpe: thanks.. will see what i can do and report back what my solution was.. because i don't want to hacky fix/patch the ansible roles..08:38
*** sshnaidm|off is now known as sshnaidm08:41
*** DanyC has joined #openstack-ansible08:46
noonedeadpunkmornings08:48
*** DanyC has quit IRC08:48
*** DanyC has joined #openstack-ansible08:48
noonedeadpunktacco: so for some reason bug is assigned to guilhermesp - maybe he will provide you with bug status...08:49
*** DanyC has quit IRC08:51
*** shyamb has joined #openstack-ansible08:59
*** DanyC has joined #openstack-ansible09:02
taccothanks.. i've tryed "heat_service_in_ldap=True" to skip the jobs..09:16
taccoworks..09:16
taccobut my question is now.. what user is required in ldap.. or what heat expects from ldap09:16
taccowill figure out.. :D09:17
*** mugsie has quit IRC09:19
taccoguilhermesp: *dodge* :D09:19
*** mugsie has joined #openstack-ansible09:21
*** yolanda has quit IRC09:28
*** yolanda__ has joined #openstack-ansible09:28
*** cosmicsound has joined #openstack-ansible09:30
jrossertacco: the only way to fix that is to patch your ansible installation (or use a forked version of ansible). You need the two patches mentioned in the bug report https://github.com/ansible/ansible/pull/42899/files and https://github.com/ansible/ansible/pull/42913/files09:31
jrosserunfortunately it is very difficult to get fixes merged into ansible for versions which are as "old" as the ones we use in Q & R09:32
taccoi see so user-domain was missing on the os_user module right?09:43
cshenand also openstack.py09:44
taccoi see. will patch this. :D09:44
taccoand os_user_role.py m(09:44
taccoanyway thanks jrosser for pointing me there.09:44
jrosserit is horrible sadly09:45
jrossertacco: depending on how you want to do it, there are instructions here about how to have a fork of ansible https://docs.openstack.org/openstack-ansible/latest/user/source-overrides/index.html09:46
jrosserif you patch it in place there is a risk that doing boostrap-ansible.sh will undo your changes09:46
jrossernoonedeadpunk: there is something wierd going on with horizon09:47
noonedeadpunkwhat do you mean?09:48
jrossersee what pymysql version it wants https://8b7bc396c22f90223dfa-906a8885044ac3b62edc724702c6de01.ssl.cf1.rackcdn.com/699123/1/check/openstack-ansible-deploy-aio_lxc-debian-buster/004e481/logs/ara-report/index.html09:48
jrosservs this https://github.com/openstack/requirements/blob/master/upper-constraints.txt#L49109:49
noonedeadpunklooks like they've upgraded django...09:50
*** jbadiapa has joined #openstack-ansible09:52
*** jftalta has joined #openstack-ansible09:55
*** Jeffrey4l has joined #openstack-ansible10:05
*** django has joined #openstack-ansible10:05
*** jmccrory has joined #openstack-ansible10:05
*** mbuil has joined #openstack-ansible10:06
*** stingrayza has joined #openstack-ansible10:07
*** szaher has joined #openstack-ansible10:07
*** irclogbot_1 has quit IRC10:08
*** irclogbot_1 has joined #openstack-ansible10:09
*** miloa has joined #openstack-ansible10:20
*** jcosmao has joined #openstack-ansible10:25
noonedeadpunkjrosser: ok, so I see the severl outcomes with horizon10:25
noonedeadpunkFirst one is to drop mysql dependency. It's used only as a session storage, so we can probably switch just to mamcached or cookies https://docs.openstack.org/horizon/latest/admin/sessions.html10:25
noonedeadpunkThe other one is to update u-c and motivate it by requirement of fresh django......10:26
jrosseri do wonder if horizon-manage meddles directly in the database?10:27
noonedeadpunkactually we are the ones who distribute horizon-manage.... https://opendev.org/openstack/openstack-ansible-os_horizon/src/branch/master/templates/horizon-manage.py.j210:27
noonedeadpunkso yeah...10:28
*** DanyC has quit IRC10:28
noonedeadpunkHorizon's original manage py just don't use mysqlclient https://opendev.org/openstack/horizon/src/branch/master/manage.py10:28
jrosseroh!10:29
*** DanyC has joined #openstack-ansible10:30
noonedeadpunkSo if we move to cookies or memcached backends we can drop out solid part of the role10:31
*** aspiers has joined #openstack-ansible10:33
*** cshen has quit IRC10:33
noonedeadpunkdownside of memcahed that it eventually will logout user if memcahced server goes down. Downsides of cookies are described in docs... So currenty our deploy is pretty balanced...10:33
*** shyamb has quit IRC10:34
noonedeadpunkoh....10:38
noonedeadpunkjrosser: I think we can try installing https://pypi.org/project/mysqlclient/10:40
noonedeadpunkinstead of PyMySQL10:40
noonedeadpunkbut for distro install it looks like it has no packages...10:42
taccojrosser: patch was already upstram in my ansible.. so didn't worked for me10:43
noonedeadpunkand it's missing in u-c10:43
*** jftalta has quit IRC11:06
*** shyamb has joined #openstack-ansible11:10
*** rpittau is now known as rpittau|bbl11:22
*** tbarron has joined #openstack-ansible11:41
jrossernoonedeadpunk: will memcached be an issue if we move to loadbalancer for that? i.e session will be stored only in 1 memcached out of N12:00
noonedeadpunkI think this will be an issue anyway, as afaik it's stored on 1 instance now as well12:02
noonedeadpunkand I think we won't be changing default behaviour due to mnaser comment12:03
noonedeadpunkbut I'll document an option how to use haproxy12:03
noonedeadpunkofc we can try and adopt mysqlclient package isntead12:04
*** cshen has joined #openstack-ansible12:08
*** dmsimard has joined #openstack-ansible12:12
*** ansmith has quit IRC12:17
*** ivve has quit IRC12:22
*** pcaruana has quit IRC12:27
*** DanyC has quit IRC12:30
*** ivve has joined #openstack-ansible12:30
*** shyamb has quit IRC12:30
*** pcaruana has joined #openstack-ansible12:33
*** cshen has quit IRC12:35
*** cshen has joined #openstack-ansible12:47
*** DanyC has joined #openstack-ansible12:51
*** goldyfruit has joined #openstack-ansible13:09
*** rpittau|bbl is now known as rpittau13:09
*** rgogunskiy has quit IRC13:09
*** ansmith has joined #openstack-ansible13:25
*** goldyfruit has quit IRC13:37
*** dave-mccowan has joined #openstack-ansible14:09
guilhermesptacco: noonedeadpunk unfortunately i havent had a chance to look at it14:20
*** cshen_ has joined #openstack-ansible14:20
*** cshen has quit IRC14:23
guilhermespthanks jrosser to point out the fixes that are present in the bug report14:28
jrosseri really am not sure there is anything we can do - unless those bugfixes are now merged to ansible and we bump the version?14:43
jrosserthat would be easy14:44
*** rgogunskiy has joined #openstack-ansible14:44
*** rgogunskiy has quit IRC14:49
*** chkumar|rover is now known as ignoreirc14:57
*** ignoreirc is now known as chkumar|rover14:58
*** KeithMnemonic has joined #openstack-ansible15:02
*** schwicht has quit IRC15:06
cjloaderhi15:07
*** schwicht has joined #openstack-ansible15:07
cjloaderlooks like we missed release dealine?15:07
*** cshen_ has quit IRC15:15
noonedeadpunkyeah, we did15:20
*** jgwentworth is now known as melwitt15:21
*** ivve has quit IRC15:21
*** goldyfruit has joined #openstack-ansible15:23
guilhermespjrosser: they seemed to be merged?15:32
jrosserhmm yes, but if they are in ansible we use for R though?15:34
jrosseri think it's really an issue of needing those patches in this version https://github.com/openstack/openstack-ansible/blob/stable/rocky/scripts/bootstrap-ansible.sh#L2615:35
cjloaderare we not going to release this cycle now?15:38
taccoguilhermesp: noonedeadpunk my workaround for now was to comment line 145,146 and 147 from os_heat/task/heat_service_setup.yml15:40
*** goldyfruit has quit IRC15:45
taccobut i guess this means all ldap users are not able to use heat.. if i got it right.15:51
jrossertacco: is it right you think those patches are merged into ansible 2.5.15?16:08
*** chkumar|rover is now known as raukadah16:14
*** goldyfruit has joined #openstack-ansible16:20
*** gregwork has joined #openstack-ansible16:21
cjloaderare we still going to be able to release? noonedeadpunk jrosser16:24
*** gyee has joined #openstack-ansible16:28
*** miloa has quit IRC16:37
jrossercjloader: the patch is in the releases repo but needs +W https://review.opendev.org/#/c/698730/16:46
cjloaderah16:46
jrosserthats for RC4 and then we need someone to merge the actual release patch16:47
* jrosser asks in #openstack-release16:48
jrosserbingo, there it goes16:49
cjloaderwoohoo16:50
cjloaderjrosser: we both rechecked17:03
cjloaderlol17:03
*** udesale has joined #openstack-ansible17:14
*** goldyfruit has quit IRC17:17
noonedeadpunkcores, we're about to release train, kindly put your votes https://review.opendev.org/#/c/698736/17:33
noonedeadpunkcjloader: jrosser thanks for re-checking it:)17:33
admin0are there plans to set a release as LTS and keep up from upgrading for say 2 years or something17:35
noonedeadpunkadmin0: I think it's worth asking tc...17:42
noonedeadpunkWe actually just following global releasing cycles17:42
jrosserisnt that 2 year cycle more of a $GIANT-VENDOR approach?17:43
*** goldyfruit has joined #openstack-ansible17:43
jrosser^ substitute hat color to suit17:43
admin0my rocky to stein is failing .. so asking :D17:43
admin0if it works, i will shut up again :)17:43
jrosseradmin0: well it should work - is something specific broken?17:46
*** goldyfruit has quit IRC17:58
*** DanyC has quit IRC17:58
*** rgogunskiy has joined #openstack-ansible18:00
*** rgogunskiy has quit IRC18:05
*** sshnaidm is now known as sshnaidm|afk18:26
*** jcosmao has left #openstack-ansible18:28
admin0jrosser, this one ..  galera-cluster-rolling-restart.yml  fails with Wait for mariadb port 3306 to be available18:52
admin0am i supposed to manually restart galera, or what is not clear18:53
jrosserI guess if 3306 is to be available then galera needs to be stopped, and isn’t?18:53
admin0am I to manually stop galera in that container ?18:54
admin0because here, the script just waits and waits and does not proceed18:54
jrosserthe upgrade script?18:54
*** rpittau is now known as rpittau|afk18:55
jrosserfwiw I don’t use the upgrade shell script, always step through he manual steps in the documentation18:56
jrosserand check very carefully what’s happening particularly with galera and rabbitmq18:56
jrosserpersonally I would run whichever galera playbook it is as the docs describe, and if it fails try to debug and look at why the service hasn’t stopped, if that’s what a previous task tried to do18:58
admin0jrosser, it is in the upgrading manually section18:58
admin02nd last step18:58
admin0fatal: [c3v_galera_container-300600e4 -> 172.29.239.164]: FAILED! => {"changed": false, "elapsed": 300, "msg": "Timeout when waiting for 172.29.237.49:3306"} == is how it ends18:59
jrosserright I read the code - just on my phone so this is awkward19:02
jrosserbut it’s looking for 3306 to be up and running and accepting connections19:02
jrosserso the answer will be in the galera log in that container about why it’s not restarted19:02
admin0[c3v_galera_container-300600e4 -> 172.29.239.164 == this 164 is my utility container19:25
admin0strange19:25
admin0in my inventory management, it does not list galeray in that IP19:25
admin0in /etc/hosts of deploy also, its listed as utility19:26
jrosseryes the ansible delegates the task there19:28
admin0oh19:28
jrossersee it fails to connect to 172.29.237.4919:29
jrosserthe utility container is guaranteed to have tcp connectivity to port 3306 of the galera host19:30
jrosserthat’s not necessarily the case for the deployment host, hence the delegate19:30
admin0ok19:30
admin0looks like mariadb crashed in this node19:31
*** gshippey has quit IRC19:31
admin0https://pastebin.com/sediTxNP19:31
mnaserjrosser: have you had a single successful rotating restart of rabbitmq .. ever?19:32
admin0it failed in that playbook with always this error .. so i think nope :D19:32
mnaserno matter what i do, if we ever restart any rabbitmq, it just goes back up but the whole thing is messed up and messages dont actually clear through19:33
admin0i have mariadb running on 2 serves, but failed on this one19:33
admin0https://pastebin.com/Wq1e9amS -- looks like the last mariadb is also stuck on something19:34
jrossermnaser: rabbitmq has been a total car crash every upgrade we have done19:38
jrosserlike massive outage19:38
mnaserjrosser: EVER if we manaully restart it, nothing19:39
mnaseri upgraded to latest which is 3.8.219:39
mnasernothing, still stuck, the only resolution, kill all vhosts, recreate them, delete all queues, restart all services19:39
mnaseras we're using k8s for some control plane components now, im very keen on running a one-replica-non-clustered rabbitmq with no persistence behind it19:39
mnaserso that a simple restart gets us a clean state which should eliminate all that stuff19:40
mnaseri even tried resetting the WHOLE entire cluster and rebuilding it from scratch.  still not sending messages.19:40
jrosserI think OSA has wrong defaults somewhere - still notifications piling up I think19:40
*** rgogunskiy has joined #openstack-ansible19:42
jrossermnaser: from memory on our S>T upgrade the whole rabbit cluster went nuts 100% on all cores19:44
mnaserjrosser: no the notifications thing is fixed19:44
mnaserthe piling up of messages i sbecause when you restart agents19:44
mnasertheir existing queues are dropped and they grab new ones for fanouts/etc19:44
mnaserso their old ones which still are part of the exchange eventually time out after 30 minutes or so19:45
*** jbadiapa has quit IRC19:46
jrosserbasically we don’t have good enough monitoring of the rabbit queues and that let to a horrid upgrade time19:47
mnaserjrosser: if you restart a single rabbitmq node right now19:47
mnaseri guarantee you will be having a bad day19:48
mnaserand we're not the only ones hitting it19:48
mnasersome folks from blizzard are too19:48
*** udesale has quit IRC19:52
jrossermnaser: i've also seen the inverse where services get wedged up and are broken wne you restart rabbitmq19:54
jrosserbut they look like they are running ok19:54
jrosseri think keystone did that to us on the S>T19:54
*** rgogunskiy has quit IRC19:55
*** gyee has quit IRC19:59
mnaserjrosser: yep exactly that :\20:07
admin0https://pastebin.com/Wq1e9amS -- looks like the last mariadb is also stuck on something20:07
admin0looks like one of my galera db won't start no matter what .. is it safe to lxc-destroy it and re-create it again ?20:10
admin0during that last part of the upgrade process20:10
*** rgogunskiy has joined #openstack-ansible20:17
*** gregwork has quit IRC20:34
*** rgogunskiy has quit IRC20:36
*** gyee has joined #openstack-ansible20:40
*** ansmith has quit IRC21:30
*** elenalindq has quit IRC21:34
*** cyberik has joined #openstack-ansible21:38
*** rgogunskiy has joined #openstack-ansible21:38
*** goldyfruit has joined #openstack-ansible21:44
*** rgogunskiy has quit IRC21:47
*** rgogunskiy has joined #openstack-ansible21:47
*** pcaruana has quit IRC21:55
*** ansmith has joined #openstack-ansible22:16
*** rgogunskiy has quit IRC22:35
*** openstackgerrit has joined #openstack-ansible23:18
openstackgerritMerged openstack/openstack-ansible-os_neutron stable/queens: l3 agent on network_hosts do dvr_snat, anywhere else dvr  https://review.opendev.org/69887823:18
*** cosmicsound has left #openstack-ansible23:20
*** tosky has quit IRC23:43

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!