Monday, 2019-11-25

*** rcernin has quit IRC00:17
*** rcernin has joined #openstack-containers00:17
flwangbrtknr: :) thank you00:37
*** vesper11 has quit IRC03:45
*** vesper11 has joined #openstack-containers03:47
*** ricolin has joined #openstack-containers03:54
*** ykarel has joined #openstack-containers04:52
*** udesale has joined #openstack-containers05:27
*** namrata has joined #openstack-containers07:04
*** rcernin has quit IRC07:09
*** CosmicSound4 has joined #openstack-containers07:14
*** CosmicSound has quit IRC07:15
*** lpetrut has joined #openstack-containers07:16
*** CosmicSound4 has quit IRC07:19
*** CosmicSound has joined #openstack-containers07:24
*** CosmicSound is now known as cosmicsound07:35
*** cosmicsound has quit IRC07:49
*** cosmicsound has joined #openstack-containers07:54
brtknrflwang: any luck?07:59
*** elenalindq has joined #openstack-containers08:00
namrataHi,brtknr this issue was discussed on friday and you asked for kubectl get pods --all-namepsaceshttp://paste.openstack.org/show/786643/The issue which we are facing "When we spin up a magnum Kubernetes cluster.We see all the pods in pending state http://paste.openstack.org/show/786532/ as which means that the cluster has been unable to select a08:05
namrataminion node for them After removing NoSchedule taint from the minion nodes manually we get it working The solution we got online  for pods in pending. https://ask.openstack.org/en/question/120442/magnum-kubernetes-noschedule-taint/ the reference link says to add [trust]cluster_user_trust = true to magnum.conf  but there is an OSA variable which08:05
namratasets  magnum_cluster_user_trust to true. And this variable is set to true by defaulthttps://opendev.org/openstack/openstack-ansible-os_magnum/commit/aceaa3e5f174af1e40060440c025f49cd1875e0b and we are not overriding it.Can I get help how to get it working without manually untainting nodes."08:05
namratacan you help me with this08:05
namratabrtknr http://paste.openstack.org/show/786643/08:23
*** ykarel is now known as ykarel|lunch08:33
*** namrata has quit IRC08:59
*** ricolin has quit IRC09:16
openstackgerritBharat Kunwar proposed openstack/magnum master: Bump up prometheus operator chart to 8.2.2  https://review.opendev.org/69544309:30
*** namrata has joined #openstack-containers09:33
*** ykarel|lunch is now known as ykarel09:37
*** namrata has quit IRC09:47
*** namrata has joined #openstack-containers10:00
brtknrnamrata: can you show me the output of `kubectl get all -A` and the content of /etc/sysconfig/heat-params10:11
brtknri can help you but a lot of the issues you are reporting have been fixed in stein/train/master so again, i will reiterate running a more recent release of magnum... it has a very loose coupling with other openstack services so it is not going to break things... we are running magnum "train 9.1.0" release with openstack "rocky" without any issues...10:13
brtknri can help you but a lot of the issues you are reporting have been fixed in stein/train/master so again, i will reiterate running a more recent release of magnum... it has a very loose coupling with other openstack services so it is UNLIKELY to break things... we are running magnum "train 9.1.0" release with openstack "rocky" without any issues...10:14
brtknr...after having lots of issues running the latest version of kubernetes on rocky10:17
namratabrtknr `kubectl get all -A` http://paste.openstack.org/show/786651/   and content of  /etc/sysconfig/heat-params on minion  http://paste.openstack.org/show/786652/10:18
namratabrtknr thanks for the suggestion we are discussing on our plan to upgrade our regions but until then magnum rocky not working is affecting many things.10:21
brtknrHow do you deploy openstack? Openstack ansible?10:23
brtknrnamrata:10:23
namratayes openstack ansibe10:23
namrata*ansible10:24
brtknrnamrata: ask in #openstack-ansible channel how to upgrade magnum train, that will be the quickest way to fix your problem if you need up to date kubernetes10:25
brtknrIf you are going to wait for a new rocky release, you will be waiting for a long time as none of the maintainers are using magnum rocky in prod10:26
brtknrwe are running stein/train on all customer sites now10:26
brtknre.g I don’t know how long it takes with OSA, but with kolla, upgrading magnum to train takes 10 mins max if using upstream image10:28
namratabrtknr i underdstand your point but these are the issues of broken images as for me v 1.10.11-1 works but other like v 1.14.8 and v1.13.12  dosent work10:30
brtknrI suggest using the default kube_tag for rocky as only those have the guarantee of being supported. Kubernetes has fast dev cycle, we try to keep up to support latest kube_tag.. as a result, magnum releases are more tightly coupled  with kubernetes version than openstack releases10:33
brtknrIf you don’t want to run latest magnum, I suggest using something like kubespray which will allow you to run the latest version of k8s10:35
namratabrtknr thanks for the information but clearly default kube_tag for rocky is exposed to billion laughs which is not ideal.10:36
brtknrnamrata: looking at your paste, i do not see openstack-cloud-controller-manager daemonset running...10:42
brtknri'd suggest trying with cloud_provider_enabled=false label but I cant think of anything else to suggest beyond that10:42
brtknrelenalindq: ^10:47
elenalindqbrtknr Roger, that.10:47
brtknrnamrata: elenalindq: I just asked the OSA folks about upgrading magnum to train on openstack deployed using OSA, they say it can be done:10:49
brtknr         brtknr | noonedeadpunk: we are running magnum train with openstack rocky deployed using kolla-ansible... was10:49
brtknr                | wondering if the same could be done with OSA10:49
brtknr  noonedeadpunk | yeah, you can define some exact version of magnum or even install it from your own repo10:49
brtknr         brtknr | magnum has loose dependency with other openstack services.... magnum train using heat templates from10:49
brtknr                | queens10:49
brtknrnoonedeadpunk | In that case it may theoretically work...10:50
*** pcaruana has joined #openstack-containers11:14
*** namrata has quit IRC11:18
*** namrata has joined #openstack-containers11:25
namratabrtknr setting `cloud_provider_enabled=false` gives me a failed cluster with status reason `"cinder" volume driver needs "cloud_provider_enabled" label to be true or unset.`11:27
namrataso its not working11:27
brtknrnamrata: so try setting volume size to 0?11:28
brtknrits because you have volume driver set to "cinder"11:29
brtknrhttps://github.com/openstack/magnum/blob/stable/rocky/magnum/drivers/heat/k8s_fedora_template_def.py#L10011:29
*** openstack has joined #openstack-containers11:53
*** ChanServ sets mode: +o openstack11:53
brtknrnamrata: not docker volume size, the volume_driver12:10
brtknrnamrata: not docker volume size, the volume_driver in cluster template12:10
namrataokay12:11
*** dave-mccowan has joined #openstack-containers12:30
*** dave-mccowan has quit IRC12:35
namratabrtknr I get a CREATE_COMPLETE cluster but no pods http://paste.openstack.org/show/786658/12:37
*** namrata has quit IRC12:40
*** namrata has joined #openstack-containers12:41
*** lpetrut has joined #openstack-containers12:53
*** goldyfruit has quit IRC13:11
*** goldyfruit has joined #openstack-containers13:14
*** goldyfruit has quit IRC13:19
*** KeithMnemonic has joined #openstack-containers13:54
*** ykarel_ is now known as ykarel14:27
*** KeithMnemonic has quit IRC14:33
*** goldyfruit has joined #openstack-containers14:38
*** spiette has joined #openstack-containers14:53
*** goldyfruit has quit IRC14:53
*** goldyfruit has joined #openstack-containers14:55
*** goldyfruit_ has joined #openstack-containers14:59
*** ykarel_ has joined #openstack-containers15:01
*** goldyfruit has quit IRC15:01
*** ykarel has quit IRC15:04
*** namrata has quit IRC15:19
*** namrata has joined #openstack-containers15:45
*** ykarel_ is now known as ykarel|away15:49
*** namrata has quit IRC16:04
*** openstackstatus has joined #openstack-containers16:14
*** ChanServ sets mode: +v openstackstatus16:14
*** elenalindq has quit IRC16:27
*** goldyfruit___ has joined #openstack-containers16:33
*** goldyfruit_ has quit IRC16:36
*** elenalindq has joined #openstack-containers16:42
*** lpetrut has quit IRC16:51
*** goldyfruit_ has joined #openstack-containers16:55
*** goldyfruit___ has quit IRC16:57
*** udesale has quit IRC16:57
*** FlorianFa has quit IRC17:29
andreinHello, I've just upgraded our openstack deployment from stein to train and I think I (partially) broke Magnum. I'm getting the following error when updating cluster templates:18:20
andreinhttps://www.irccloud.com/pastebin/TQ1bAz6A/18:20
andreinlooking at the history, it looks like the /hidden attribute is non-removable since ~10 months ago, could this actually be a horizon bug?18:23
*** ykarel|away has quit IRC18:42
*** colin- has joined #openstack-containers20:24
flwangandrein: still round? the hidden issue is a known one20:37
flwangandrein: can you login into your db can check the value of the 'hidden' for templates table?20:38
flwangbrtknr: no, no luck for the api issue, i'm still working on that20:40
andreinflwang: checking now20:40
andreinflwang: all my cluster templates show "hidden: 0"20:43
flwangok, then could you pls clarify what's your current probnlem?20:44
flwangcan't update template?20:45
andreinyes20:45
andreinit used to work on stein, though20:45
flwangwhat's your current version of magnum client?20:46
andreinI'm using the horizon dashboard, was just installing magnum client to try it this way20:47
andreinit works with python-magnumclient==2.15.020:48
flwangandrein: just to be clear, so now the CLI works for you, but the dashboard still doesn't work?20:50
andreinyes, that's correct20:50
flwangandrein: did you set the microversion in local_settings.py for horizon?20:52
andreinthat's a good question, I don't think so. I deployed this with kolla-ansible.20:53
andreinI'll check the horizon container in a sec20:53
andreinwhat am I looking for in local_settings.py?20:55
flwangOPENSTACK_API_VERSIONS20:59
andreinOPENSTACK_API_VERSIONS = {"identity": 3}21:01
andreinthat's all I have there21:01
*** elenalindq has quit IRC21:02
andreinflwang: the service catalog has "http://172.16.0.250:9511/v1" listed as the magnum endpoint21:03
flwangOPENSTACK_API_VERSIONS = {"identity": 3, "container-infra": "latest"}  try this21:04
andreinflwang: this might be complicated to set up, that variable is not overridable in kolla-ansible21:08
flwangandrein: ok, then i will leave it to you :) that's the thing you need to check and test21:10
andreinflwang: actually, it's overridable, deploying it now21:12
andreinflwang: still no luck, same error. The response from horizon is "HTTPResponse instance has no attribute 'headers'", status code 500. no longer see anything in the magnum logs.21:22
andreinso I think it's not getting past horizon21:22
flwangdo you mean the 500 error is in horizon or magnum?21:23
andreinhorizon throws 500, magnum used to throw 400 (as I posted above)21:24
andreinnow I don't see anything at all in magnum's log21:24
flwangok, did you run 'python manage.py collectstatic' and then compress ?21:24
flwangyou shouldn't need this, but with the new change, you shouldn't see any error caused by the new config21:25
flwangi'm not familiar with your horizon deployment way, but at least you need to restart the horizon21:26
andreinI deployed a new container with the new setting21:26
andreinusing kolla-ansible21:27
andreinI think I have a few stacktraces in horizon's log now21:27
flwangthen you can check your horizon log see why21:27
flwangif you can get correct response with cli, that means magnum server is working as expected at least21:28
andreinflwang: here's the exception, I can't make much sense of it unfortunately https://www.irccloud.com/pastebin/R9qpjZWz/21:29
*** rcernin has joined #openstack-containers21:31
flwangandrein: can you see the magnum client version of your horizon env?21:31
andreinflwang: the horizon container also has python-magnumclient=2.15.021:32
flwangandrein: sorry, i don't have any clue with the exception, you may have to add some break points to understand why21:35
*** pcaruana has quit IRC21:38
andreinflwang: ok, figured it out. there are actually no changes, my elasticsearch cluster was just slow. magnum still throws a 400 response with: Client-side error: '/hidden' is a mandatory attribute21:44
brtknrflwang: the api issue doesn’t exist on master branch, maybe you missed a backport to your special stein branch22:18
flwangbrtknr: good point, i will take another look22:29
flwangbrtknr: btw, you mentioned that you're using stable/train right?22:29
flwangdid you see any issue related to node group?22:29
*** KeithMnemonic has joined #openstack-containers22:41
*** cosmicsound has quit IRC23:45
*** cosmicsound has joined #openstack-containers23:45

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!