Friday, 2020-03-27

*** cshen has joined #openstack-ansible00:04
*** cshen has quit IRC00:08
*** DanyC has joined #openstack-ansible00:35
*** DanyC has quit IRC00:41
*** cshen has joined #openstack-ansible00:44
*** maharg101 has quit IRC00:46
*** maharg101 has joined #openstack-ansible00:47
*** cshen has quit IRC00:49
*** miloa has quit IRC00:51
*** macz_ has joined #openstack-ansible00:54
*** macz_ has quit IRC00:59
*** prometheanfire has joined #openstack-ansible01:02
*** DanyC has joined #openstack-ansible01:08
*** DanyC has quit IRC01:13
*** rh-jelabarre has quit IRC02:21
*** cshen has joined #openstack-ansible02:45
*** cshen has quit IRC02:50
*** gyee has quit IRC02:53
*** cshen has joined #openstack-ansible03:01
*** gkadam has joined #openstack-ansible03:03
*** cshen has quit IRC03:05
*** gkadam_ has joined #openstack-ansible03:16
*** gkadam has quit IRC03:16
*** gkadam_ has quit IRC03:17
*** mloza has quit IRC04:45
*** cshen has joined #openstack-ansible05:02
*** udesale has joined #openstack-ansible05:06
*** cshen has quit IRC05:06
*** cshen has joined #openstack-ansible05:17
*** cshen has quit IRC05:22
*** evrardjp has quit IRC05:36
*** evrardjp has joined #openstack-ansible05:36
*** cshen has joined #openstack-ansible06:00
*** cshen has quit IRC06:05
*** cshen has joined #openstack-ansible07:00
*** shyamb has joined #openstack-ansible07:14
*** miloa has joined #openstack-ansible07:29
*** shyamb has quit IRC07:36
*** DanyC has joined #openstack-ansible08:00
*** rpittau|afk is now known as rpittau08:23
*** shyamb has joined #openstack-ansible08:30
*** wpp has quit IRC08:31
*** shyamb has quit IRC08:31
*** shyamb has joined #openstack-ansible08:31
*** DanyC has quit IRC08:33
*** shyamb has quit IRC08:49
*** DanyC has joined #openstack-ansible08:57
*** wpp has joined #openstack-ansible09:00
openstackgerritChandan Kumar (raukadah) proposed openstack/ansible-role-python_venv_build master: Switch to CentOS-8 based tripleo Job  https://review.opendev.org/71536509:07
*** shyamb has joined #openstack-ansible09:09
openstackgerritChandan Kumar (raukadah) proposed openstack/ansible-config_template master: Switch to CentOS 8 based Tripleo job  https://review.opendev.org/71536709:11
openstackgerritChandan Kumar (raukadah) proposed openstack/openstack-ansible-os_tempest master: Switch to CentOS-8 based TripleO job  https://review.opendev.org/71536809:13
*** itandops has joined #openstack-ansible09:15
itandopsHi all, I'm installing openstack 20.0.1 but I get this error http://paste.openstack.org/show/791222/.the setup-hosts playbook works well but this arises at the end of setup-infrastructure.Any suggestion to solve this please ?09:21
itandopsI get this error also in 20.0.209:21
*** tosky has joined #openstack-ansible09:29
*** udesale_ has joined #openstack-ansible09:35
*** udesale has quit IRC09:38
openstackgerritChandan Kumar (raukadah) proposed openstack/openstack-ansible-os_tempest master: Added tempest_tempestconf_profile_ specific vars  https://review.opendev.org/71460109:48
*** sshnaidm|afk is now known as sshnaidm|off09:50
*** gshippey has joined #openstack-ansible10:16
*** shyamb has quit IRC10:17
*** shyamb has joined #openstack-ansible10:20
*** kopecmartin has quit IRC10:22
*** kopecmartin has joined #openstack-ansible10:27
*** jbadiapa has joined #openstack-ansible10:41
jamesfreeman959Hello all - I had a problem last night deploying openstack-ansible v19.0.11 - this is a completely fresh install working from this example: https://docs.openstack.org/openstack-ansible/stein/user/ceph/full-deploy.html10:47
jamesfreeman959I hit a failure running setup-openstack.yml10:47
jamesfreeman959TASK [os_ceilometer : Initialize Gnocchi database by creating ceilometer resources] *******************************************************************************************************************************************************************************************10:48
jamesfreeman959 [WARNING]: Module remote_tmp /var/lib/ceilometer/.ansible/tmp did not exist and was created with a mode of 0700, this may cause issues when running as another user. To avoid this, create the remote_tmp dir with the correct permissions manually10:48
jamesfreeman959fatal: [infra1_ceilometer_central_container-9d1d0923]: FAILED! => {"changed": false, "cmd": ["/openstack/venvs/ceilometer-19.0.11/bin/ceilometer-upgrade"], "delta": "0:00:14.575090", "end": "2020-03-26 21:45:09.850635", "msg": "non-zero return code", "rc": 1, "start": "2020-03-26 21:44:55.275545", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []}10:48
jamesfreeman959No immediate errors seen in either the gnocchi or the ceilometer containers on infra110:48
jamesfreeman959If anyone has any ideas on how to debug or resolve I would be grateful!10:48
*** shyamb has quit IRC10:52
*** shyamb has joined #openstack-ansible10:54
*** spatel has joined #openstack-ansible10:57
*** spatel has quit IRC11:01
*** arxcruz is now known as arxcruz|off11:19
*** jamesden_ has joined #openstack-ansible11:22
*** rpittau is now known as rpittau|bbl11:32
*** shyamb has quit IRC11:37
noonedeadpunkjamesfreeman959: have you tried running /openstack/venvs/ceilometer-19.0.11/bin/ceilometer-upgrade manually? any output?11:39
noonedeadpunkjrosser: now seems that pytest is failing for rocky.... https://zuul.opendev.org/t/openstack/build/f1f26735282d4c37a706894c6869c89e/log/job-output.txt#1001111:41
noonedeadpunkoh, I think we just get master u-c11:42
noonedeadpunkI guess you've patched that for train?11:42
noonedeadpunkfound it https://review.opendev.org/#/c/703979/11:44
*** macz_ has joined #openstack-ansible11:46
jrosseriirc we didn’t see problems on rocky at the time11:48
jrosserbut those should really be backported further I think11:48
*** macz_ has quit IRC11:51
jamesfreeman959noonedeadpunk: Please forgive me inexperience - would I run that script inside the LXC container on infra1?11:53
noonedeadpunkyes, inside ceilometer container11:54
jamesfreeman959Ok - will fire up the environment and test and get back to you. Thanks!11:54
*** mathlin_ has quit IRC11:57
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_tempest stable/rocky: Use contraints for tempest plugins  https://review.opendev.org/71540712:04
jamesfreeman959noonedeadpunk: I just powered up the VM's again, and looked in /openstack/venvs/ - it only contains: "cinder-19.0.11  neutron-19.0.11"12:08
jamesfreeman959no ceilometer directory12:08
noonedeadpunkhm. I think you;re not supposed to have neutron and cinder inside ceilometer-center lxc container12:09
noonedeadpunk*ceilometer-central12:10
jamesfreeman959noonedeadpunk: ah - this is on the bare host. I powered the nodes down overnight. Powered them up this morning but no containers have come up. Do I need to run the playbook again for them to come up?12:11
noonedeadpunkyou may have them in case of metal build...12:11
jamesfreeman959last night there were LXC containers running12:11
noonedeadpunkwhat does lxc-ls say?12:13
jamesfreeman959It returns no output12:13
noonedeadpunkactually lxc containers should spawn up with node12:13
noonedeadpunkdo you run it with root privileges?12:14
jamesfreeman959omg - I'm so sorry - clearly not enough coffee this morning12:14
jamesfreeman959ok - now that I've got over that "moment" - running ceilometer-upgrade returns no output, and the exit code is 112:15
*** rh-jelabarre has joined #openstack-ansible12:17
noonedeadpunkok. can you enter utility container, source openrc, and run openstack endpoint list --service metric ?12:18
jamesfreeman959It hung for a while, and now has returned "Gateway Timeout (HTTP 504)"12:21
noonedeadpunkhum12:22
noonedeadpunkI'd say you might have issues with keystone...12:23
noonedeadpunkat the moment, which might be result of reboot... Is mariadb and other infra running ok?12:24
noonedeadpunklike memcached, rabbitmq and stuff12:24
jamesfreeman959mariadb is down completely on all 3 nodes12:26
noonedeadpunkI guess you'll need to repair galera then first12:28
*** DanyC has quit IRC12:34
*** DanyC has joined #openstack-ansible12:34
*** DanyC has quit IRC12:39
jamesfreeman959noonedeadpunk: Running through https://docs.openstack.org/openstack-ansible/stein/admin/maintenance-tasks.html#galera-cluster-recovery - all nodes are down and I don't have a clear one to bootstrap the cluster from. Also no backups as I'd only just got this partially built. What's would you say is my most efficient strategy to recover?12:41
*** cshen has quit IRC12:43
noonedeadpunkI'd say you should jsut select one host and start new cluster on it. After that other 2 nodes should start without issues and join12:45
noonedeadpunkliek https://mariadb.com/kb/en/getting-started-with-mariadb-galera-cluster/#bootstrapping-a-new-cluster12:46
*** tlunkw has joined #openstack-ansible12:47
*** shyamb has joined #openstack-ansible12:48
*** tlunkw has quit IRC12:48
*** tlunkw has joined #openstack-ansible12:49
*** tlunkw has quit IRC12:51
*** tlunkw has joined #openstack-ansible12:51
*** partlycloudy has quit IRC12:53
*** partlycloudy has joined #openstack-ansible12:53
*** rholloway has joined #openstack-ansible12:55
*** spatel has joined #openstack-ansible13:00
*** openstackstatus has quit IRC13:01
*** openstack has joined #openstack-ansible13:05
*** ChanServ sets mode: +o openstack13:05
jamesfreeman959noonedeadpunk: Ok I backtracked - got the Galera cluster fixed and running. Now "openstack endpoint list --service metric" looks sane in the utility container I think13:05
jamesfreeman959+----------------------------------+-----------+--------------+--------------+---------+-----------+-----------------------------------+13:06
jamesfreeman959| ID                               | Region    | Service Name | Service Type | Enabled | Interface | URL                               |13:06
jamesfreeman959+----------------------------------+-----------+--------------+--------------+---------+-----------+-----------------------------------+13:06
jamesfreeman959| 568b598751694a26a73d4b4786f4bb43 | RegionOne | gnocchi      | metric       | True    | admin     | http://172.29.236.9:8041          |13:06
jamesfreeman959| 7ec4b935131c4a1ba9b898a93ceece5d | RegionOne | gnocchi      | metric       | True    | public    | http://openstack.example.org:8041 |13:06
jamesfreeman959| 80e0cf1cbf0f4cbabf86db15b2037a04 | RegionOne | gnocchi      | metric       | True    | internal  | http://172.29.236.9:8041          |13:06
jamesfreeman959+----------------------------------+-----------+--------------+--------------+---------+-----------+-----------------------------------+13:06
noonedeadpunkjamesfreeman959: please use paste.openstack.org :p13:06
jamesfreeman959Sorry - and thanks13:06
noonedeadpunkbut I got the idea13:06
noonedeadpunkSo does ceilometer-upgrade still silently fails with no output?13:07
jamesfreeman959Yes, and exit code is 113:07
*** DanyC has joined #openstack-ansible13:07
noonedeadpunkdoes `openstack metric resource-type list` have any types listed?13:08
noonedeadpunk(that's what ceilometer-upgrade should populate)13:09
jamesfreeman959ok - that doesn't look so good: /openstack/venvs/utility-19.0.11/lib/python2.7/site-packages/ujson.so: undefined symbol: Buffer_AppendShortHexUnchecked13:10
*** rpittau|bbl is now known as rpittau13:10
noonedeadpunkjamesfreeman959: BTW, can ceilometer container resolve openstack.example.org?13:10
jamesfreeman959Just checked - yes it can - it corresponds to the VIP I set up in the inventory13:11
noonedeadpunkhm, does `gnocchi resource-type list` results in the same error?13:12
jamesfreeman959yes - same error13:13
*** cshen has joined #openstack-ansible13:14
noonedeadpunkHm. WHat about TOKEN=$(openstack token issue -c id -f value); curl http://172.20.0.9:8041/v1/resource_type -H "X-Auth-Token: $TOKEN"13:15
noonedeadpunkwait13:16
noonedeadpunkTOKEN=$(openstack token issue -c id -f value); curl http://172.29.236.9:8041/v1/resource_type -H "X-Auth-Token: $TOKEN"13:16
*** tlunkw has quit IRC13:16
*** itandops has quit IRC13:16
jamesfreeman959from the utility container, it complains that curl cannot be found13:17
jamesfreeman959hold on - will try from the host13:17
noonedeadpunkjust install curl :p13:18
jamesfreeman959[{"attributes": {}, "state": "active", "name": "generic"}]13:18
noonedeadpunkok, so this feels like gnocchi itself is ok...13:19
*** macz_ has joined #openstack-ansible13:21
*** macz_ has quit IRC13:21
noonedeadpunkI'm not sure if ceilometer-upgrade does use ujson for interaction with gnocchi or not...13:21
noonedeadpunkbut ujson is probably broken badly http://lists.openstack.org/pipermail/openstack-discuss/2020-January/012285.html13:21
*** macz_ has joined #openstack-ansible13:21
*** shyamb has quit IRC13:22
noonedeadpunkjamesfreeman959: so let's return to the first question - was ceilometer installation intended one? Ie do you need it in your deployment or you're just following guide?13:22
noonedeadpunkjamesfreeman959: yeah, ceilometer-upgrade uses gnocchi client which is apparently broken because of ujson:(13:26
noonedeadpunkwhich is bad news actually...13:28
jamesfreeman959noonedeadpunk: so right now we want to build a reference architecture for the business on openstack-ansible - the example I'm following seemed good because it has HA storage and infrastructure nodes.13:58
jamesfreeman959ultimately we will need some monitoring/stats but I'm agnostic at this stage on how we achieve it13:58
jamesfreeman959is this a bigger issue then if bad news?13:59
*** itandops has joined #openstack-ansible14:00
itandopshi all, any suggestion about my issue ?14:01
itandopshttp://paste.openstack.org/show/791222/14:01
noonedeadpunkjamesfreeman959: so ceilometer gives you info about resource usage by instances. So this type of monitoring is pretty usefull in terms of billing I'd say14:08
noonedeadpunkso to fix this I think patching gnocchiclient is required14:08
noonedeadpunkbut gnocchi is not supported anymore...14:08
jamesfreeman959noonedeadpunk: that sounds like the kind of thing we'd need - we're a services company so those kind of metrics would be useful14:09
noonedeadpunkok, I see.14:09
jamesfreeman959is there a way forwards if gnocchi is not supported?14:09
jamesfreeman959for example if I move to train or ussuri, do they still use gnocchi?14:10
noonedeadpunkCeilometer supports several publishers - like prometheus or monasca. But I guess neither of them were tested with osa14:10
noonedeadpunkyeah we still use gnocchi by default, and don't have roles for deploying other engines, as don't have really much ppl using telemetry14:11
jamesfreeman959that's good to know14:12
noonedeadpunkANd actually gnocchi is working pretty good at the moment14:12
noonedeadpunkexcept it's client has problems...14:12
jamesfreeman959would this be a case of rolling back ujson?14:12
noonedeadpunkI guess discussion has ended up in using rapidjson instead...14:16
noonedeadpunkactually there might be a way of making ujson to work14:16
noonedeadpunkI didn't dig into that much tbh as never faced real issues with that until now14:17
noonedeadpunkthere was no ujson release in the last 4 years so didn't really get what you mean under rolling it back.14:17
jamesfreeman959ah ok - I didn't research the release history. I know often when I've had library problems, I've fixed it by reverting to an earlier version. But as the ujson is 4 years old that won't work....14:18
noonedeadpunkyeah - so the thing is that it was working ok but now it fails with some compilers14:19
noonedeadpunkout of that ML "14:19
noonedeadpunkThe original issue is that the released version of ujson is in14:19
noonedeadpunknon-spec-conforming C which may break randomly based on used compiler14:19
noonedeadpunkand linker.14:19
noonedeadpunkBut this part of code which gives you failure runs only when gnocchi is deployed. So if you select another dispatcher for ceilometer it's role should work for you14:21
jamesfreeman959this would be a manual patch on top of osa I guess?14:24
noonedeadpunkyou mean changing ceilometer dispatcher or what?14:24
jamesfreeman959yes - or I suppose to ask a broader question - how would you recommend I proceed?14:26
noonedeadpunkI guess you'll need just to use ceilometer_ceilometer_conf_overrides to set event_dispatchers and meter_dispatchers but I'd say to reffer to ceilometers docs14:26
jamesfreeman959ok got it14:27
noonedeadpunkActually I'd probably try out fixing ujson in the ceilometer and gnocchi venvs....14:28
noonedeadpunkNot sure how good this idea is though14:28
noonedeadpunkyou can actually ask telemetry folks what solution they would recommend14:30
noonedeadpunkas I guess they're more familiar with what's going on with their project...14:30
jamesfreeman959a bit of searching seems to indicate particular issues around Ubuntu 18.04, which is my build env14:31
noonedeadpunkhm, on rocky I had it working on ubuntu 18/04 though14:32
noonedeadpunkand ujson==1.3514:32
noonedeadpunkoh, wait14:33
noonedeadpunkit has new release 3 days ago?14:33
noonedeadpunkyeah, so it released 2.0 version on march 714:34
noonedeadpunkjamesfreeman959: what version do you have?:)14:34
jamesfreeman959in the ceilometer container venv, 1.3514:35
jamesfreeman959I was just reading their issue tracker - it looks like they were pushing a fix that should be in 2.014:35
noonedeadpunkthey do have 2.0.3 now14:36
noonedeadpunkso yeah - try isntalling it manually14:36
noonedeadpunkit's version is updated only on master.14:37
noonedeadpunkboth train and stein will install 1.35 by default14:37
jamesfreeman959ok - so on infra1 only, I attached to the ceilometer container, activated the venv, and upgraded ujson. ceilometer-upgrade now runs for longer - it returns no output still but exit code is now 014:40
jamesfreeman959ujson == 2.0.314:41
noonedeadpunkso TOKEN=$(openstack token issue -c id -f value); curl http://172.29.236.9:8041/v1/resource_type -H "X-Auth-Token: $TOKEN" should give you the way richer result :p14:41
jamesfreeman959lots of lovely JSON :)14:45
noonedeadpunkyeah, so I guess now role should run just fine. Unless it downgrade ujson or re-create venv :(14:48
jamesfreeman959I'll test and report back - I presume I can just run "openstack-ansible setup-openstack.yml" again?14:48
noonedeadpunkThe thing is we can't jsut change ujson version in role since we have to stick to constraints provided by https://releases.openstack.org/constraints/upper/stein14:49
jamesfreeman959that's fine - I'm happy to maintain a local patch for myself - as long as I've got written down what I need to do, I'm happy14:49
jamesfreeman959once Ussuri is released I presume this might be in there? (you mentioned it was in the master)14:50
noonedeadpunkYeah, it will be in U14:50
noonedeadpunkSo yes, you can either run setup-openstack.yml or just launch missing roles manually https://opendev.org/openstack/openstack-ansible/src/branch/master/playbooks/setup-openstack.yml#L29-L4114:51
noonedeadpunk(if you have some that you need)14:51
jamesfreeman959ok awesome - thanks so much for all your amazing help - I've learned a lot!14:51
jamesfreeman959I'll make the change and test - fingers crossed!14:51
noonedeadpunkyeah, sure, you're welcome:)14:52
noonedeadpunkactually if you see smth worth changing upsteam - you can submit a patch14:53
*** dave-mccowan has joined #openstack-ansible14:53
jamesfreeman959will do14:54
itandopsany feedback please ? http://paste.openstack.org/show/791222/14:57
noonedeadpunkitandops: is it rocky?14:58
noonedeadpunkI guess no since it's using python_venv_build role14:59
itandopsnoonedeadpunk: I get this error in both Stein and Train15:01
noonedeadpunkdo you have /root/.pip/pip.conf in container?15:01
chandankumarnoonedeadpunk, jrosser https://review.opendev.org/715365 and https://review.opendev.org/71536815:04
itandopsI check utility and keystone containers no one container pip.conf file15:04
*** Open10K8S has joined #openstack-ansible15:04
jrosseryou could go in the utility container and try to curl the url that fails15:05
noonedeadpunkjrosser: what do we still store on repo_server for pip? wheels?15:06
jrosseri think so yes15:08
itandopsjrosser I don't understand which link do you want me to curl ?15:09
noonedeadpunkI guess http://172.29.236.11:8181 at least15:13
noonedeadpunkis 172.29.236.11 your VIP?15:13
ionihey guys15:13
ioniquestion15:13
ionihow do i configure aio to install and configure ceph infra15:14
ionii've seen something related to bootstrap_host_scenarios_expanded15:14
ionibut i don't know how to enable scenarios15:14
noonedeadpunk./scripts/gate-check-commit.sh aio_lxc_ceph15:15
ionii want have an aio that has cinder+ceph15:15
noonedeadpunkwill deploy aio in lxc containers with ceph15:15
noonedeadpunkit should do that15:16
ioniafter running boostrap aio?15:16
noonedeadpunkioni: isntead of everythng. Just clone repo and run this :)15:16
noonedeadpunk(run as root)15:16
noonedeadpunkand kinda ensure that you are able to login as root15:18
*** udesale_ has quit IRC15:20
ioninoonedeadpunk, ok, i mostly want nova, neutron, cinder and ceph15:20
noonedeadpunkit will install all of it + horizon iirc15:21
ioninoonedeadpunk, cool15:22
noonedeadpunkoh, actually it's probably better to run  `./scripts/gate-check-commit.sh aio_lxc_ceph deploy source ` to be more specific15:22
ioni{% if 'octavia' in bootstrap_host_scenarios_expanded %}15:23
ioniwhat about this?15:23
ionii think i got it15:23
ionihttps://docs.openstack.org/openstack-ansible/latest/user/aio/quickstart.html15:23
noonedeadpunkif you want octavia in addition just add _octavia to scenario ie aio_lxc_ceph_octavia15:23
ioniexport SCENARIO='aio_lxc_barbican_ceph'15:23
ionicool cool15:24
ionii got it!15:24
ionithanks15:24
noonedeadpunkioni: so that's what aio include https://opendev.org/openstack/openstack-ansible/src/branch/master/tests/roles/bootstrap-host/vars/main.yml#L2815:24
ionicool15:24
itandopsnoonedeadpunk: yes 172.29.236.11 is my VIP. but curl  http://172.29.236.11:8181 fails because no process is listening to 818115:27
itandopsnoonedeadpunk: the haproxy should be installed before the playbook setup-openstack ?15:28
noonedeadpunkitandops: so actually haproxy is supposed to listen on that port and forward requests to repo server15:28
noonedeadpunkdo you have hosts in repo_all group?15:29
*** velmeran has joined #openstack-ansible15:29
velmeranIs this a good place to ask about a newbies deployment problem?15:32
noonedeadpunkitandops: so before setup-openstack you should launch setup-hosts.yml and setup-infrastructure.yml15:32
noonedeadpunkvelmeran: yeah, go on :)15:33
noonedeadpunkin case you deploy via OSA :p15:33
velmeranI've gotten through setup-host and setup-infrastructure, no errors reported.  but on the setup-openstack section, its failing on TASK [os_keystone : Create fernet keys for Keystone]  because the keystone-manage executable is not in the directory /openstack/venvs/keystone-20.0.0/bin/15:35
velmeranI've found that folder, and there are other things in there, but no keystone-manage.  I tried reinstalling, still all good through the first two playbooks, but it seems I'm just missing that file on the setup-openstack part.15:35
itandopsnoonedeadpunk:The repo_all section into /etc/openstack_deploy/openstack_inventory.json contains: "repo_all": { "children": [ "pkg_repo"], "hosts": [] },15:36
noonedeadpunkitandops: do you have `repo-infra_hosts` defined in openstack_user_config.yml? If not try setting it to `repo-infra_hosts: *infrastructure_hosts` and re-run setup-infrastructure.yml15:38
*** gyee has joined #openstack-ansible15:40
noonedeadpunkvelmeran: what release are you trying to deploy?15:43
velmeranI've believe its the Train release, I've been following this guide: https://docs.openstack.org/project-deploy-guide/openstack-ansible/train/run-playbooks.html15:44
noonedeadpunkvelmeran: what do you have once you inside /opt/openstack-ansible run `git branch`?15:48
noonedeadpunkoh, ok, I see, its 20.0.015:49
noonedeadpunkso I'd recommend you to use 20.0.2 - in folder /opt/openstack-ansible/  run `git checkout 20.0.2`15:50
noonedeadpunkafter that run ./scripts/bootstrap-ansible.sh15:50
noonedeadpunkand run setup-openstack.yml again15:50
velmeranokay, doing those steps now15:51
noonedeadpunkbtw, what's the output of `ls -l /openstack/venvs/keystone-20.0.0/bin/ | grep keystone` ?15:52
velmerannothing in the containers folder with keystone in the name.15:54
noonedeadpunkhm....15:54
noonedeadpunkit feels like things might fail a bit earlier than you paste...15:55
velmeranjust some active, easy install, pip, python, and wheel files/executables15:55
jamesfreeman959noonedeadpunk: ok - getting closer. Re-running the setup-openstack playbook gets further (so ujson didn't get rolled back). infra1 now looks ok, but the cinder service is showing problems on infra2 and infra315:56
jamesfreeman959http://paste.openstack.org/show/791233/15:56
noonedeadpunkok, just try out 20.0.2 release first anyway15:56
velmeranokay, its running through the setup-openstack now15:56
noonedeadpunkjamesfreeman959: it doesn;t feel futher since cinder role is the way before of ceilometer...15:57
noonedeadpunkjamesfreeman959: so what's the status of cinder-api service in the failed container?15:58
jamesfreeman959noonedeadpunk: This looks a bit fatal: "--- no python application found, check your startup logs for errors ---"15:59
jamesfreeman959(from systemctl status cinder-api)15:59
noonedeadpunkso cinder api is laucnhed via uwsgi. It's config is placed in /etc/uwsgi/cinder-api.ini16:01
velmeranI might be hung on: TASK [python_venv_build : Install python packages into the venv] its been on that step a while, and target host are all not showing much activity compared to when running the rest of the steps.16:03
noonedeadpunkjamesfreeman959: so is wsgi-file location exists?16:04
noonedeadpunkvelmeran: at this step packages should be isntalled via apt/yum16:05
velmeranhumm, no yum running on any of my three host16:06
velmeranno errors from the playbook yet, just sitting on this step16:06
noonedeadpunkvelmeran: not sure I can suggest smth here.... you can try re-run that playbook16:07
velmeranok16:08
noonedeadpunkvelmeran: you can also try re-creating that container16:10
noonedeadpunkvelmeran: like destroy them with openstack-ansible playbooks/containers-lxc-destroy.yml --limit keystone_all16:11
noonedeadpunkand create again with openstack-ansible playbooks/containers-deploy.yml --limit keystone_all,lxc_hosts16:11
noonedeadpunkafter that re-run os-keystone-isntall.yml playbook16:12
*** melwitt is now known as jgwentworth16:12
*** DanyC has quit IRC16:13
velmeranokay might try that.  interestingly this time, I got to the step where it checks for the fernet keys and says they exist, but then goes to the create step and fails due to missing file, but then just continued on past that step instead of droping out.16:13
*** DanyC has joined #openstack-ansible16:14
*** DanyC has joined #openstack-ansible16:15
noonedeadpunkcan you psot output to paste.openstack.org?16:15
velmeranokay, then it fails out on os_keystone : Wait for web server to complete starting16:15
velmerangoing to remove and reinstall keystone and see if that helps.16:15
noonedeadpunkyeah....16:15
velmeranreinstall failed, http://paste.openstack.org/show/791239/  looks like its failing to find yum repo16:21
noonedeadpunkvelmeran: actually it's 404 for me as well...16:25
noonedeadpunkdoes yum on bare metal host works ok?16:26
noonedeadpunkbut it seems it's already in container...16:26
velmeranyea, I removed the container again and re-ran install, this time it connected. so I think the repo's are having some outages/overloaded.16:29
velmerantrying the os-keystone-install playbook now16:30
*** cshen has quit IRC16:36
*** cshen has joined #openstack-ansible16:37
*** cshen has quit IRC16:42
jamesfreeman959noonedeadpunk: Sorry for the delay - just checked - everything looks complete. Python binary is there, uwsgi, /etc/uwsgi/cinder-api.ini16:48
jamesfreeman959all looks complete16:48
noonedeadpunkhave you tried restarting service?16:48
jamesfreeman959literally just tried - the output of systemctl status cinder-api looks better now - I'm re-running the playbook16:50
noonedeadpunkI guess this might be result of failed mariadb after all nodes being down16:50
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_tempest stable/rocky: Use contraints for tempest plugins  https://review.opendev.org/71540716:52
velmeran@noonedeadpunk I found the internal ip for my deployment wasn't pingable except locally, that was causing issue when a python script was looking for http on it so I fixed that.  but now I'm back to install steps failing due to repo's just not responding in time.16:53
jamesfreeman959noonedeadpunk: makes sense - playbook is running now16:53
jamesfreeman959noonedeadpunk: Possibly a dumb question, but is it expected that mariadb will be down on reboot? I shut all nodes down cleanly last night (no hard power off) and so I expected the database cluster to resume. Did I do something wrong?16:58
noonedeadpunkvelmeran eventually epel shoould be dropped here https://opendev.org/openstack/openstack-ansible-openstack_hosts/src/branch/master/tasks/main.yml#L6016:59
noonedeadpunkso not sure why it makes problems16:59
*** idlemind_ has quit IRC17:00
*** idlemind has joined #openstack-ansible17:00
noonedeadpunkjamesfreeman959: actually when galera lose all cluster participants it should be started with recovery iirc. So the last one shuts down the first raised up17:01
noonedeadpunkBut I actually try not to let galera be fully down - that's the point of cluster actually hehe17:01
jamesfreeman959noonedeadpunk: That's good advice - I'm used to the old corosync, pacemaker, etc., where it sorts out who will be master on boot. Galera is a bit new to me17:03
noonedeadpunkso actually galera also has marker which node was down the last, but I'd say it alnmost never work17:05
noonedeadpunkexcept I guess cases when you manually stop mysqld one by one17:07
jamesfreeman959I'm arranging a lab environment in a datacenter so that can be up 24/7 - but right now this is all on a high powered workstation so I need to shut it down from time to time17:07
jamesfreeman959However if I know what to look for, and what needs resolving on a restart then all is fine17:08
noonedeadpunkjamesfreeman959: I guess you could have just single maria deployed17:08
noonedeadpunkthat would resolve this cluster issues on startup17:08
noonedeadpunksince there would be no cluster actually)17:09
*** cshen has joined #openstack-ansible17:09
jamesfreeman959noonedeadpunk: That's a good plan - my brief was to build a scale model of what we'd build in production, so I built the cluster17:09
jamesfreeman959I made work difficult for myself! :-D17:09
*** pcaruana has quit IRC17:10
openstackgerritMerged openstack/ansible-role-systemd_mount master: Missing document start "---"  https://review.opendev.org/71510717:10
openstackgerritMerged openstack/ansible-role-python_venv_build master: Switch to CentOS-8 based tripleo Job  https://review.opendev.org/71536517:13
*** cshen has quit IRC17:13
velmeranhumm, since I did that update to the 20.0.2 release, all attempts to run openstack_hosts : Add requirement packages (repositories gpg keys packages, toolkits...) on keystone fail due to "One of the configured repositories failed (Unknown)".  Is there an easy way to find that command so I can see what repo it might be trying to use that is now17:16
velmeranmissing/broken?17:16
openstackgerritMerged openstack/openstack-ansible-os_tempest master: Switch to CentOS-8 based TripleO job  https://review.opendev.org/71536817:16
noonedeadpunkvelmeran: so it's here https://opendev.org/openstack/openstack-ansible-openstack_hosts/src/branch/master/tasks/openstack_hosts_configure_yum.yml#L68-L7117:17
noonedeadpunkand packages it tries to install is here https://opendev.org/openstack/openstack-ansible-openstack_hosts/src/branch/master/vars/redhat.yml#L88-L10117:17
noonedeadpunkso actually it's jsut 1 package - yum-plugin-priorities17:18
velmeranokay.  humm, is there a way to lxc-console to these and try to run that command?17:21
velmeranits asking me for a login/pass17:21
noonedeadpunklxc-attach -n container-name17:22
velmeranah17:22
*** pcaruana has joined #openstack-ansible17:23
*** miloa has quit IRC17:25
*** rholloway has quit IRC17:25
openstackgerritMerged openstack/openstack-ansible stable/rocky: Bump OSA stable/rocky  https://review.opendev.org/71492617:28
*** evrardjp has quit IRC17:36
*** evrardjp has joined #openstack-ansible17:36
mnaserebbex: welcome! :)17:44
*** itandops has quit IRC17:48
velmeranwell, I think its networking, which I thought I understood a bit. but it seems all my attempts to get things working at first might have goofed things up as none of the containers seemed to be able to ping or get dns outside.  think I'm going to start fresh with the containers, see if that helps.17:54
*** theintern_ has joined #openstack-ansible18:06
*** theintern_ has quit IRC18:09
noonedeadpunkjrosser: bump for rocky jsut landed... Do we want https://review.opendev.org/#/c/715407/ to be in EM as well? I guess it might ensure working tempest which was working only because pip_install_upper_constraints is defined in openstack-ansible-tests18:09
*** velmeran has quit IRC18:12
*** jbadiapa has quit IRC18:15
jrossernoonedeadpunk: yeah let’s include that otherwise we know it that branch will break18:15
noonedeadpunkin CI it probably won't but otherwise yeah...18:16
noonedeadpunkok let's try quickly merge it then...18:16
jrosseri +2 it18:23
*** velmeran has joined #openstack-ansible18:28
*** velmeran76 has joined #openstack-ansible18:34
*** velmeran has quit IRC18:35
*** velmeran76 has quit IRC18:36
*** DanyC has quit IRC18:38
*** velmeran has joined #openstack-ansible18:42
noonedeadpunkmnaser: guilhermesp can you kindaly vote on https://review.opendev.org/#/c/715407/ ?18:43
*** spatel has quit IRC18:44
noonedeadpunkor maybe ebbex whant to join party?:)18:44
noonedeadpunknice, thanks mnaser18:49
velmeranIs there a good guide that goes over host network bridge setup for centos7? I think my stuff breaks as it makes containers that can't reach the internet or even local gateway/dns18:52
noonedeadpunkin case you're going to use simple linuxbridges there's actually nothing special in it18:52
noonedeadpunkso actually containers are connecting to the internet via lxc-br, which is created by lxc itself.18:53
velmeranwell, I think I over complicated things at first, so trying to step back to the beginning.18:54
noonedeadpunkAlso lxc should create src-nat rules in iptables18:54
*** rpittau is now known as rpittau|afk18:57
noonedeadpunkI guess basic diagram of networking is here https://docs.openstack.org/openstack-ansible/latest/reference/architecture/container-networking.html#network-diagrams18:57
noonedeadpunkalso https://docs.openstack.org/openstack-ansible/latest/user/network-arch/example.html#network-interfaces might be usefull18:58
noonedeadpunkbut configs there are for deb only18:58
velmeranyea.  I was trying to just give a seperate vm nic to each bridge, with each nic being on a seperate vlan.19:00
noonedeadpunkI think you ned routed environment then19:04
noonedeadpunklike https://docs.openstack.org/openstack-ansible/latest/user/l3pods/example.html ?19:05
noonedeadpunkah, wait19:06
noonedeadpunkgot you wrong19:06
noonedeadpunkI think it's ok to have each nic in each bridge19:06
noonedeadpunkbut you need another one for lxcbr anyway19:06
noonedeadpunkwhich will have access to internet19:07
velmeranokay, it made lxcbr, but it put it off on 10.0.3.1, which wasn't anything I specified, not sure how the other containers were talking with it, if they even were.19:07
*** cshen has joined #openstack-ansible19:09
velmeranI'm just looking at the openstack_user_config.yml.test.example file, trying to figure out the ip spaces I need to change so things match my network.19:09
velmeranseems like I have 3 cidr networks "container, tunnel, storage", and 4 bridges I would need "br-mgmt, br-vxlan, brvlan, br-storage"19:10
velmeranI had changed each cidr to a real vlan on my switches, made the bridges on each host (I forgot br-vlan, so that is a problem...), but was having issues with the containers connecttivity.19:12
*** cshen has quit IRC19:14
*** cshen has joined #openstack-ansible19:25
*** cshen has quit IRC19:30
*** joshualyle has joined #openstack-ansible19:30
*** joshualyle has quit IRC19:32
CeeMacjamesfreeman959: regarding galera cluster, I've found setting all but one of the containers not to auto boot then manually stopping mariadb on each of those containers and wait 2 mins inbetween. When all mariadb except one (on remaining autoboot container) are shutdown the one left should have the safe to boot flag and can be shut down. I find it works best if you do a rolling graceful shutdown of all the19:43
CeeMacnodes/containers that write to DB so that there is no traffic coming on to the cluster when you shut the last mariadb service down. Obviously you need to ensure that container starts up before any of the others to ensure the cluster is there.19:43
CeeMacStill borks sometimes though *shrug*19:43
*** djhankb7 has joined #openstack-ansible19:44
*** djhankb has quit IRC19:45
*** djhankb7 is now known as djhankb19:45
*** NewJorg has quit IRC19:46
*** NewJorg has joined #openstack-ansible19:46
*** Soopaman has joined #openstack-ansible19:50
*** DanyC has joined #openstack-ansible19:51
*** DanyC has quit IRC19:56
*** thuydang has joined #openstack-ansible19:56
*** thuydang has quit IRC19:58
*** thuydang has joined #openstack-ansible19:58
*** itsjg has joined #openstack-ansible20:09
noonedeadpunkvelmeran so on ctl hosts lxc should create one more bridge named lxcbr0 where all lxc interfaces will be attached with eth020:13
noonedeadpunkIt do not take any interface from controller - just use src-nat20:14
noonedeadpunkthis should be done automatcally on containers creation20:15
noonedeadpunkso all container should have at least 2 eth0 and eth120:15
noonedeadpunketh0 will probably have 10.0.3.0/24 and containers should be able to talk via it only on this spesific node - this network is not shared between controllers and used by containers to reach the world20:16
noonedeadpunkand eth1 - mgmt etwork through which containers should communicate with each other between nodes20:17
noonedeadpunkbtw probably you won't need br-vlan. also br-vxlan is also not so required - it can be regular interface in case you place neutron-agents on bare metal (without containers) as suggested in docs20:19
noonedeadpunkbut to simplify things for the beggining you may leave them :)20:20
velmeranyea, I was trying to figure out how I was going to handle vlan and vxlan with my setup as I'm not passing all vlans into the host.  I think I can just make some private networks in vmware for everything but br-mgmt, that one would be a nic back into my network where it could reach the internet etc.20:22
velmeranso long as that lxcbr0 is using that as the way out, it should work...20:22
noonedeadpunkactually vlan and vxlan both used for tenants private networks. and mostly only vxlan is used as it's more convenient and has less limitations20:24
*** jamesden_ has quit IRC20:25
noonedeadpunkand to make use of vxaln just intrfave (or another vlan) can be used since there never be any interface in that bridge except 1 from host20:25
mnaserhmm20:34
mnaserdoes anyone have any idea why https://github.com/openstack/openstack-ansible-openstack_hosts/blob/master/vars/debian.yml#L71 is there?20:34
mnaserit doesn't seem like having lvm2 installed on all hosts is something that's necessary20:34
mnasersome of these seem like they should live in specific repos20:35
mnaserlike bridge-utils20:35
guilhermespyeah at least official docs doesnt mention a computing needing lvm220:35
mnaser(context: lvm2 seems to be crashing debian nodes that guilhermesp is trying to deploy on)20:36
mnasernot sure if others have seen something similar :\20:36
mnaserguilhermesp: can you try pushing a change that makes that list empty and see what breaks?  ideally, we should fix the roles to install what they need..20:36
mnaserwhile you wait for reinstalling nodes :p20:36
spotzmnaser: Only crashing Debian?20:37
mnaseryes20:37
spotzI wonder if they changed anything though Debian isn't known for changing studd vs ubuntu20:37
spotzstuff....20:37
*** gshippey has quit IRC20:40
*** jamesden_ has joined #openstack-ansible20:46
velmeranSo I will need a br-vlan on my host, is it just my infra node, or also compute node?20:53
*** cshen has joined #openstack-ansible21:02
*** cshen has quit IRC21:07
jrossergood evening everyone21:09
jrosservelmeran: it pretty much depends on how you want your external and tenant networking to work21:10
jrosserbut as a good starting point for a simple life you might want to try to make the bridges uniform across all the nodes, even if they don't go anywhere (no vlan type networks on the compute nodes for example, but make the bridge anyway)21:11
jrosserthen the neutron config can be uniform everywhere21:11
jrosserbut you can set this up however you like really - the examples are just a suggestion21:12
openstackgerritMerged openstack/openstack-ansible-os_tempest stable/rocky: Use contraints for tempest plugins  https://review.opendev.org/71540721:13
jrossergit diff origin/master origin/stable/train defaults/main.yml21:15
jrosser-keystone_upper_constraints_url: "{{ requirements_git_url | default('https://releases.openstack.org/constraints/upper/' ~ requirements_git_install_branch | default('master')) }}"21:15
jrosser+keystone_upper_constraints_url: "{{ requirements_git_url | default('https://opendev.org/openstack/requirements/raw/' ~ requirements_git_install_branch | default('master') ~ '/upper-constraints.txt') }}"21:15
jrossermnaser: noonedeadpunk ^ we have this difference in pretty much all our roles - i think something might go wrong when we cut the U branch from master21:16
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible stable/rocky: Bump tempest role  https://review.opendev.org/71555421:17
*** cshen has joined #openstack-ansible21:18
*** cshen has quit IRC21:23
*** macz_ has quit IRC21:25
openstackgerritJonathan Rosser proposed openstack/ansible-config_template master: Switch to CentOS 8 based Tripleo job  https://review.opendev.org/71536721:33
*** jamesden_ has quit IRC21:43
*** macz_ has joined #openstack-ansible22:09
*** macz_ has quit IRC22:18
*** rh-jelabarre has quit IRC22:25
mnaserjrosser: is it because we use branch 'stable/xxx' ?22:31
*** thuydang has quit IRC22:49
jrossermnaser: yes it is - the url we use on master is 404 if you try to put a branch in there “stable/blah”22:55
jrosserrather than a release name on its own22:56
*** thuydang has joined #openstack-ansible22:58
velmeranhumm, my containers are able to get out to the internet, but they can't seem to see my internal or external lb_vip_addresses which are in my container cidr.23:03
velmeranI'm getting stuck on [python_venv_build : Install python packages into the venv], in the logs its failing: Getting page http://10.0.50.111:8181/os-releases/20.0.2/centos-7.7-x86_6423:06
velmeranwith 10.0.50.111 being my internal vip23:06
velmeranhitting that page, haproxy is returning a 50323:07
*** DanyC has joined #openstack-ansible23:08
*** DanyC has quit IRC23:13
*** cshen has joined #openstack-ansible23:19
openstackgerritMerged openstack/openstack-ansible-tests stable/train: Set requirements_git_url during functional tests  https://review.opendev.org/71448623:24
*** cshen has quit IRC23:25
*** cshen has joined #openstack-ansible23:36
*** NewJorg has quit IRC23:38
*** NewJorg has joined #openstack-ansible23:39
*** cshen has quit IRC23:40
*** thuydang has quit IRC23:58

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!