Tuesday, 2020-10-27

csmartre hardening, are the containers actually running SSH though?00:05
csmartI'm showing my ignorance, but it seems a little strange to me to SSH directly into a container, i.e. they might have configs but they aren't used00:06
*** macz_ has quit IRC00:09
*** gyee has quit IRC00:21
*** spatel has joined #openstack-ansible00:36
*** MickyMan77 has joined #openstack-ansible00:39
*** NewJorg has quit IRC00:41
*** rfolco has quit IRC00:42
*** NewJorg has joined #openstack-ansible00:42
*** MickyMan77 has quit IRC00:47
*** NobodyCam_ has joined #openstack-ansible00:52
*** NobodyCam has quit IRC00:52
*** NobodyCam_ is now known as NobodyCam00:52
*** mugsie has quit IRC01:00
*** mugsie has joined #openstack-ansible01:04
*** MickyMan77 has joined #openstack-ansible01:18
*** dave-mccowan has quit IRC01:20
*** cshen has joined #openstack-ansible01:28
*** cshen has quit IRC01:32
*** MickyMan77 has quit IRC01:53
*** MickyMan77 has joined #openstack-ansible02:49
*** spatel has quit IRC03:15
*** MickyMan77 has quit IRC03:23
*** cshen has joined #openstack-ansible03:28
*** cshen has quit IRC03:33
*** MickyMan77 has joined #openstack-ansible04:20
*** MickyMan77 has quit IRC04:55
*** cshen has joined #openstack-ansible05:28
*** evrardjp has quit IRC05:33
*** evrardjp has joined #openstack-ansible05:33
*** cshen has quit IRC05:33
*** MickyMan77 has joined #openstack-ansible05:50
*** MickyMan77 has quit IRC06:09
*** MickyMan77 has joined #openstack-ansible06:09
*** cshen has joined #openstack-ansible06:30
*** cshen has quit IRC06:35
*** recyclehero has joined #openstack-ansible06:40
*** miloa has joined #openstack-ansible06:43
openstackgerritamolkahat proposed openstack/openstack-ansible-os_tempest master: Migrate to content provider jobs/templates  https://review.opendev.org/75928707:20
*** rpittau|afk is now known as rpittau07:42
noonedeadpunkcsmart: we don't ssh into containers. We ssh only to lxc_hosts and use lxc tooling to enter container rather than ssh07:44
csmartnoonedeadpunk: yep that's what I thought đź‘Ť makes sense to me, ta07:45
*** nurdie has quit IRC07:47
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible master: Actually destroy container even if keeping data  https://review.opendev.org/72953307:48
*** cshen has joined #openstack-ansible07:51
*** andrewbonney has joined #openstack-ansible08:10
*** jbadiapa has quit IRC08:28
ptoWhy have you not chosen an pre-building approach, where the images are build and staged, then deployed to the hosts?08:30
*** macz_ has joined #openstack-ansible08:33
*** tosky has joined #openstack-ansible08:36
*** macz_ has quit IRC08:37
*** jbadiapa has joined #openstack-ansible08:38
*** cshen has quit IRC08:50
*** cshen has joined #openstack-ansible08:50
cshennoonedeadpunk: Hi, I found an etherpad link from you. https://etherpad.opendev.org/p/osa-rocky-bionic-upgrade Just want to know if this doc is still up2date?08:53
noonedeadpunkcshen: yes and no:) it can be used for some basic understanding. but thanks to ebbex we also have this https://docs.openstack.org/openstack-ansible/rocky/admin/upgrades/distribution-upgrades.html09:02
cshennoonedeadpunk: thank you very much!09:03
noonedeadpunkbut I think that etherpad might be still helpful09:07
kleiniwould be the upgrade from bionic to focal similar?09:08
ebbexkleini: Most of the steps should be similar, yes. Problems may appear in handling packages to repo_server, and might need some modifications to https://review.opendev.org/#/c/714483/09:12
cshennoonedeadpunk: yes, I read the page. It seems to mainly focus on control plane. Your etherpad link is also very useful.09:13
kleiniebbex, thanks. Unfortunately I have absolutely not idea about the repo server. Need to understand that first and then try to do the upgrade.09:16
MickyMan77I have added a certificate for the setting haproxy_user_ssl_cert and it's wokring perfect when I access the Horizon gui via a url. But when I try to access the console on the instances it point me the the ip addess instead of the url that point to the external vip addess.09:19
MickyMan77how can I change from ip addess to url ?09:20
ebbexMickyMan77: perhaps look at how nova_spice_html5proxy_base_uri resolves09:27
MickyMan77https://computersweden.idg.se/2.2683/1.741771/lacka-gunnebo-it-attack09:29
ebbexlooks like it uses external_lb_vip_address09:29
*** pto has quit IRC09:35
*** pto has joined #openstack-ansible09:40
*** pto_ has joined #openstack-ansible09:41
*** pto__ has joined #openstack-ansible09:41
*** pto_ has quit IRC09:45
*** pto__ has quit IRC09:53
*** pto has joined #openstack-ansible09:55
openstackgerritGaudenz Steinlin proposed openstack/openstack-ansible master: Use TCP mode for console if SSL is configured  https://review.opendev.org/57415310:03
*** spatel has joined #openstack-ansible10:05
*** spatel has quit IRC10:10
*** rfolco has joined #openstack-ansible10:31
*** pto has quit IRC10:39
*** pto_ has joined #openstack-ansible10:40
*** pto_ has quit IRC10:51
*** gshippey has joined #openstack-ansible10:55
*** macz_ has joined #openstack-ansible11:25
*** macz_ has quit IRC11:30
*** pto has joined #openstack-ansible11:31
*** pto has quit IRC11:36
*** pto_ has joined #openstack-ansible11:36
*** pto has joined #openstack-ansible11:38
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-repo_server master: Allow remote detection of repo sync status  https://review.opendev.org/66078811:40
*** pto_ has quit IRC11:42
*** rh-jelabarre has joined #openstack-ansible11:53
*** cshen_ has joined #openstack-ansible11:57
*** cshen has quit IRC12:00
ptoI dont quite understand the purpose of nova_libvirt_images_rbd_pool? Is this a flag to enable the Nova host to support ceph? Shoudl it be the same as the volume_backend_name defined in cinder_backends?12:09
noonedeadpunkpto: it's the storage for nova ephemerl drives, config drives or swap ones12:10
ptoAah... that makes sense. So it should be a sepperate pool from where the volumes are stored?12:11
ptoIs there a supported way to use an external ceph cluster (not provisioned from the openstack ansible)?12:13
guilhermesphey there, has anyone experienced octavia in a env with lxc + neutron-linuxbridge?12:18
admin0pto, external ceph is supported12:25
ptoadmin0: Cool. Do you have a hit how? Currently I have not defined any ceph nodes in openstack_user_config.yml, which makes the play skip the ceph client part12:28
admin0you have to give the fsid, pools and mons in user_variables12:28
ptoI have already done that and the play completes, but it does not create auth and pools. Is that working as intended?12:30
admin0it will not12:30
admin0you have to create the auths and pools yourself12:30
admin0that is the whole idea isn't it12:30
admin0you create the pools and auths12:30
admin0and then it will just copy the files for it to work12:30
admin0but skip touching ceph altogether12:31
admin0https://gist.github.com/a1git/1274f8593bf8257644d48579f151909912:31
admin0you have to create the pools and auths beforehand12:31
ptoadmin0: thanks. I see the ceph_client role is skipped when: ceph_mons | list | length == 012:33
ptoadmin0: Just wondering, if the mons should be defined in the group ceph_mons?12:34
ptoadmin0: Or will that make the play configure the existing mons?12:34
admin0the mons must be defined, so that it can ssh to the mons and copy the ceph configs12:34
admin0it will not create the mons for you .. just ssh and download the configs from it12:35
ptoadmin0: So, should i define  ceph-mon_hosts to the existing mons in the openstack_user_variables.yml?12:36
admin0i gave you a paste ^^12:36
ptoadmin0: sure, the user_variables.yml parts. Not the openstack_user_config.yml12:38
admin0user config will have nothing that says ceph12:39
admin0no mention of it12:39
jrosserguilhermesp: we use octavia like that here, lxc + linuxbridge12:40
admin0cinder will use this driver:  cinder.volume.drivers.rbd.RBDDriver12:40
admin0and then you will use the username and pool name12:40
ptoThe official docs says: https://docs.openstack.org/openstack-ansible-ceph_client/latest/config-from-file.html very similar to your gist, but if you define ceph_keyrings_dir: it should suck up the keys from there, but the play is skipped until you define ceph-mon_hosts. Im confused?12:40
jrosserpto: theres a subtely between the ceph mons ansible group and the list you can provide ceph_mons12:41
jrosserone is an ansible group that will create mons, the other is a list you can make that says "my externally provisioned  mons are here"12:41
ptoI see... Thanks allot! I will give it a test drive right away12:42
jrosserthey are named very similar so in the code, maybe _ vs - in the name so needs careful reading12:42
*** spatel has joined #openstack-ansible12:57
ptoThanks for the clarification. With the attach gist, will openstack obtain the keys from the mons directly?13:02
admin0it will .. if the deploy can ssh to it13:03
jrosserpto: is that OK that the deploy host can ssh to the mons? you can give all the keys manually if that is not possible13:05
ptojrosser: That was also my understanding, but for some reason, the pools were not created on the ceph cluster. It might be due to the unsupported --cluster parameter. I have tried to fix it with symlinks so far13:08
jrosserpto: from memory it is ceph-ansible (for example) that would create the pools13:10
jrosserso if you have an external cluster then i think it would be an action for the cluster admin to make the pools and keys13:10
ptoThank you all for helping. You are awsome!13:12
ptoIs the manila project a mandatory part of Ussuri? It seems impossible to disable it13:15
ptoand the installation is broken with default config (only config is manila-infra_hosts & manila-data_hosts set to infra hosts). The installation breaks at Copy manila configs with "msg": "'null' is undefined"13:18
admin0projects will not be added unless you defined it in the configs13:19
admin0maybe you had older artifacts ?13:19
*** hindret has joined #openstack-ansible13:22
noonedeadpunkneeded another vote for https://review.opendev.org/#/c/759308/313:23
*** mmethot has joined #openstack-ansible13:23
ptoadmin0: Nope. Its a fresh install and fresh config13:24
admin0then you can redo again :D13:24
admin0backup the /etc/openstack_deploy folder  and try again13:25
jrosserpto: when you do your own deploy (i.e not with an AIO) you will get what you define in /etc/openstack_deploy/openstack_user_config.yml13:28
jrosserthere is no default for that and you can opt in/out of different services exactly as you wish13:28
jrosserthere are a bunch of examples here https://github.com/openstack/openstack-ansible/tree/master/etc/openstack_deploy13:29
jrosserbut really they are kind of just inspiration / hints rather than something you must take completely13:29
ptojrosser: I think there is a bug in the os_manila role then. It were partly deployed without the openstac_user_variables.yml containing the defintions. Im gonna redeploy later this week, and then I will verify if its a bug and report it, if so,13:32
jrosserthere may be additional config required for manila13:33
jrosserquite a good way to find out about that is to look at the CI jobs, for example OSA manila is here https://review.opendev.org/#/q/project:openstack/openstack-ansible-os_manila13:33
jrosseryou would be able to run up a local AIO with manila as an exact copy of those CI runs13:34
ptoCool! I will have a look at it13:34
*** cshen has joined #openstack-ansible13:35
noonedeadpunkwell yeah, we even see issues with manila with lxc deployment, so some work probably needs to be done. In CI we run with ceph-nfs backend (with ganesha)13:35
*** cshen_ has quit IRC13:35
noonedeadpunkbut generic drivers should also work13:35
jrosseryou can look in the stuff that sets up the CI job to see the settings we use https://github.com/openstack/openstack-ansible/blob/master/tests/roles/bootstrap-host/templates/user_variables_manila.yml.j213:35
ptojrosser: I finally got the federated identity working on ussuri. Its far from hitless, but it seems to work. The major "but" is that the setup of federated domain cannot be done before the keystone-manage db_sync --contracts has been run13:35
jrosserpto: that is really interesting13:36
ptojrosser: When keystone comes up initially, there are multiple SQL triggers which protects the db until it has been patched with the contracts13:36
jrosseri do not really know enough about how the db migrations are done to say if it is a bug in keystone or not13:36
jrosserbut what you describe does make sense, in deployments here we have always done a simple first pass without federation13:37
jrosserthen enabled federation later once we were happy the deployment was good13:37
*** nurdie has joined #openstack-ansible13:38
ptojrosser: I dont know why the actual db_migrates is executed last? https://github.com/openstack/openstack-ansible/blob/47e5a90a7fcc78adc44bbd0803e0faabb56197b6/playbooks/os-keystone-install.yml#L135 I would think it should be run as one of the first things in the actual os_keystone role?13:38
jrossernoonedeadpunk: ^ i'm not sure what we should do about that, part of the db migrations is in post_tasks in the os_keystone playbook13:39
jrosserfeels kind of wierd13:39
jrosseroh actually it's a second play entirely, there must be a really specific reason for that13:39
noonedeadpunkwell it was added during upgrades13:40
ptojrosser: I have no clue why is has been put so late, but the play https://github.com/openstack/openstack-ansible-os_keystone/blob/bc3f922b903efeeb99164010bfdea744f3d3de6b/tasks/main.yml#L230 will fail if the db has not been migrated before13:41
noonedeadpunkAnd I think it's run after first host is installed or smth like that13:41
jrosserthis is why https://github.com/openstack/openstack-ansible/commit/ecf32e20784d733fa89e1b0a392fd4b565ca4f4113:41
ptoI actually discovred the problem in the queens release and the same workaround worked (commenting the part out, and finalize the keystone install, then comment it in and run it again)13:42
jrosseri think that the safest change would be to remove the federation setup completely out of main.yml in os_keystone13:43
jrosserthen we could call the role again right at the end of os_keystone with tasks_from: federation_setup.yml13:43
ptoAlternative, the flag need_db_contract = False in  /etc/ansible/facts.d/openstack_ansible.fact could be checked and the task skipped13:45
ptoBut anyway it should be moved to later in the play13:45
jrosserwell it's in a role, thats the trouble13:48
jrossernot the play13:48
*** dave-mccowan has joined #openstack-ansible13:48
openstackgerritJames Denton proposed openstack/openstack-ansible-os_neutron master: Implement uWSGI for neutron-api  https://review.opendev.org/48615614:01
openstackgerritMerged openstack/openstack-ansible-lxc_hosts master: Determine latest base image available  https://review.opendev.org/75922914:05
*** sshnaidm|rover has quit IRC14:05
*** sshnaidm has joined #openstack-ansible14:05
*** sshnaidm is now known as sshnaidm|rover14:06
openstackgerritJames Denton proposed openstack/ansible-role-systemd_networkd master: Add GPG Key for EPEL8 Repo  https://review.opendev.org/75914514:08
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible master: Fix upgrade jobs for bind-to-mgmt  https://review.opendev.org/75846114:09
openstackgerritMerged openstack/openstack-ansible-tests stable/stein: Pin virtualenv<20 for python2 functional tests  https://review.opendev.org/75930814:17
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_adjutant master: Make role fit to the OSA standards  https://review.opendev.org/75631314:20
*** cshen has quit IRC14:20
noonedeadpunkjrosser: did we want to bump amount of threads in this patch? https://review.opendev.org/#/c/705680/2/defaults/main.yml14:37
noonedeadpunkI think just number of processes?14:38
noonedeadpunkglance_api_threads just confusing dublicate name of glance_api_workers14:38
*** macz_ has joined #openstack-ansible14:49
openstackgerritJames Denton proposed openstack/openstack-ansible-os_neutron master: Clean up Neutron metering configuration  https://review.opendev.org/75988214:53
openstackgerritJames Denton proposed openstack/openstack-ansible-os_neutron master: Clean up Neutron metering configuration  https://review.opendev.org/75988214:54
*** macz_ has quit IRC14:54
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_aodh master: Reduce number of processes on small systems  https://review.opendev.org/75988314:55
jrossernoonedeadpunk: oh yes that is really confusing15:01
jrosserextra variable for really no purpose15:02
*** macz_ has joined #openstack-ansible15:10
*** macz_ has joined #openstack-ansible15:10
*** yann-kaelig has joined #openstack-ansible15:31
*** gyee has joined #openstack-ansible15:34
*** miloa has quit IRC15:41
admin0TASK [lxc_hosts : Ensure image has been pre-staged] -- this tries to access  http://cdimage.ubuntu.com/ubuntu-base/releases/18.04/release/ubuntu-base-18.04.3-base-amd64.tar.gz  which does not exist anymore15:55
kleiniadmin0, https://review.opendev.org/#/c/75922915:57
noonedeadpunkadmin0: http://lists.openstack.org/pipermail/openstack-discuss/2020-October/018280.html15:58
noonedeadpunkbut yeah, this patch  works:)15:59
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_barbican master: Reduce number of processes on small systems  https://review.opendev.org/75989516:03
*** tosky_ has joined #openstack-ansible16:07
gshippey@noonedeadpunk  are we doing bug triage in the meeting now, or at following the PTG schedule later in the day?16:07
*** tosky has quit IRC16:08
noonedeadpunkgshippey: I think we agreed to skip PTG16:08
noonedeadpunkbut yes, I think we need to discuss some topics regarding bugs as well16:10
openstackgerritJames Denton proposed openstack/openstack-ansible-os_neutron master: Implement uWSGI for neutron-api  https://review.opendev.org/48615616:10
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-lxc_hosts stable/ussuri: Determine latest base image available  https://review.opendev.org/75990016:12
jrosseroh, hmm16:12
noonedeadpunkhttps://review.opendev.org/#/c/759298/ ?16:13
jrosseryeah, just realised16:13
noonedeadpunkI've placed for all branches at once16:13
noonedeadpunkjust for stein there's extra dependency16:13
jrosseryeah makes sense16:13
noonedeadpunkas we can't backport renos :) and different set of os anyway16:14
jrossernoonedeadpunk: is it meeting time?16:22
noonedeadpunkmmmm16:25
noonedeadpunksummer saving time damn it16:26
noonedeadpunk#startmeeting openstack_ansible_meeting16:26
openstackMeeting started Tue Oct 27 16:26:22 2020 UTC and is due to finish in 60 minutes.  The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot.16:26
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.16:26
*** openstack changes topic to " (Meeting topic: openstack_ansible_meeting)"16:26
openstackThe meeting name has been set to 'openstack_ansible_meeting'16:26
noonedeadpunk#topic office hours16:27
*** openstack changes topic to "office hours (Meeting topic: openstack_ansible_meeting)"16:27
noonedeadpunkWell I realized we have some topics to discuss after PTG16:27
noonedeadpunkfirst of all, I clean forgot to raise topic with our bugs16:28
noonedeadpunkI think we need to somehow cleanup current bug reports, as there are tons of obsolete ones or issued for non supported releases16:29
jrosserwere do we say non supported is?16:30
noonedeadpunkI was thinking about ones in EM16:30
noonedeadpunkBut we don't say this directly at the moment I think16:31
noonedeadpunkbut we can close them conditionally when we obviously see this is should not be relevant anymore16:31
noonedeadpunklike https://bugs.launchpad.net/openstack-ansible/+bug/173672616:33
openstackLaunchpad bug 1736726 in openstack-ansible "Pike update causes parallel restart of Galera containers" [High,Confirmed] - Assigned to Jesse Pretorius (jesse-pretorius)16:33
noonedeadpunkI never faced this issue tbh16:33
noonedeadpunkBut it seemed relevant these days16:33
noonedeadpunkor should we try to solve https://bugs.launchpad.net/openstack-ansible/+bug/1778663 ?16:33
openstackLaunchpad bug 1778663 in openstack-ansible "Pike→Queens upgrade fails on final step in run_upgrade.sh (running haproxy-install.yml)" [High,In progress] - Assigned to Antony Messerli (antonym)16:33
jrosseri guess relatedly we should move one/some more releases to EM16:33
noonedeadpunkwell yes, it's stein...16:34
noonedeadpunkI want to do last point release with fixed images and move to stable/stein branch afterwards16:34
noonedeadpunklike we did for rocky16:35
noonedeadpunkSo ideas what should we do with pretty old bugs? Should we try to resolve them? Or close with saying that in case it happens for currently supported release please re-submit?16:38
noonedeadpunkAs I really want to start keeping track on our bugtracker16:38
jrossergiven we can only really actively support the later branches i would be closing the bugs where possible16:40
gshippeyhttps://bugs.launchpad.net/openstack-ansible/+bug/1744014 - have to still fix up my patch but hopefully we can get some movement on this one soon. I think we should cull the older bugs, especially if they're minor and ask for resubmission's where necessary.  https://bugs.launchpad.net/openstack-ansible/+bugs?orderby=-id&memo=&start=300&direction=backwards <- bugs listed oldest to newest, might be useful16:40
openstackLaunchpad bug 1744014 in openstack-ansible "[Docs] Expected Backups in OpenStack-Ansible" [Medium,Confirmed]16:40
*** SecOpsNinja has joined #openstack-ansible16:41
noonedeadpunkyeah, it is:)16:42
jrossersome of this is really easy, its for deprecated things16:43
Adri2000I wouldn't be shocked if older bugs, especially if no one can reproduce them on recent releases, were closed16:43
Adri2000(even my bugs :))16:43
noonedeadpunk#agreed to close old bugs for non supported / EM releases with asking to re-create in case still relevant16:44
noonedeadpunkI will try to go through them16:44
*** cshen has joined #openstack-ansible16:45
noonedeadpunkRegarding bug https://bugs.launchpad.net/openstack-ansible/+bug/190161916:46
openstackLaunchpad bug 1901619 in openstack-ansible "Ansble-hardening role is not applied to containers" [Undecided,New]16:46
noonedeadpunkI'd say it's invalid?16:46
noonedeadpunkoh, wait16:46
SecOpsNinjahi everyone. im  having a strange error iin nova api regargind oslo.messaging._drivers.impl_rabbit ... Connection failed: [Errno 113] EHOSTUNREACH (retrying in 2.0 seconds): OSError: [Errno 113] EHOSTUNREACH but i have confirmed that rabbitmq cluster is runing ok with 3 runing nodes and all the porst configured in /etc/nova/nova.conf are accesable using telnet test. what could be the cause o16:46
SecOpsNinjaf this failure?16:46
noonedeadpunklet's first decide if we want to return recent bug triage?:)16:46
noonedeadpunkI think that might be pretty useful to get other team opinion on bugs16:48
noonedeadpunkand maybe will bring more involvment into meetings (hopefully but unlikely)16:48
noonedeadpunkAs I feel that doing triage was a good thing back then16:49
jrosseri think getting more folk involved is good, particularly anyone using older releases16:49
*** cshen has quit IRC16:49
gshippeyđź‘Ť16:49
openstackgerritMerged openstack/openstack-ansible-lxc_hosts stable/ussuri: Determine latest base image available  https://review.opendev.org/75929816:50
jrossermine tend to be N-1 age, others may have a different perspective16:50
jrosserebbex: is this something you can help with?16:50
noonedeadpunkI think we here do N-1.5 or smth like this:) So not using every release but jumping through one, so doing upgrades on yearly basis16:51
noonedeadpunk#agreed doing bug triage during meetings16:54
noonedeadpunkso, https://bugs.launchpad.net/openstack-ansible/+bug/1901619 :)16:54
openstackLaunchpad bug 1901619 in openstack-ansible "Ansble-hardening role is not applied to containers" [Undecided,New]16:54
noonedeadpunkanyone thinks we should run hardening against containers as well?16:55
gshippeyperhaps from a technical perspective, but doesnt the hardening role take ages to run16:56
noonedeadpunkI don't think it's pretty useful there as we're using minimal images, time is taken from host and containers shouldn't be directly accesible16:56
noonedeadpunkgshippey: well it's pretty fast16:56
*** tosky_ is now known as tosky16:56
gshippeydon't know why i feel it always takes an age - ignore me!16:56
noonedeadpunktakes 2 mins according to ara:)16:57
noonedeadpunkhttps://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_da2/759229/4/check/openstack-ansible-deploy-infra_lxc-centos-8/da27ca7/logs/ara-report/playbooks/2.html16:57
noonedeadpunkI think I know why - it has so many tasks....16:58
noonedeadpunkso easy thing we can do if we want hardening to run against containers - is just change the order of execution in setup-hosts.yml16:58
noonedeadpunkactually, hosts set is defined with a variable16:59
openstackgerritMerged openstack/openstack-ansible-lxc_hosts stable/train: Determine latest base image available  https://review.opendev.org/75929916:59
noonedeadpunkso see no reason in not changing the order and allowing ppl to decide if they want it or not16:59
jrosserthey containers still do need some ssh17:01
jrossertheres some rsync and things isnt there? and keystone key distribution17:02
jrosserand rotation17:02
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible master: Run hardening after container deployment  https://review.opendev.org/75990717:03
noonedeadpunkwell yes, I saw ssh running17:03
noonedeadpunkBut I thought that it's for repo server only17:04
*** odyssey4me is now known as odyssey4me|PTO17:04
noonedeadpunkwell we've returned it to the list of the packages but I forgot what removal of openssh breaks except repo17:05
noonedeadpunkso should we try to run hardening against all hosts by default17:06
noonedeadpunk?17:06
*** aedc_ has joined #openstack-ansible17:10
*** aedc has quit IRC17:11
jrossersorry just dealing with other things here17:11
noonedeadpunklet me end meeting then and we can continue later:)17:15
noonedeadpunk#endmeeting17:16
*** openstack changes topic to "Launchpad: https://launchpad.net/openstack-ansible || Weekly Meetings: https://wiki.openstack.org/wiki/Meetings/openstack-ansible || Review Dashboard: http://bit.ly/osa-review-board-v2"17:16
openstackMeeting ended Tue Oct 27 17:16:03 2020 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)17:16
openstackMinutes:        http://eavesdrop.openstack.org/meetings/openstack_ansible_meeting/2020/openstack_ansible_meeting.2020-10-27-16.26.html17:16
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/openstack_ansible_meeting/2020/openstack_ansible_meeting.2020-10-27-16.26.txt17:16
openstackLog:            http://eavesdrop.openstack.org/meetings/openstack_ansible_meeting/2020/openstack_ansible_meeting.2020-10-27-16.26.log.html17:16
*** MickyMan77 has quit IRC17:16
openstackgerritJames Denton proposed openstack/openstack-ansible-os_neutron master: Clean up Neutron metering configuration  https://review.opendev.org/75988217:16
*** rpittau is now known as rpittau|afk17:34
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_neutron master: Switch functional tests to focal  https://review.opendev.org/75991417:43
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_cinder master: Reduce number of processes on small systems  https://review.opendev.org/75991617:50
*** MickyMan77 has joined #openstack-ansible18:01
*** MickyMan77 has quit IRC18:10
*** ThiagoCMC has quit IRC18:17
*** andrewbonney has quit IRC18:21
*** cshen has joined #openstack-ansible18:25
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible master: Bump SHAs for master  https://review.opendev.org/75597319:04
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible master: Added Openstack Adjutant role deployment  https://review.opendev.org/75631019:04
*** MickyMan77 has joined #openstack-ansible19:06
*** itsjg has joined #openstack-ansible19:07
SecOpsNinjais there any way to check what ERROR oslo.messaging._drivers.impl_rabbit [req-6f0baa57-aa50-4652-9539-a6ed93f25bad - - - - -] is causing this?19:09
itsjgSecOpsNinja It's strange, I see this message on both my controllers and compute nodes every so often as well. However I always confirm my rabbitmq cluster is healthy and I don't have network connectivity issues. I've never seen it directly impact services if the message is occasional. If you're seeing it non-stop and your noticing services crashing that's probably an indication of greater rabbitMQ issue. Does it appear to be impacting services?19:14
*** MickyMan77 has quit IRC19:15
SecOpsNinjayep im trying to start a vm and doesn t work it gets stuck. but all services say they are up and runing19:15
SecOpsNinjaeven when going to rabbitmq cluster it says is has all 3 nodes19:15
SecOpsNinjaand from nova containers where im seeing does logs i can ping each rabbitmq cluster and cluseck that is ports are open19:16
SecOpsNinjanova placement is workign correctly19:16
SecOpsNinjabut i can put any vm up and runing....19:16
SecOpsNinja*can't19:17
itsjgAhh... I believe I ran into this once before and had to restart all my RabbitMQ containers. Couldn't figure out the root cause, though. If you go into the rabbitmq container and check in /var/log/rabbitmq, do you see any hints?19:17
admin0is there a known issue in creating galera users ?19:18
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_tempest stable/train: Bump magnum tempest plugin  https://review.opendev.org/75993219:18
noonedeadpunkadmin0: not I'm aware about19:18
openstackgerritDmitriy Rabotyagov (noonedeadpunk) proposed openstack/openstack-ansible-os_magnum stable/train: Add deployment of keystone_auth_default_policy  https://review.opendev.org/75947219:19
SecOpsNinjaitsjg,  i have restart all containers and changed keepalive to confirm that all control panel where restarted19:19
SecOpsNinjain /var/log/rabbitmq/rabbit\@*.log i have information that various machines connected to rabbitmq19:20
noonedeadpunkSecOpsNinja: what release is it?19:20
SecOpsNinjaussuri i believe19:20
admin0The PyMySQL (Python 2.7 and Python 3.X) or MySQL-python (Python 2.X) module is required.19:21
admin0this is 20.x branch on 18.0419:21
noonedeadpunkSecOpsNinja: and how instance creation fails?19:22
noonedeadpunkI mean what error does it gives?19:22
noonedeadpunkAs porbably rabbit may be not directy related here19:22
noonedeadpunk*probably19:23
noonedeadpunkadmin0: hm, and you don't have any overrides of *_db_setup_host or maybe you've pulled some master role?19:26
admin0its on 20.0.4 branch19:26
SecOpsNinjanoonedeadpunk, the problem was strating vms is being stuck in starting phase. now i was able to start one but console doesn't work i was able to confirm its runign in one compute host with virsh list.  but if i try to create a new instance from scratch  it stays in building forever. i try to create a new volume (based on a image at gives erro very quickly19:26
noonedeadpunkwell it feels like rabbit indeed19:27
noonedeadpunkwhat I always recommend with rabbit - run rabbitmq-install.yml -e rabbitmq_upgrade=true19:28
noonedeadpunkthis re-creates cluster and queues19:29
noonedeadpunkat least doesn't makes things worse19:29
SecOpsNinjaok i will trty that19:29
SecOpsNinjai think its a rabbitmq problem but it doesn seem to be a connection but probably queues19:29
noonedeadpunkadmin0: for 20.0.4 we usually delegate to galera_all[0] node. and use system pip there19:31
noonedeadpunk*system python19:31
noonedeadpunkbut I'd expect it to have PyMySQL module installed19:31
noonedeadpunkthe only thing that may go wrong, is if ansible used py2 but now is using py3. so it has installed dependency for py2 but now tries to use another binary which does not have it19:32
*** mmethot has quit IRC19:33
noonedeadpunkthe most safe thing would be just install PyMySQL on that container manually, or you can try re-running galera_client role19:33
noonedeadpunkhm, looking through role I can't find where we did install PyMySQL there...19:35
jrosseradmin0: it would be useful to see a paste of the error and task output, there could be something we fixed on master and did not backport that far19:35
SecOpsNinjanoonedeadpunk,  is there any list i can check regarding what queues i have available in cluster ? because when i do rabbitmqctl list_queues it only shows Timeout: 60.0 seconds ... and Listing queues for vhost / ...19:36
jrossernoonedeadpunk: it's kind of here on master https://github.com/openstack/openstack-ansible/blob/master/inventory/group_vars/utility_all.yml#L5819:37
noonedeadpunkjrosser: I think we could add https://opendev.org/openstack/openstack-ansible-galera_server/commit/021b1621e4a9d4ac5de606884eecd8ffb9efcc7c after 20.0.419:37
noonedeadpunknot sure19:37
*** mmethot has joined #openstack-ansible19:37
jrosserthese need to be in the utility venv with pip i think19:38
noonedeadpunkSecOpsNinja: we have a separate vhost for each service19:38
jrosserbecasue it's ansible tasks which do the db user creation19:38
noonedeadpunkjrosser: we were delegating to different host in train19:38
jrosserooooohhhh19:38
noonedeadpunkwhich was galera_all[0]19:38
noonedeadpunkwe've change that just in master I think19:39
noonedeadpunkSecOpsNinja: so you need to rabbitmqctl list_queues -p nova (or whatever)19:39
noonedeadpunkbut rabbit was able to stuck very weirdly, when it was operating from the first sight, but queues were broken somehow19:40
noonedeadpunknot sure how to explain it as never understood, but also didn't faced with that since train...19:40
SecOpsNinjanoonedeadpunk, i idn't understand the -p part. i should put the name of the contaiener or the host?19:42
noonedeadpunkof the rabbitmq vhost....19:42
noonedeadpunkrabbitmqctl list_vhosts19:42
* noonedeadpunk needs to go off for today19:43
SecOpsNinjanoonedeadpunk,  thanks gain for all the help19:43
SecOpsNinjaregarding nova-scheduler is this normal? nova.scheduler.host_manager [req-90773d87-a4ed-45f7-ba7e-5429daf0a0fa - - - - -] Received a sync request from an unknown host 'osa-compute-02'. Re-created its InstanceList.19:55
dmsimardthat's a somewhat normal message, iirc it means it's the first time that node is checking in to nova conductor20:00
SecOpsNinjaok i enable debug mode in nova api to see if i can understand that error and try to understand what is causing that20:03
SecOpsNinjai can't understand this error...20:04
*** ChiTo has joined #openstack-ansible20:04
SecOpsNinjais there a way to check that req id to see who started it and to where is trying to connect ?20:05
itsjgWhat kind of operation was it, resize, migration, etc? You can get more info if you do a grep for the req ID in journalctl (helps to know what the operation was to know which container to look in, assuming you don't have centralized logging)20:06
*** melwitt has joined #openstack-ansible20:06
SecOpsNinjabecause if i start a instance by dashbord, it rerunt 504 timeout and when i reload it it says in p+owering on and doesn't leave that state20:07
itsjgOh I see. It would be helpful to run a journalctl on the compute node hosting that and grep for the reqID to get more detailed info, you may also have to enable debug logging on the compute node's nova.conf if there isn't much info20:09
*** ChiTo has quit IRC20:10
*** MickyMan77 has joined #openstack-ansible20:11
SecOpsNinjayep that is the very strange part . i dont see any error in nova-compute or neutron service in compute hosts....20:12
SecOpsNinjaregarding tthe req that gives that Connection failed: [Errno 113] EHOSTUNREACH the previous logs messages regarding that is http://paste.openstack.org/show/799444/ where is talks about Releasing lock "compute-rpcapi-router" ... any ideia what maybe causing this?20:15
*** rpittau|afk is now known as rpittau20:18
SecOpsNinjaand the again another starnge fact... the vms started a very long wait lol20:18
SecOpsNinja*after a very long wait in powering up....20:19
*** MickyMan77 has quit IRC20:20
SecOpsNinjai will try to check tomorow with fresh eyes. thanks everyone o720:25
*** SecOpsNinja has left #openstack-ansible20:31
*** cshen has quit IRC20:31
admin0i removed all /etc/ansible/roles and did a fresh start ( to ensure nothing from any other/master branch has an influence)20:50
admin0if i see the error, i will paste it20:50
*** gouthamr has quit IRC20:58
*** logan- has quit IRC20:58
*** gouthamr has joined #openstack-ansible20:59
*** gouthamr has quit IRC20:59
*** gouthamr has joined #openstack-ansible20:59
*** logan- has joined #openstack-ansible21:01
*** rpittau is now known as rpittau|afk21:03
*** MickyMan77 has joined #openstack-ansible21:05
*** MickyMan77 has quit IRC21:13
*** yann-kaelig has quit IRC21:29
*** spatel has quit IRC21:38
*** aedc_ has quit IRC21:42
*** sshnaidm|rover is now known as sshnaidm|afk21:50
*** rfolco has quit IRC22:12
*** cshen has joined #openstack-ansible22:24
*** cshen has quit IRC22:29
*** MickyMan77 has joined #openstack-ansible22:40
*** pcaruana has quit IRC22:46
*** MickyMan77 has quit IRC22:48
fridtjof[m]An issue I've consistently been having over multiple releases on initial deployments:23:00
fridtjof[m]When setting up everything (currently Train, ubuntu 18.04, simple storage node with lvm), creating instances fails with a stacktrace containing the message "Connector doesn't have required information: initiator".23:02
fridtjof[m]plugging this into google, I stumble upon this old post: https://ask.openstack.org/en/question/118921/unable-to-attach-volume-due-to-error-invalidinput-invalid-input-received-connector-doesnt-have-required-information-initiator/23:03
fridtjof[m]which tells me to start iscsid on the compute hosts (it's enabled, so a reboot would work too, I guess). After doing that, everything works fine.23:03
fridtjof[m]Seems like a fairly simple to fix oversight in some playbook?23:03
*** cshen has joined #openstack-ansible23:27
*** gshippey has quit IRC23:32
*** cshen has quit IRC23:32
*** macz_ has quit IRC23:37
*** MickyMan77 has joined #openstack-ansible23:45
*** tosky has quit IRC23:52
*** MickyMan77 has quit IRC23:53

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!