Wednesday, 2021-07-28

*** rpittau|afk is now known as rpittau07:02
qwebirc64230hi 07:44
qwebirc64230i m running the ansible-playbook setuphost.yml 07:45
qwebirc64230but gettting an error no host found 07:45
qwebirc64230though i have set up the host ip in the user-config yaml file. 07:45
qwebirc64230anything else that i can check for this issue07:47
jrosserqwebirc64230: you can paste any interesting output to paste.opendev.org if you want someone to look at it07:49
jrosseralso you can check that the inventory is parsed properly with the script openstack-ansible/scripts/inventory-manage.py07:50
qwebirc64230[root@tstosdep01 scripts]# ./inventory-manage.py --list-host +----------------------------------------------+----------+--------------------------+---------------+----------------+----------------+--------------------------+ | container_name                               | is_metal | component                | physical_host | tunnel_address | ansible_host   | container_types          | +----------------------------------------------+-------07:56
qwebirc64230i could see the containers 07:57
qwebirc64230i could see the containers details 07:57
qwebirc64230but the ip addresses are something which are different from the openstack-user-config file 07:57
qwebirc64230172.29.236.1407:57
qwebirc64230in my openstack user config yaml i have the CIDR mentioned : 10.0.160.0/2408:02
qwebirc64230but if i run the ./inventory-manage.py --list-host i could see the ip address of the container are in the CIDR : 172.29.236.1408:03
qwebirc64230still if i run the stup host playbook i get an error no host matching found 08:03
jrossercan you paste the openstack user config file to paste.opendev.org?08:13
qwebirc64230done08:14
jrosseryou'll need to share the url here08:15
qwebirc64230https://paste.opendev.org/show/807754/08:20
qwebirc64230my infra hosts are in different subnet like 10.0.16.0/2208:24
qwebirc64230and my CIDR are in the different subnet 08:24
qwebirc64230maybe that is causing the issue08:24
jrossersorry i have a meeting for a while - back later08:32
qwebirc64230no worries take ur time08:33
*** zbr is now known as Guest254409:09
jrosserqwebirc64230: i don't think you need cidr_networks defined inside global_overrides09:23
jrosseralso you should think about reserving IP addresses in your subnets with used_ips09:23
jrosserbecause it is coming up with addresses in 172.29.236.x i would guess that this is a host you've previously done a deployment on and have now changed the networking to use 10.x addresses instead09:24
jrosserthe only way that 172.29.236.x addressing can be used if it has (or used to be) set up that way in openstack_user_config09:25
jrosserthere are state files kept in /etc/openstack_deploy, and those are read to display the output when you use the inventory_manage script09:26
jrosserthats why I think you have an older config and your current one all somehow mixed up together09:26
qwebirc64230is it compulsory to provide the used ip's09:35
qwebirc64230ideally it should pick the free ip addresses within the CIDR09:36
qwebirc64230i have removed the generated files which are ip adress and inventory 09:37
qwebirc64230and now the playbook picks up my CIDR range09:37
qwebirc64230so the infra hosts should be in the CIDR range which is specified at the top or i can have a different subnet for my infra hosts 09:39
*** sshnaidm|afk is now known as sshnaidm09:45
jrosserqwebirc64230: it does not know which IP that you have used for routers, other hardware, ssh bastions in those subnets which are out of the scope of OSA09:47
jrosserit does pick free IP addresses in the subnets, but if you want the first 10(?) ip to be reserved for default gateway / vrrp / whatever.... then thats what used_ip is for09:48
jrosserqwebirc64230: you can have one subnet that is used for ssh (and by extension ansible) to access your hosts, like your 10.0.16.x09:49
jrosserand you can use completely different ones for mgmt / storage / tunnel, however you like09:50
jrosserthis is all very flexible and there's no particularly right or wrong answer09:50
qwebirc64230can i use 1 subnet for all management storage and tunneling 09:50
qwebirc64230is that possible09:51
jrossermost deployments make the mgmt network the same subnet as the physical hosts, but that certinaly is not a fixed rule and there are plenty that dont09:51
jrosseryou can perhaps collapse it all down to one network but i think there is much more risk of that not working, as generally on-one does that09:52
jrosserhighly un-tested09:52
qwebirc64230the deployment node where i have my ansible places should be in the same subnet as the mgmt network 09:52
jrosserit does not have to be09:52
jrosserit needs to be able to ssh to the bare metal hosts on the IP addresses you specify on openstack_user_config09:52
jrosseryou can make this be the mgmt network if you want09:52
jrosseror you can have a subnet dedicated to ssh / bare metal access - totally flexible here09:53
qwebirc64230got it 09:53
jrosserfor example in my deployments PXEboot and ssh happen on a network thats not mgmt network, they're seperate09:54
jrosserthis can be set up to fit around how you want to manage the hosts09:54
qwebirc64230cool10:14
opendevreviewAndrew Bonney proposed openstack/ansible-role-python_venv_build master: Add distro/arch to requirements file path  https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/80173810:46
jrosseroh no they did it again GPG error: https://packages.erlang-solutions.com/ubuntu focal Release: The following signatures were invalid: BADSIG D208507CA14F4FCA Erlang Solutions Ltd. <packages@erlang-solutions.com>11:27
evrardjpgood morning11:31
evrardjpmmm, that's not fun jrosser... :( 11:31
evrardjpDid they propose a new gpg signature file?11:31
jrosserthey do it every time :/11:31
jrosserno, when they release new packages they snafu up their repo11:32
evrardjpat some point, I recalled that I wanted to catch up with system provided rabbitmq, so that we can not care about those anymore...11:32
evrardjpI suppose that didn't happen? :)11:32
jrosseroh well we just got into a big big mess here with that on a bionic -> focal upgrade11:32
jrosserso sticking with the same version is most preferable11:33
evrardjp:D11:33
evrardjpunderstandable11:33
* jrosser lunchtime11:34
evrardjpI was checking our osa integrated repo,  and I am surprised (well, consider I am old!) to see collection requirements file..11:34
jrosseroh yes we went in fully on collections11:35
evrardjpI see we now have in tree roles,  out of tree roles,  collections11:35
jrosserit decouples the ansible version from the module version and is very nice11:35
jrosserwe can use up to date openstack modules without needed bleeding edge ansible11:35
evrardjpoh that's interesting, but what's the difference if it was in tree from a submodule/subtree? 11:36
evrardjpI don't see ceph-ansible using collections11:36
evrardjp(and will we move all roles to collections?)11:36
evrardjpis there a documented "future" state of collections for OSA,  so I can understand it a bit better?11:37
jrosserhmm?11:50
jrosserso far this is just for ansible modules11:50
jrosserwe install ansible-base which doesnt give you any modules at all11:51
evrardjpok11:51
evrardjpis there a plan to move all our roles into a certain collection? 11:51
jrosserso then everything thats actually needed like openstack / rabbitmq / mysql etc has to be specificed in the collection requriements file11:51
evrardjpok so that's not really using galaxy for the collections, it's just manually done, did I get that right?11:52
jrosserit uses the galaxy cli to install them, but the locations are git repos rather than the published collections11:53
jrosserhttps://github.com/openstack/openstack-ansible/blob/4d6c3a2ec743e149505e5b9c936dacee6d6d4379/scripts/get-ansible-collection-requirements.yml#L54-L6211:53
jrosserthats becasue the service that sits behind ansible galaxy is not sufficiently reliable for our CI use case11:54
evrardjphaha, same when we moved to use galaxy for a-r-r,  if you remember ;)11:54
jrossersoooooo much job failures in the past that we switched it to install from git repos hosted elswhere11:54
jrosserthe collections documentation discourages installation from git sources, but tbh real life has said otherwise11:55
jrosserevrardjp: would be good to have you reviewing code again :)11:58
evrardjpjrosser: I sadly don't have time for that! But I can bring more ppl on the table,  which isn't bad either :) 12:13
evrardjpjrosser: yes I am not surprised about the "do not install from git sources" .  But I am more puzzled nowadays on how we managed to make ansible more complex than what it should be ... 12:14
* evrardjp shrugs12:14
jrosseryeah, i find the "roles in collections" thing kind of hard to understand12:15
jrossertoo used to things just being simple and in git repos i guess12:15
evrardjpwell, I will be honest, and explain why I am here: I feel things too complex,  and I want to help simplify. Maybe not me directly,  but at least indirectly.  But for that I need proper input of the community ;)12:18
evrardjpAre you still using ceph-ansible jrosser ?12:19
jrosserwe are, though we deploy outside of OSA framework12:19
evrardjpI am wondering if we shouldn't be simple like them... finally admit ansible.cfg in our repo ;)12:19
evrardjpyeah that's what City Network is doing too12:19
evrardjpI find it "good enough" :)12:20
jrosserimho ceph-ansible is not a good example12:20
evrardjpoh ? 12:20
evrardjpdo you have a better example?12:20
jrosserlooked at from a step back "is this stuff going to change randomly and break my stuff?" -> probably12:20
evrardjp(Fun story,  I dug up one old documentation of an "ideal wrapper" that I wanted to bring in RAX long ago)12:21
evrardjpwell, that's not linked to the structure, that's linked to the content ;)12:21
jrossertheres a lot of tech-debt to address in general though12:23
jrosserso we can be busy on things related to SSL, secrets store, finally lose openstack-ansible-tests <- very very close now12:24
jrosserwould be interesting to hear what you think is too complex12:25
evrardjpit's the management of x clouds from a CI perspective12:26
evrardjpI would be happy to hear what you do there :) 12:26
jrosseryou mean getting toward CD?12:29
evrardjpcorrect12:31
evrardjpwell, CD is relatively easy: Just run the plays in prod 12:31
evrardjpproper testing of the plays in an pre-prod environment matching prod is always a challenge... 12:32
evrardjpI was seeing this problem at RAX,  but we structured things differently than CN,  so.... 12:32
evrardjpI am just wondering how the rest of the people are doing12:32
evrardjp(one of the technical annoyances is the multiple repos management of multiple sources for multiple environments, and simplifying this sounds key in our place)12:33
jrosserfrom on OSA perspective i think that one of the toughest things is that it's a toolbox12:38
jrosserwe run a quite large pre-prod environment which is as close to production as we can make it12:39
jrosserthough it's on an internal network rather than internet which makes that harder12:39
jrosserbut it is a very large overhead, in fact right now we're rebuilding it to address divergence between it and the prod environment12:40
jrosserfor multiple environments we create a virtualised deploy host per environment12:47
jrosserand those are reproducible, i.e we can destroy one and bring it back including all the state if necessary12:47
jrosserultimately i guess there still needs to be some sort of manifest for that, the OSA SHA, the role SHA, the collection versions, overrides for all our forked repos (doh!) 12:50
opendevreviewSatish Patel proposed openstack/openstack-ansible-os_neutron master: Change OVN metadata protocol to https  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/80213512:51
spateljrosser do you want me to add centos-8-stream job in this place also? - https://github.com/openstack/openstack-ansible/blob/master/zuul.d/jobs.yaml#L485-L548 12:54
jrosserspatel: for the OVN job?12:55
spatelfor ovs and ovn both 12:56
jrosserfor those i think they only need to go in the os_neutron repo12:56
spatelok i have already added stream in os_neutron repo just need to commit, this is what i going to add - https://paste.opendev.org/show/807765/12:57
spatelThat is all we need right? i think we should add ovn job also for stream12:59
jrosseroh, well at the moment we are only running ovn jobs on focal though?13:01
jrosseri would make small patches that do one thing, like switch the existing centos-8 repos for OVS jobs to stream13:02
jrosserthen if you want to add centos jobs for OVN make that another, there might be tons of stuff to fix there13:02
spatelgot it 13:03
spateli will keep testing stream in lab for ovn and then pushing it to CI 13:03
jrosserlooks like the calico job is broken13:03
spatellet me just add ovs at present 13:03
jrossersure13:04
spatelI have noticed calico job, i will try to see what is going on.. look like tempest is complaining 13:04
jrosseryeah, also thats still on bionic13:04
jrosserreally should be focal but i'm guessing there is some error in the neutron services13:05
opendevreviewSatish Patel proposed openstack/openstack-ansible-os_neutron master: Adding centos-8-stream job for ovs  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/80270113:05
spateli will take a look at that... 13:06
jrosserif you want to just switch the centos-8 jobs over to stream there is no need to keep the old ones13:06
anskiyovn with stream is a little bit broken now13:07
spatelit turned out to be easy way fix for ovn-metadata patch - https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/802135/2/templates/neutron_ovn_metadata_agent.ini.j213:07
spateljrosser should i remove centos-8 then? 13:07
spateli thought for sometime we will keep it but if you want i can remove it and just keep stream13:07
jrosserpart of the work to do for the next release is to remove centos-8 support13:08
jrosserso may as well start13:08
jrosserbut like i said yesterday keep an eye to what is happening for wallaby13:09
jrosserfor example see here, stable/wallaby is broken for OVN jobs right now https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/79888113:10
spateljrosser this is broken because of hostname and this patch will fix it https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/80213413:11
jrosserok so the thing preventing that merging is calico13:11
spatelyes only calico is holding it back.. i can take a look at that otherwise we can set non-voting 13:12
anskiyI'm currently expiriencing problem with OVN service name on stream, like, there is no ovn-central13:12
jrosseranskiy: is this a real deployment or in an AIO?13:14
anskiyit's non-aio lab in VMs13:14
jrosserif you are able to reproduce this in an AIO it would be super useful13:14
evrardjpinstead of making non voting, it's probably a good idea to call maintainers to fix it... I am pretty sure some of OSA users are using calico, and they don't want it to break. 13:14
jrosseranskiy: then we can compare it really directly with what is happening in our CI and getting it fixed quickly13:15
jrosserand it would also give us a data point if it does work in AIO but not multinode then it points to different kinds of bugs13:15
anskiyjrosser: well, I kinda have a fix, but it's not complete bc the path on rpm-based dists is /etc/sysconfig/<something>, not /etc/default: https://paste.opendev.org/show/807766/13:16
anskiyso I've just manually symlinked it to continue my tests :)13:16
spateljrosser i am removing centos-8 and re-committing patch 13:17
spatellets remove which is going to die anyway in few month 13:17
opendevreviewSatish Patel proposed openstack/openstack-ansible-os_neutron master: Adding centos-8-stream job for ovs  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/80270113:18
spateljrosser what is calico build for? does it doing something different? sorry i have less knowledge around that :)13:29
spateljrosser look like some metadata service is failing for calico build based on logs - curl http://169.254.169.254/latest/meta-data/public-ipv4' failed, exit status: 52 13:37
jrosserspatel: calico is an alternative network driver13:40
jrosserhttps://docs.projectcalico.org/about/about-calico13:40
spatelMay be haproxy all SSL vip issue - https://b2cd002d267de9376201-ea96d6532bea611ab21e2fe90ffd8bb3.ssl.cf2.rackcdn.com/802134/4/check/openstack-ansible-deploy-aio_metal_calico-ubuntu-bionic/b6c1e88/logs/etc/host/calico/felix.cfg.txt13:40
jrosseroh thats definatly possible13:40
spatelmay be doesn't know what protocol to use 13:40
spatelreading this and not sure it has option for https vs http - https://docs.projectcalico.org/reference/resources/felixconfig13:42
spatelGotta go will be back in 1 hour 13:43
opendevreviewAndrew Bonney proposed openstack/openstack-ansible master: haproxy: decrease check frequency for letsencrypt back ends  https://review.opendev.org/c/openstack/openstack-ansible/+/80271614:22
opendevreviewAndrew Bonney proposed openstack/openstack-ansible master: haproxy: decrease check interval for letsencrypt back ends  https://review.opendev.org/c/openstack/openstack-ansible/+/80271614:25
*** rpittau is now known as rpittau|afk16:21
dmsimardnoonedeadpunk: the delegate_to stuff sent me down quite the rabbithole17:35
dmsimardthe two main issues being that ansible returns the data about delegate_to differently depending if it's a loop task or not and then there's the potential for a task to be delegated to multiple hosts at the same time (i.e, with_items: {{ some_host_group }} and then delegate_to: {{ item }}17:36
dmsimardI have a workaround for the loop thing but the part about potentially multiple hosts being delegated to for a single task makes it tricky17:38
*** prometheanfire is now known as Guest262119:34
opendevreviewDavid Moreau Simard proposed openstack/openstack-ansible master: DNM: Test ara 1.5.7rc2 with --diff  https://review.opendev.org/c/openstack/openstack-ansible/+/69663423:04

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!