Saturday, 2021-02-20

*** luksky has quit IRC00:03
*** tosky has quit IRC00:10
*** waxfire7 has joined #openstack-ansible01:13
*** waxfire has quit IRC01:13
*** waxfire7 is now known as waxfire01:13
*** spatel has joined #openstack-ansible01:16
*** LowKey has quit IRC01:25
*** LowKey has joined #openstack-ansible02:20
*** LowKey has quit IRC02:25
*** spatel has quit IRC03:15
*** spatel has joined #openstack-ansible03:17
*** waxfire4 has joined #openstack-ansible03:41
*** waxfire has quit IRC03:41
*** waxfire4 is now known as waxfire03:41
*** evrardjp has quit IRC05:38
*** evrardjp has joined #openstack-ansible05:39
*** spatel has quit IRC06:18
*** gyee has quit IRC06:31
*** lemko has quit IRC07:30
*** lemko7 has joined #openstack-ansible07:30
*** lemko7 is now known as lemko07:30
MrClayPoleadmin0: A 169.254.x.x address is assigned to a DHCP client when it can't get a DHCP address. Could it be the script is running to early in the boot process?07:32
*** waxfire has quit IRC07:36
*** waxfire has joined #openstack-ansible07:36
*** dirtwash has joined #openstack-ansible09:00
dirtwashhi, im running openstack victoria with ceph rbd backend. snapshots are failing with rbd incomplete write errors, heres glance log, maybe someone has any idea or hint? http://paste.openstack.org/show/802855/09:00
*** luksky has joined #openstack-ansible09:22
*** tosky has joined #openstack-ansible10:13
openstackgerritMerged openstack/openstack-ansible master: Stop installing client packages for ansible venv  https://review.opendev.org/c/openstack/openstack-ansible/+/77447210:20
*** pcaruana has quit IRC10:27
*** waxfire3 has joined #openstack-ansible10:33
*** waxfire has quit IRC10:35
*** waxfire3 is now known as waxfire10:35
*** jbadiapa has quit IRC10:45
admin0MrClayPole, the script is passed via cloud init10:59
MrClayPoleAh, so its more Likely the script is being proulated with the wrong value?11:01
jrosser169.254.169.254 is the correct address for the metadata service11:02
jrosserit's a link-local address used before the real ip is known11:03
CeeMacadmin0: any issues or delays with the vif port binding?11:08
*** noonedeadpunk has quit IRC11:16
*** noonedeadpunk has joined #openstack-ansible11:17
admin0this is what I get in the cloud-init logs .. https://gist.github.com/a1git/c7e8b5de5b4cfb28e203d01a3d164b67  .. but as soon as it boots and i login and do the curl, its fine .. i can curl the metadata service11:20
CeeMacAnything in nova logs? If there was a delay plugging the vif it could explain this maybe?11:23
admin0actually , just booted up a new instance and i cannot curl to that metadata service even after it booted up11:24
admin0so there were almost 20 instances like this where i have the logs saying cannot connect and then it was able to connect11:24
admin0tried a new one today, and this one cannot connect to the metadata service even after its booted up11:25
CeeMacAll on same host or?11:25
admin0when i do ip route, i see 2 entries for metadata .. one for .1 and one for .22( could be the metadata ip)11:25
CeeMacI presume the metadata service is up and alive?11:26
admin0via admin , i can see that all neutron agents and services are alive and working11:26
CeeMacIf you can check the events for one of the new instances to get the req id check nova logs for that req id and see if there are any vif port bind errors11:27
admin0why would there be 2 entries for the metadata service11:27
admin0one to 192.168.0.1 and 192.168.0.2211:28
admin0i am checking what .22 is11:28
*** LowKey has joined #openstack-ansible11:28
admin0so .22 is network dhcp11:29
CeeMac1 would be the router I'd guess?11:30
admin0yeah11:30
admin0do you guys limit the number of dhcp per network ?11:31
CeeMacIp range?11:31
admin0neutron_conf_overrides: dhcp_agents_per_network11:33
CeeMacOh, no I dont do that11:34
admin0is metadata service provided by one of the dhcp agents ?11:34
CeeMacI thought there were specific metadata agents11:35
CeeMacStill wortg checking nova logs to see what's occurring during the deployment phase11:35
admin0deploying a new vm to check this out11:39
admin0using anfi-affnity so that i can spawn one in every host11:39
admin0which i did before and all worked fine11:39
admin0so it seems to be not affecting every node11:40
admin0maybe its a specific compute => network node combo11:40
CeeMacCould be one host having issues with its agents?11:40
admin0is there a way to validate if a metadata service is working correctly ?11:44
jrosseryou go in the active router network namespace and you will see an instance of haproxy running bound to 169.254.169.25411:45
admin0ok .. so what could cause a new route in the routing table pointing 169.254.169.254 to a dhcp instance -- and also 1 exist for router11:51
admin0i meant 2 routing entries11:52
*** LowKey has quit IRC12:00
*** LowKey has joined #openstack-ansible12:00
CeeMacWhat do you see on other tenant networks?12:17
admin0i am first going to limit the number of network nodes12:23
admin0so that  i can isolate this12:23
admin0then delete all workload and then redo it again12:23
*** waxfire9 has joined #openstack-ansible12:52
*** waxfire has quit IRC12:53
*** waxfire9 is now known as waxfire12:53
*** waxfire has quit IRC13:35
*** waxfire has joined #openstack-ansible13:35
openstackgerritMerged openstack/openstack-ansible master: Use zuul provided requirements repo during bootstrap-ansible.sh  https://review.opendev.org/c/openstack/openstack-ansible/+/77444413:45
*** SiavashSardari has joined #openstack-ansible13:58
*** pto has joined #openstack-ansible14:25
admin0i had some hosts configured as network hosts .. now I don't want them to be network hosts . how shoud I remove them ? i delete the agents .. btu they come back up14:26
CeeMacI think the agents have to be stopped / disabled before you delete them14:27
*** LowKey has quit IRC14:28
CeeMacYou'll have to update your config file and prune the inventory I guess too14:29
*** luksky has quit IRC14:29
*** pto has quit IRC14:29
CeeMacNot sure if you'd need to manually remove the netns14:29
CeeMacMight be better purging the host from inventory and redepolying it?14:30
*** waxfire has quit IRC14:31
*** waxfire has joined #openstack-ansible14:31
admin0CeeMac, how do I purge those hosts ?14:40
admin0just delete them ?14:40
admin0they are not controllers14:40
admin0so no chance of ips being mismatch14:40
admin0yeah  .. seems to have workd for the infra14:44
admin0so i had to rm -rf /openstack/venv/neutron  and rerun the playbooks14:44
CeeMacIf you were changing/removing roles on a host I would probably -r them from the inventory, clear out the ansible_facts, update the group mappings in o_u_c.yml then RE deploy / pxe the o/s and run the required playbooks to add it back as a fresh node14:47
CeeMacRemoving the venv doesn't clean up systemctl units etc so you'd be left with bits and bobs lurking around14:48
admin0doing a pxe was not an option .. now what i have to do is manually remove the systectl for l3 agent, metadata, and dhcp15:00
admin0i did the -r from inventory, cleaned up the uc.yml  and rm -rf /etc/neutron and /venv before re-running the playbooks15:00
SiavashSardarihey everyone. I have and old lab setup, today I rebooted the controllers and now most of lxc containers won't start. this is the out put of lxc-start --logfile15:02
SiavashSardarihttp://paste.openstack.org/show/802858/15:02
SiavashSardariany Idea about this issue?15:02
*** tosky has quit IRC15:35
*** luksky has joined #openstack-ansible15:54
admin0SiavashSardari, is it an aio ?15:57
admin0normally if controllers are restarted, lxc containers start up just fine15:57
SiavashSardarinot exactly, but controllers are computes too15:58
SiavashSardarikinda like an ha aio15:58
admin0for aio .. when used with swift/ceph .. i know it does some loopback device mapping that does not go well during reboots and re-running of the playbooks16:03
admin0so i am not sure16:03
admin0what i normally do is create virtual machines, virtual ceph and then run my lab ( and never use an aio )16:03
SiavashSardarididn't know that. Thank you.16:16
*** ioni has quit IRC16:34
*** waxfire0 has joined #openstack-ansible16:45
*** waxfire has quit IRC16:46
*** waxfire0 is now known as waxfire16:46
*** d34dh0r53 has quit IRC17:09
*** SiavashSardari has quit IRC18:16
*** waxfire7 has joined #openstack-ansible18:18
*** waxfire has quit IRC18:20
*** waxfire7 is now known as waxfire18:20
*** lemko has quit IRC18:48
*** lemko has joined #openstack-ansible18:49
*** lemko5 has joined #openstack-ansible19:42
*** lemko has quit IRC19:42
*** lemko5 is now known as lemko19:42
*** tosky has joined #openstack-ansible20:21
*** pto has joined #openstack-ansible20:26
*** pto has quit IRC20:30
*** luksky has quit IRC20:42
*** luksky has joined #openstack-ansible21:01
*** ioni has joined #openstack-ansible21:45
*** luksky has quit IRC22:54
*** dave-mccowan has quit IRC23:41

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!