Tuesday, 2021-01-26

*** luksky has quit IRC00:08
*** cshen has quit IRC00:26
*** pcaruana has quit IRC01:12
*** pcaruana has joined #openstack-ansible01:12
dotnettedHey all - I can ping the br-vlan gateway from the host through the br-vlan interface without issue, but pinging from the namespaced external network results in an unanswered ARP request for the gateway IP - I'm a networking noob, anything obvious I might be missing here? Thanks! http://paste.openstack.org/show/Ci6iMM3MjlSBtbfitFkg/01:40
*** macz_ has quit IRC01:43
*** tinwood has quit IRC02:08
*** tinwood has joined #openstack-ansible02:11
*** macz_ has joined #openstack-ansible02:21
*** cshen has joined #openstack-ansible02:22
*** cshen has quit IRC02:26
*** spatel has joined #openstack-ansible02:27
*** spatel has quit IRC02:29
*** spatel has joined #openstack-ansible02:30
*** macz_ has quit IRC02:30
*** spatel has quit IRC02:31
*** spatel has joined #openstack-ansible02:33
*** cshen has joined #openstack-ansible02:45
*** spatel has quit IRC02:49
*** cshen has quit IRC02:49
*** spatel has joined #openstack-ansible02:55
*** johanssone has quit IRC03:00
*** johanssone has joined #openstack-ansible03:01
*** cshen has joined #openstack-ansible04:45
*** macz_ has joined #openstack-ansible04:47
*** cshen has quit IRC04:50
*** macz_ has quit IRC04:52
*** evrardjp has quit IRC05:33
*** evrardjp has joined #openstack-ansible05:33
*** spatel has quit IRC05:47
*** gyee has quit IRC06:07
*** cshen has joined #openstack-ansible06:30
*** cshen has quit IRC06:35
*** miloa has joined #openstack-ansible06:43
*** SiavashSardari has joined #openstack-ansible06:49
*** spotz has quit IRC06:54
*** antonym has quit IRC06:56
*** antonym has joined #openstack-ansible06:57
*** cshen has joined #openstack-ansible07:01
fricklerdotnetted: seems like you might be stacking two vlans? you have br-vlan on ethxxx.80 and then use eth1.80, likely one of them is too much.07:35
*** SiavashSardari has quit IRC07:42
*** SiavashSardari has joined #openstack-ansible07:46
*** luksky has joined #openstack-ansible07:50
*** luksky has quit IRC07:53
*** rpittau|afk_ is now known as rpittau07:57
SiavashSardarimorning, does anyone have volume.provider.capacity.allocated metric in the OpenStack Ceilometer/Gnocchi? I'm having trouble with volume provider metrics in my setup08:00
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_zun master: Move zun pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_zun/+/77230008:03
*** andrewbonney has joined #openstack-ansible08:19
*** maharg101 has joined #openstack-ansible08:28
noonedeadpunkSiavashSardari: I don't collect it :( But I think I had disk.device.allocation and it was working (except windows machines)08:45
*** tosky has joined #openstack-ansible08:45
jrossermorning08:47
CeeMaco/08:48
SiavashSardarinoonedeadpunk I was thinking about disk.device.allocation but we have to sum all values for all volumes to have an overview about ceph allocations.08:48
SiavashSardarior maybe there is another solution for that.08:49
noonedeadpunkI guess it would be as incorrect as inside ceph because of not using discard properly unless you set volumes to use iscsi mode instead of ide08:50
MickyMan77good morning,08:50
MickyMan77I do get an error when I deploy version 21.2.2..08:51
MickyMan77http://paste.openstack.org/show/801975/08:51
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_ironic master: Move ironic pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_ironic/+/77228408:51
noonedeadpunkbut to get overwiew of ceph alklocation I guess I'd rather use ceph df or smth like that...08:51
SiavashSardarinoonedeadpunk yeah you are right about that. but the problem with ceph df is that ceph images are thin provisioned and while we use e.g. 100G but the allocation might be 1T08:53
noonedeadpunkOk, wait. So you want so see how much you've sold and not how much is currently occupied?09:00
noonedeadpunkBecause allocation I think should be currently occupied space and iirc that's what ceph df shows?09:01
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_murano master: Use the utility host for db setup tasks  https://review.opendev.org/c/openstack/openstack-ansible-os_murano/+/74723609:05
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_murano master: Move murano pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_murano/+/77229209:08
*** jbadiapa has joined #openstack-ansible09:10
SiavashSardarinoonedeadpunk yeah exactly.09:11
fricklerMickyMan77: seems you somehow managed to create multiple placement service records. try "openstack service list" and check for duplicates and remove them09:12
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_gnocchi master: Move gnocchi pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/77228109:12
noonedeadpunksince otherwise you can use volume.size which will be totall size of cinder volume. and it's working perfectly for me09:13
admin0morning \o09:13
admin0noonedeadpunk, we do a query in the db directly to get allocated size vs ceph df to get current size to get an idea on it09:14
SiavashSardariBTW while we are on the subject of metrics, gnocchi role on branch stable/ussuri didn't work for me. I managed to get it running manually with updating tooz and ceilometer pip packages and I forgot to ask here if something should get fixed.09:14
admin0mysql read only queries are fast and direct .. ( less demanding ) then using ceilometer/gnocchi and callign the apis just to get the same data09:15
admin0in my opinion .. so i create a read-only user and run the data directy into a custom dashboard09:15
noonedeadpunkbut you need to follow all db migrations which is kind of burden imo09:16
noonedeadpunkmaintenance of direct db request results in more work that needs to be done if you want it to work reliably09:16
SiavashSardariadmin0 that sounds like a good solution. did you implement that with ansible and have a role/playbook?09:16
noonedeadpunkand all api changed are pretty well documented, with release notes and some term of deprecation09:16
admin0SiavashSardari, its a small php dashboard with direct mysql connection to the openstack db .. and i generate pie charts like   most used images, most used flavor,  vm created/deleted line graphs in the last 24 hours, last week .. etc09:18
noonedeadpunkbtw, I guess to get data from cinder volumes you should have appropriate timerd https://opendev.org/openstack/openstack-ansible-os_cinder/src/branch/master/tasks/main.yml#L118-L12709:18
admin0at some point, i was also thinking of making it open source and maintaining it09:18
admin0it gives stats that openstack does not provide at once .. like projects that are using most ram, cpu, ip addresses, volumes etc09:19
admin0those kind of stats and graphs09:19
noonedeadpunkregarding gnocchi - not sure tbh. maybe it needs fixing, but have no idea atm09:19
noonedeadpunkSiavashSardari: if you have written steps that you used to make it working that would be helpful to check things out09:19
admin0mysql ENTER ( in the util container)  ;      select sum(size) from cinder.volumes where status = 'in-use' and attach_status='attached';09:24
admin0will give you the total size of volumes vs df from the pool = you have the data09:24
SiavashSardarinoonedeadpunk actually the cinder-volume-usage-audit caught my eyes too, but I didn't know what should I expect from it and how to get data from that service. I checked it's log and sounds like it's getting enough data to solve this issue. but I didn't know how to take advantage of that and didn't find anything on docs09:24
noonedeadpunkthat service push useage report into notifications queue that is listened by ceilometer in this case.09:26
*** lkoranda has joined #openstack-ansible09:26
*** SecOpsNinja has joined #openstack-ansible09:34
SiavashSardarinoonedeadpunk about gnocchi, my first problem was with tooz error on get_lock https://github.com/openstack/tooz/blob/stable/ussuri/tooz/drivers/zookeeper.py#L516 which I found out it was resolved on victoria release. then I had another error regarding redis which in kolla they fixed it in09:35
SiavashSardarihttps://review.opendev.org/c/openstack/kolla-ansible/+/668592 . and at last I had an error like https://bugzilla.redhat.com/show_bug.cgi?id=1883733 which was happening in gnocchi 4.3.3.dev (I'm not sure about the version, but it was something like that) which is the version installed by osa. then I upgrade gnocchi to 4.4.0 and then it worked09:35
openstackbugzilla.redhat.com bug 1883733 in gnocchi "Gnocchi API 500 Error with 'TypeError: keys must be a string'" [High,Verified] - Assigned to mmagr09:35
SiavashSardariproperly.09:35
noonedeadpunkSiavashSardari: but we don't have redis at all?09:36
SiavashSardarithe redis part is not osa issue, but we should update the docs about gnocchi with redis in osa docs09:36
noonedeadpunkum?09:36
noonedeadpunkHonestly I'd say we should use zookeeper instead09:37
SiavashSardarihttps://docs.openstack.org/openstack-ansible/latest/user/prod/gnocchi_redis.html09:37
noonedeadpunkoh09:37
SiavashSardarino we use zookeeper but I was testing to use redis as measure store and ceph as aggregate store09:38
noonedeadpunkaha, yes, I got now why redis raised here09:38
noonedeadpunkyeah, we should update docs for sure09:38
SiavashSardarithere was some minor changes which I don't remember right now. I will check them out and upload a batch for that09:40
openstackgerritMerged openstack/openstack-ansible-os_cloudkitty master: Move cloudkitty pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_cloudkitty/+/77227909:42
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Switch gnocchi release to 4.4  https://review.opendev.org/c/openstack/openstack-ansible/+/77248509:44
SiavashSardarinoonedeadpunk I didn't test it but updating gnocchi will update tooz too?09:50
openstackgerritMerged openstack/ansible-role-python_venv_build master: Optimize the link address in docs  https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/57975609:50
noonedeadpunkI think tooz version will go from u-c https://releases.openstack.org/constraints/upper/ussuri09:51
openstackgerritMerged openstack/openstack-ansible-os_aodh master: Fix documentation formatting  https://review.opendev.org/c/openstack/openstack-ansible-os_aodh/+/58701909:51
openstackgerritMerged openstack/openstack-ansible-tests master: Use setuptools constraint from global-requirements-pins rather than u-c  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/77177009:51
openstackgerritMerged openstack/ansible-role-python_venv_build master: Add source code and bugs url link to README  https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/59858009:52
SiavashSardariI'm not sure which version is equivalent of victoria release but I have 2.8.0 and it works fine.09:53
admin0how to allow horizon to upload images ? i forgot that variable .. and 2nd qustion is .. if there is already ceph in the cluster, how to tell swift to use ceph as a backend ?10:10
jrosserif you want to use ceph radosgw for object storage then you don't have swift10:12
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_zun master: Move zun pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_zun/+/77230010:17
*** dtantsur|afk is now known as dtantsur10:18
MickyMan77 frickler: Can I remove the placement service and then rerun the deployment ?10:22
MickyMan77http://paste.openstack.org/show/801978/10:22
fricklerMickyMan77: hmm, wait, looks like I misread your log (and ansible could be more verbose), can you also show the output of "openstack project list"?10:24
frickleradmin0: jrosser: I also stumbled about horizon failing to upload images, they always end up in state "queued" and horizon pops up an error. is that a known issue?10:26
admin0frickler, there needs to be a variable set  in user_variables .. which is not  set by default ..  some legacy mode .. but i forgot it10:26
MickyMan77 frickler: http://paste.openstack.org/show/801980/10:28
MickyMan77the first "openstack service list" output was from wrong farm... :)10:29
jrosserMickyMan77: you have somehow got two 'service' projects, thats not right10:30
frickleradmin0: HORIZON_IMAGES_UPLOAD_MODE can be legacy or direct, but I'm wondering why the default isn't a working setup10:30
MickyMan77Yes, I did see that..10:30
admin0frickler, the default is that it does not work and need to use the cli10:33
admin0why exactly, not sure10:33
fricklerbad user experience for sure10:33
MickyMan77I think that is old settings for the LDAP connection... I will remove one of them.10:33
openstackgerritSiavash Sardari proposed openstack/openstack-ansible master: Update Gnocchi with Ceph and Redis documentation.  https://review.opendev.org/c/openstack/openstack-ansible/+/77249410:34
fricklerMickyMan77: be careful to remove the correct one, otherwise your whole setup might break10:34
MickyMan77yes. I will remove that one that was use for ldap...10:35
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Install rsync for CentOS bootstrap  https://review.opendev.org/c/openstack/openstack-ansible/+/77249510:40
openstackgerritMerged openstack/openstack-ansible-os_masakari master: Move masakari pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_masakari/+/77228810:43
openstackgerritMerged openstack/openstack-ansible-os_blazar master: Move blazar pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_blazar/+/77226110:44
openstackgerritMerged openstack/openstack-ansible-os_designate master: Move designate pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_designate/+/77228010:44
openstackgerritMerged openstack/openstack-ansible-os_senlin master: Move senlin pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_senlin/+/77229610:47
openstackgerritMerged openstack/openstack-ansible-os_swift master: Move swift packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_swift/+/77229710:50
jrosserfrickler: if you are able to help improve any of these user experience things it would be awesome10:50
openstackgerritMerged openstack/openstack-ansible-os_sahara master: Move sahara pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_sahara/+/77229510:52
*** SiavashSardari has quit IRC11:07
*** rpittau is now known as rpittau|bbl11:09
*** SiavashSardari has joined #openstack-ansible11:18
*** masterpe has quit IRC11:19
*** ioni has quit IRC11:19
*** csmart has quit IRC11:19
*** manti has quit IRC11:19
*** fridtjof[m] has quit IRC11:20
noonedeadpunkjrosser: shouldn't we also merge https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/771195 ?11:24
jrossernoonedeadpunk: yes we should11:26
noonedeadpunkit seems it was working....11:26
*** masterpe has joined #openstack-ansible11:31
jrosserfeels like some effort needed on a few roles though, unrelated to the pip changes11:32
openstackgerritMerged openstack/openstack-ansible-os_neutron master: Prevent neutron-l3-agent killing keepalived on restart  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/77179111:33
jrossershould we backport that ^^11:37
*** fridtjof[m] has joined #openstack-ansible11:58
*** ioni has joined #openstack-ansible11:58
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible-os_neutron master: L3 agent cleanup_on_shutdown  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/77182611:58
*** csmart has joined #openstack-ansible11:58
*** manti has joined #openstack-ansible11:58
noonedeadpunkI'd say yes11:58
noonedeadpunkat least to V11:59
noonedeadpunkmurano needs a lot I think and not sure if ipv6 solved octavia issue?12:00
jrosseri think the ipv6 thing just revealed another problem12:00
jrosseri was wondering if this existing setting is wrong https://github.com/openstack/openstack-ansible/blob/master/tests/roles/bootstrap-host/templates/user_variables_octavia.yml.j2#L3012:01
jrosserironic something is broken too12:01
noonedeadpunkoh, yes, but it's broken on tempestconf12:02
noonedeadpunkwell, at least setting is present https://github.com/openstack/octavia-tempest-plugin/blob/master/octavia_tempest_plugin/config.py#L14312:03
noonedeadpunkoh, you mean format...12:04
noonedeadpunkit's wrong :(12:04
noonedeadpunkit should be just `test_with_ipv6: false` indeed12:04
jrosseryeah, so my patch may be not needed12:06
*** jbadiapa has quit IRC12:06
jrosserwell depends, if we want optional ipv6 tests or no ipv6 tests12:06
noonedeadpunkit should adjust just different thing :)12:06
noonedeadpunkI'd say we can live without ipv612:07
jrosseri was surprised tbh that it worked at all12:07
noonedeadpunkwe should ensure that lb is functional and that's it I guess12:07
noonedeadpunkmaybe it was valid at some point... unlikely though12:07
openstackgerritJonathan Rosser proposed openstack/openstack-ansible master: Disable octavia ipv6 tempest test when the node does not have an ipv6 address  https://review.opendev.org/c/openstack/openstack-ansible/+/77237912:09
openstackgerritJonathan Rosser proposed openstack/openstack-ansible master: Disable octavia ipv6 tempest test by fixing config error  https://review.opendev.org/c/openstack/openstack-ansible/+/77237912:09
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Install rsync for CentOS bootstrap  https://review.opendev.org/c/openstack/openstack-ansible/+/77249512:31
openstackgerritDmitriy Rabotyagov proposed openstack/openstack-ansible master: Remove note about metal/horizon compatability  https://review.opendev.org/c/openstack/openstack-ansible/+/77157312:37
jrosserironic is still broken with tempestconf12:45
jrosserit now does use the internal endpoint to get the service catalog, but then it uses the public endpoints from the catalog12:45
jrosserand fails again with SSL error in CI12:46
jrosserthere seems to be a patch https://opendev.org/osf/python-tempestconf/commit/f146f810695e83d2a8ce0fcdb94ff32e75ebdb2012:46
jrosserbut i do really fail to understand, --insecure passed where......12:47
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_tempest master: Use internal endpoint for tempestconf and respect tempest_keystone_interface_insecure  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/76996612:58
*** rpittau|bbl is now known as rpittau12:59
*** jbadiapa has joined #openstack-ansible13:03
*** rh-jelabarre has joined #openstack-ansible13:04
*** macz_ has joined #openstack-ansible13:45
*** macz_ has quit IRC13:50
openstackgerritMerged openstack/openstack-ansible-galera_server master: Bring db setup vars in line with other roles  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/77165213:57
*** spatel has joined #openstack-ansible14:02
*** MickyMan77 has quit IRC14:02
*** macz_ has joined #openstack-ansible14:06
*** MickyMan77 has joined #openstack-ansible14:08
*** macz_ has quit IRC14:11
*** SiavashSardari has quit IRC14:11
spatelnoonedeadpunk: or jrosser I am dealing with very bizarre issue with RC1 release, whenever i add new compute node and run playbook os-nova-install.yml or os-neutron-install.yml i hit this error http://paste.openstack.org/show/801986/14:11
spatelDo you guys have any history around this issue?14:12
spatelworkaround is to re-run -e venv_rebuild=yes14:12
noonedeadpunkmy guess would be that something drops your contraints from repo container14:12
noonedeadpunkmaybe 2 running lsyncd on different containers?14:12
jrosseri would say some sort of partial upgrade odd behaviour, look at the versions14:13
jrossernova==22.0.2.dev5 and your venv is /openstack/venvs/nova-22.0.0.0rc114:13
spatelhmm, i thought about that so i remove 2 repo and kept just single repo in LB14:13
jrosserrelease candidate vs. stable/victoria as a guess14:14
spateljrosser: i am also confused about that so not sure from where this is coming14:14
jrosserdid you forget to bootstrap-ansible.sh after moving the branch forward, something like that?14:14
spateli am 100% sure i checkout tag when i was deploying OSA and never touch it after that14:15
jrosserwhat SHA do you have in openstack-ansible repo?14:15
spatellet me check14:16
*** dotnetted_alt has joined #openstack-ansible14:16
*** dotnetted has quit IRC14:17
spatelgit log showing commit ac43c0ce9442184c7041146c6d4fddd1fcbf9d36 (HEAD, tag: 22.0.0.0rc1)14:17
spatelis that what you asking?14:18
admin0from which variable can i update/overrride the default lxcbr0 ip range used in openstack to something else ( that is not used in the network ) that is conflicting14:19
jrosserspatel: try /opt/ansible-runtime/bin/python setup.py --version14:20
spatel22.0.0.0rc114:20
jrosserhmm14:20
*** sshnaidm|ruck is now known as sshnaidm|afk14:20
spatelit always works with  -e venv_rebuild=yes14:21
spatellet me check lsyncd stuff14:24
jrosseralso check the wheel build log on the repo server to see what actual version of the nova wheel it built, and check the constraints file that is shown in your paste too that calls it up14:26
spatelnoonedeadpunk: hey, does lsyncd run on all 3 repo container right?14:26
jrosseronly on the first iirc14:26
spatelthat is what i am seeing its only running on first container14:26
jrosserthe wheel build should happen on the first, and the results be then rsync out to the rest14:26
noonedeadpunkmaybe build host !=lsyncd host?14:27
jrosserthat would certainly make things go odd14:27
jrosserspatel: check the ansible log for which host is delegated to during python_venv_build role wheel builds14:28
jrossermake sure that is the same on as you see lsyncd running on14:28
spatelin my LB repo-1 has high priority and that one is running lsyncd service14:28
spatelchecking logs on repo-114:29
jrosserit's not really to do with the loadbalancer14:29
spatelif you want i can disable other repo-2 and 314:29
spatelOk i got what you saying its on playbook to pick repo14:29
jrosserthere are two sides to it, wheels are built in one place and should be rsynced to the others14:29
jrosserthen all repo servers have the content to serve via the LB14:30
spatelshould i run playbook with -vvvv to find out which repo its picking also tailing logs on repo to see what is going on /var/log/python_wheel_build.log14:31
spateljrosser: on repo-2/3 file python_wheel_build.log doesn't existing so its clear repo-1 building wheels14:32
jrosserok, then you should check the lsyncd log on repo-114:33
spatelso far no error but let me tail logs and run playbook to see14:35
spateljrosser: tell me one thing, why lsyncd is important here if i shutdown repo-2 and repo-3 ?14:35
spatelthen lsyncd has no role right?14:35
jrossertrue14:35
spatelif i disable repo-2/3 in my F5 then no request will land there also.14:36
spateljrosser: my os-releases/22.0.0.0rc1/nova-22.0.0.0rc1-constraints.txt has nova==22.0.2.dev514:38
spateldoes that normal to you?14:38
jrosserspatel: the latest tag on nova is https://github.com/openstack/nova/tags 22.0.114:39
jrosserso it is reasonable that a commit ahead of that is 22.0.2.dev<N-commits-ahead>14:40
spatelin my lab i am running 22.0.0.0b2 and i am not seeing any issue there, its working fine.14:40
jrossernow you should check the pythin_venv_build log on the host that failed, there will be lots more detail than the ansible log14:41
spatelLet me find fresh file of pythin_venv_build.log from compute host14:41
*** dave-mccowan has joined #openstack-ansible14:44
*** dave-mccowan has quit IRC14:46
spateljrosser: whole logs files look good so far and in end it barf with this error - http://paste.openstack.org/show/801989/14:55
spatelERROR: Could not find a version that satisfies the requirement nova==22.0.2.dev514:56
spatelvery odd that its saying not able to find so where actually its trying to find, does it searching in some kind of array to match?15:01
spatelhmm! if you check logs Found link https: is searching for each nova version one by one15:04
spateland matching hash15:05
jrosseryes you have to read through the log carefully and see if it finds the one from the repo server15:05
jrosserthen try to figure out why pip decided that was not a suitable candidate15:05
jrosseryou should have requirements/constraints files dropped on the compute host which are inputs to pip15:06
spatelthat file isn't there - /openstack/venvs/nova-22.0.0.0rc1/15:07
spatelplaybook deleting itself when build fails15:07
spatelit was there while build was running15:07
*** sshnaidm|afk is now known as sshnaidm|ruck15:07
spatelits interesting venv_rebuild=yes not complaining about anything and it works15:08
MickyMan77Hi all,  I have a panic issue. the galera-container is not able to start up mysql...15:09
spatelpip is trying to find 22.0.2 which is not existing anywhere because latest version is 22.0.115:10
spatelhow that version endup in nova-22.0.0.0rc1-constraints.txt15:11
openstackgerritAndrew Bonney proposed openstack/openstack-ansible-os_neutron stable/victoria: Prevent neutron-l3-agent killing keepalived on restart  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/77253815:20
openstackgerritMerged openstack/openstack-ansible-os_trove master: Move trove pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_trove/+/77229915:21
openstackgerritAndrew Bonney proposed openstack/openstack-ansible-os_octavia master: Add variable to override keypair setup python interpreter  https://review.opendev.org/c/openstack/openstack-ansible-os_octavia/+/77255915:23
*** gary_perkins has quit IRC15:28
jrosserspatel: the python_venv_build role creates those constraints files, and the contents is based on the actual wheels that got built15:29
jrosserso if a wheel for 22.0.2.dev5 was built, thats what gets written to the constraints15:29
openstackgerritAndrew Bonney proposed openstack/openstack-ansible master: Remove localhost references from galera docs following bind change  https://review.opendev.org/c/openstack/openstack-ansible/+/77256015:29
jrossercomment out the tasks that delete the constraints files on venv build failure, then take a look at them15:30
spatelsounds good15:30
spatellet me try that15:30
spateli thought playbook download  constraints file directly from repo server so it should be same what we have on repo15:31
openstackgerritDmitriy Rabotyagov proposed openstack/ansible-hardening master: Make possible to avoid aide installation  https://review.opendev.org/c/openstack/ansible-hardening/+/77256115:32
openstackgerritMerged openstack/openstack-ansible-os_rally master: Ensure that the rally venv build is self contained  https://review.opendev.org/c/openstack/openstack-ansible-os_rally/+/77142115:33
spateljrosser: i can see all 3 file on compute machine constraints.txt, global-constraints.txt & requirements.txt and they are similar to what we have on repo server15:41
openstackgerritMerged openstack/openstack-ansible-os_magnum master: Move magnum pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_magnum/+/77228515:58
noonedeadpunk#startmeeting openstack_ansible_meeting16:00
openstackMeeting started Tue Jan 26 16:00:43 2021 UTC and is due to finish in 60 minutes.  The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot.16:00
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.16:00
*** openstack changes topic to " (Meeting topic: openstack_ansible_meeting)"16:00
openstackThe meeting name has been set to 'openstack_ansible_meeting'16:00
noonedeadpunk#topic office hours16:00
*** openstack changes topic to "office hours (Meeting topic: openstack_ansible_meeting)"16:01
noonedeadpunko/16:01
jrosserhello16:01
*** macz_ has joined #openstack-ansible16:02
noonedeadpunkSo, regarding renos. First of all we need to stop publishing new ones. I placed bunch of PRs to cover that. Once that will be done, I will go to infra team regarding removing already published ones16:03
*** macz_ has joined #openstack-ansible16:03
noonedeadpunkwhich means that we can just abandon changes like https://review.opendev.org/c/openstack/openstack-ansible-os_mistral/+/76866316:03
jrosseryeah, we should do that.... the only one (if its there?) which should stay is on openstack-ansible i guess16:04
*** gshippey has joined #openstack-ansible16:04
noonedeadpunkyeah. Also I'm not sure if we should leave for ansible-hardening...16:04
noonedeadpunkI've pushed PR but now I'm not so sure16:05
noonedeadpunkconsidering repo is also tagless, then probably yes...16:05
jrosserok well -W any you are not sure about16:05
noonedeadpunkthat was the only one I think.16:07
jrosseri think i may have figured out this tempestconf stuff16:07
noonedeadpunkthey are kind of broken anyway there16:07
noonedeadpunkoh, rly? I tried to push some patch but it was not the cause https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/76996616:08
noonedeadpunkor it ws not the only one16:08
jrosseryes i fiddled with that today16:08
noonedeadpunkaha, yes, just noticed that16:09
jrosserjust seems i make silly error with the command16:09
jrosser[Errno 2] No such file or directory: '/root/workspace/etc/profile.yaml --insecure False'16:09
jrosserlike it takes the whole thing as the --profile parameter, which is strange16:10
*** jamesdenton has quit IRC16:11
*** jamesden_ has joined #openstack-ansible16:11
noonedeadpunkmaybe it's because of ""16:11
jrosseralso https://opendev.org/openstack/openstack-ansible-os_tempest/src/branch/master/defaults/main.yml#L228 kind of not what i expected either16:11
noonedeadpunksince they start before profile and end at the end16:11
jrosserthere must be a better variable for that16:12
noonedeadpunkwe use keystone_service_internaluri_insecure everywhere16:13
noonedeadpunkor what do you mean&16:13
jrosseri think i need a variable that talks about the external IP16:13
noonedeadpunkI think we should just make tempestconf use internalurl for interaction?16:14
jrosserwell, your patch made it do that16:14
jrosserit gets the service catalog from the internal endpoint16:15
jrosserthen uses the public entries in the catalog itself by the look of it16:15
noonedeadpunkuh....16:15
jrosseri think that may be what is happening here http://paste.openstack.org/show/801994/16:16
noonedeadpunkthere sould be some extra I guess to select internal instead of public16:16
noonedeadpunkactually the same issue spatel reported about senlin16:16
noonedeadpunkthat it connects to keystone through internal but takes public endpoints from catalog16:16
jrosseri can't otherwise see how it has discovered the IP of eth016:17
spatelYes, its known issue and may required senlin code change16:17
noonedeadpunkI'm pretty sure you're right about tempestconf picking up public endpoint16:18
jrosseranyway, my hope was that adding --insecure would make it not worry about the certificate16:18
jrosseri think it may be legitimate behaviour as tempest kind of pretends to be an end user16:18
noonedeadpunkyeah16:18
noonedeadpunkwill see...16:18
spatelsenlin acting like end-user and using all public endpoint by default16:18
jrossertheres is two things, making the tempestconf discovery not worry about the cert16:18
jrosserthen it has to propery write out a tempest.conf that *also* doesnt worry about the cert for the actual tests16:19
jrosserseems only recently this is possible https://opendev.org/osf/python-tempestconf/commit/f146f810695e83d2a8ce0fcdb94ff32e75ebdb2016:19
noonedeadpunkok, we can set verify: false16:20
*** lkoranda has quit IRC16:20
jrosserdo you know where we would do that?16:21
jrosserthe documentation for this is kind of sparse16:21
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_tempest master: Use internal endpoint for tempestconf and respect tempest_keystone_interface_insecure  https://review.opendev.org/c/openstack/openstack-ansible-os_tempest/+/76996616:22
noonedeadpunkmaybe add to tempest_tempestconf_profile ?16:22
jrosseras far as i can see you can't specify this in the profile16:23
jrosseri think that this must be picking up a clouds.yaml from somewhere16:24
jrosseranyway, we maybe should move on a bit?16:25
noonedeadpunkwhere does it takes auth creds from?16:25
noonedeadpunkclouds.yaml?16:25
noonedeadpunkyeah, let's move16:25
jrosserandrewbonney is doing a U->V upgrade in our lab today16:25
noonedeadpunkWe've done T->V in lab and it went beautiful16:26
noonedeadpunkPlanning do it in prod next week16:26
jrosserwe have a bunch of issues and will make some patches16:26
jrosseractually some already done16:26
jrossermaybe most surprisng was something during setup-hosts restarting all the api containers at the same time16:27
MickyMan77jrosser: can you help out, I'm not able to start, stop or restart the mysql in the galera_container. It just hanging.. I can't found any log about the issue...16:27
jrossersome releasenotes may be missing too, like bind-to-mgmt means rabbitmq containers need /etc/hosts fixing up a bit16:29
openstackgerritGaudenz Steinlin proposed openstack/openstack-ansible-os_cinder stable/ussuri: Define credentials for nova interaction  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/77253916:29
openstackgerritMerged openstack/openstack-ansible master: Use TCP mode for console if SSL is configured  https://review.opendev.org/c/openstack/openstack-ansible/+/57415316:30
jrosserdid we have a bug / explanation for the designate pool UUID issue?16:30
andrewbonneyjrosser:  fwiw our designate issue is this one: https://bugs.launchpad.net/designate/+bug/189793616:31
openstackLaunchpad bug 1897936 in Designate "Pool update fails when zones exist" [Undecided,Fix released] - Assigned to Mark Goddard (mgoddard)16:31
noonedeadpunkNo, I can't remeber if we have one16:31
noonedeadpunkbut it was quite silly that we used config param that was not preset in designate anymore16:32
noonedeadpunkwhat caused containers to restart?16:32
noonedeadpunkhave you upgraded to focal?16:33
admin0MickyMan77, is the container broken or only mysql ?16:33
spatelI thought designate issue has been resolved, i have deployed couple of time with UUID patch and no issue so far16:33
admin0if container is broken, is it only 1 container, or 3 containers ( if you are in HA setup)16:33
jrosserandrewbonney: looks like that patch was reverted https://review.opendev.org/c/openstack/designate/+/75542916:33
noonedeadpunkbtw is it smth you mentioned about rabbitmq? https://bugs.launchpad.net/openstack-ansible/+bug/182485716:35
openstackLaunchpad bug 1824857 in openstack-ansible "Rabbitmq join cluster fail" [Undecided,New]16:35
openstackgerritMerged openstack/openstack-ansible-apt_package_pinning master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-apt_package_pinning/+/77200716:36
jrosseryes, as part of bind-to-mgmt we had 3 patches about that16:36
*** gary_perkins has joined #openstack-ansible16:36
noonedeadpunkhave we missed smth from V?16:36
prometheanfirehas anyone upgraded from ussuri to victoria?16:37
jrossernoonedeadpunk: currently this has to be dealt with manually https://review.opendev.org/c/openstack/openstack-ansible-openstack_hosts/+/670706 https://review.opendev.org/c/openstack/openstack-ansible-lxc_container_create/+/670705 https://review.opendev.org/c/openstack/openstack-ansible/+/67039216:37
jrosserprometheanfire: we discuss this right now for the last 10 mins :)16:38
*** gyee has joined #openstack-ansible16:40
noonedeadpunkI think https://review.opendev.org/c/openstack/openstack-ansible-openstack_hosts/+/670706 should not be relevant now?16:40
noonedeadpunkor dunno...16:40
noonedeadpunknot quite understand why issue is raised now...16:40
jrosseri think the activation of bind-to-mgmt across everything comes in for V16:41
noonedeadpunkwe now generate hosts file quite differently with blockinfile16:42
jrosserthat is true16:42
jrosserdo you have a link to that patch?16:42
noonedeadpunkso we just need to drop everything except this block?16:42
noonedeadpunkhttps://opendev.org/openstack/openstack-ansible-openstack_hosts/commit/c64e1caf72c20a2ffcce7b1d92e8b8cc8093a80816:42
jrosserright so even blockinfile won't remove stuff thats wrong16:42
jrosseri think our issue was the hosts files are long lived and had accumulated stuff from many releases16:43
noonedeadpunkthat';s interesting. I think I might have exactly the same issue16:44
prometheanfirejrosser: good timing :D16:45
noonedeadpunkso eventually dropping all except generated block should help it?16:45
jrossermaybe a bit too much to assume the deployer has not put things in there16:46
jrossernot sure OSA 'owns' then whole file16:46
openstackgerritMerged openstack/openstack-ansible-plugins master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/77205416:48
noonedeadpunknah, ofc it's not what I'm thinking to script, but actually how it might be solved _when_ we will face it here:)16:49
jrosseri think it was a trivial fix manually once we figured what was going on16:50
noonedeadpunkyeah...16:52
noonedeadpunkI'm still not sure what caused bug https://bugs.launchpad.net/openstack-ansible/+bug/1824857 since what you said should not be the issue for clen deployments16:53
openstackLaunchpad bug 1824857 in openstack-ansible "Rabbitmq join cluster fail" [Undecided,New]16:53
noonedeadpunkWill try to set up 3 nodes sandbox tomorrow16:53
openstackgerritMerged openstack/openstack-ansible-lxc_container_create master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-lxc_container_create/+/77201316:54
* jrosser just looking at bugs16:54
noonedeadpunkhttps://bugs.launchpad.net/openstack-ansible/+bug/1911482 has come here in IRC and said that it was issue in inventory or smth16:58
openstackLaunchpad bug 1911482 in openstack-ansible "neutron-l3-agent broken after train upgrade" [Undecided,New]16:58
noonedeadpunkI'm wondering if we can move it to incomplete or invalid...16:58
jrossercomment "resolved via IRC" and incomplete sounds good17:00
noonedeadpunkShould we do anything regarding https://bugs.launchpad.net/openstack-ansible/+bug/1877421 ?17:01
openstackLaunchpad bug 1877421 in openstack-ansible "Cinder-volume is not able to recognize a ceph cluster on OpenStack Train." [Undecided,Confirmed]17:01
jrosseroh,17:05
*** rh-jelabarre has quit IRC17:06
jrosserwasnt there something magical about RBD, in that you couldnt ever have ceph@RBD becasue it's some kind of keyword?17:06
jrosserthere was a huge long irc thread about this some time ago17:06
noonedeadpunkyeah, I can recall smth like that17:06
jrosseras soon as you use ceph@some-other-pool-name it's all ok17:06
noonedeadpunkI'm wondering if we have defined smth like this in docs?17:07
noonedeadpunkoh....17:08
noonedeadpunk#endmeeting17:08
*** openstack changes topic to "Launchpad: https://launchpad.net/openstack-ansible || Weekly Meetings: https://wiki.openstack.org/wiki/Meetings/openstack-ansible || Review Dashboard: http://bit.ly/osa-review-board-v3"17:08
openstackMeeting ended Tue Jan 26 17:08:16 2021 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)17:08
openstackMinutes:        http://eavesdrop.openstack.org/meetings/openstack_ansible_meeting/2021/openstack_ansible_meeting.2021-01-26-16.00.html17:08
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/openstack_ansible_meeting/2021/openstack_ansible_meeting.2021-01-26-16.00.txt17:08
openstackLog:            http://eavesdrop.openstack.org/meetings/openstack_ansible_meeting/2021/openstack_ansible_meeting.2021-01-26-16.00.log.html17:08
noonedeadpunkyeah, we have that in docs17:09
noonedeadpunkbut we have it defined in aio as well...17:09
openstackgerritMerged openstack/openstack-ansible-os_cloudkitty master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_cloudkitty/+/77202617:09
noonedeadpunkhttps://opendev.org/openstack/openstack-ansible/src/branch/master/tests/roles/bootstrap-host/tasks/prepare_ceph.yml#L93-L10217:10
jrosseri wonder how that works in CI17:11
noonedeadpunkor you can't name it as "ceph"...17:11
noonedeadpunkuh17:12
jrossermine are rbd_<storagetype> with hdd/ssd/nvme....17:13
jrosserin cinder_backends17:13
noonedeadpunkI also have SSD/HDD/STOR17:14
noonedeadpunkebbex: maybe you have some newer context?17:15
noonedeadpunkfor https://bugs.launchpad.net/openstack-ansible/+bug/187742117:15
openstackLaunchpad bug 1877421 in openstack-ansible "Cinder-volume is not able to recognize a ceph cluster on OpenStack Train." [Undecided,Confirmed]17:15
openstackgerritMerged openstack/openstack-ansible-tests master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/77206017:19
openstackgerritJonathan Rosser proposed openstack/openstack-ansible-os_horizon master: Move horizon pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_horizon/+/77028317:20
*** rh-jelabarre has joined #openstack-ansible17:22
*** rh-jelabarre has quit IRC17:23
*** rh-jelabarre has joined #openstack-ansible17:23
*** rpittau is now known as rpittau|afk17:27
spatelnoonedeadpunk: what -e venv_rebuild=yes does behind the scenes ?17:27
spatelI think i am done with debug, not able to find why its not able to find version but when i use venv_rebuild it does find version17:28
spatelor better i should upgrade RC1 to RC2 ?17:30
jrossercan you paste the relevant part of the python_venv_build log from the compute host?17:31
noonedeadpunkvenv_rebuild force re-creates contriaints files on repo containers and rebuilds wheels and recreate venvs17:31
spatelYes let me do that, log file is very very log so let me find out how i can share17:35
jrosseronly the last part where it tries and fails to do nova17:36
jrosserthe sections should be clear where it deals with each package17:36
spatelok17:36
spateljrosser: here is the sample - http://paste.openstack.org/show/802008/17:40
spatelI am having same issue with nova/neutron both (both saying not able to find version but if i run venv_rebuild=yes works)17:41
noonedeadpunkdo you have wheel built up in /var/www/repo/os-releases/22.0.0.0rc1/ubuntu-20.04-x86_64/ inside repo containers?17:44
spatellet me show you content of that folder17:45
jrosseralso do wget on http://10.65.0.16:8181/os-releases/22.0.0.0rc1/ubuntu-20.04-x86_6417:45
spatelhttp://paste.openstack.org/show/802009/17:45
jrosserand see if the thing that it says is not available (like neutron==17.1.0.dev119) is listed in what you get back17:45
spatelyes i can wget all file from that repo17:46
noonedeadpunkhm, I think it should containe neutron, but it's not17:46
jrosserin your last paste there is no...17:46
jrosser^that17:46
spateli don't have anything start with neutron* in ubuntu-20.04-x86_64 foler17:46
jrosserso something is wrong with the wheel build17:46
openstackgerritMerged openstack/ansible-role-python_venv_build master: Replace pip requirement source specs with the package name for built wheels  https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/77119517:47
jrosserthe error from the venv build is absolutely correct17:47
spatelhmm17:47
jrosserfrom the POV of pip on the compute node, the wheel of that version is not availble at the places its been given to look17:47
noonedeadpunkactually these dev can't be found anywhere if there are no wheels built17:48
jrosserspatel: on the repo server, do you have /var/log/python_wheel_build.log?17:48
spatelyes17:49
spatelits very long and big17:49
jrosserok, so just rename that file to something new17:49
spateldo you want me to remove it and re-run playbook with fresh error?17:49
spatelok17:49
jrosserthen re-run just one of your failing things17:49
spatelon it17:49
spatelrunning playbook17:50
jrosseryou should see the python_venv_build role do the wheel build17:51
jrossersee if those tasks do anything, or are all skipped17:51
jrosserif they do something, the task header will show if they are delegated to another host (hopefully the repo server)17:51
*** maharg101 has quit IRC17:55
spatelthat log file is massive17:55
spateli trying to find keyword skipped and found nothing17:56
spatelhold on damn it i was looking into python_venv_build.log file not wheel :)17:57
jrosserlets say you re-ran nova playbook17:57
spateljrosser: it didn't created any python_wheel_build.log file17:57
spatelI am on repo-117:57
spatelvery odd17:58
jrosserright, so check the ansible log if the wheel build steps were skipped, or where executed17:58
openstackgerritMerged openstack/openstack-ansible-memcached_server master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-memcached_server/+/77201517:58
spateljrosser: here its full output of playbook - http://paste.openstack.org/show/802013/17:59
jrosserTASK [python_venv_build : Build wheels for the packages to be installed into the venv] ***********************************************************************17:59
jrosserskipping: [ostack-phx-comp-sriov-1-7]17:59
jrosser^ there17:59
jrosserdid you do -e venv_rebuild=true?18:00
jrosserwhich if i recall, is when it works?18:01
jrosserspatel: i have to go, but you see what we are doing18:02
jrosserstart with the error message and work back18:02
spatelthanks let me see what i can do18:02
spatellatest connect  later18:03
openstackgerritMerged openstack/openstack-ansible-openstack_openrc master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-openstack_openrc/+/77201918:04
openstackgerritMerged openstack/openstack-ansible-os_heat master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_heat/+/77203118:04
openstackgerritMerged openstack/openstack-ansible-os_keystone master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_keystone/+/77203418:11
openstackgerritMerged openstack/openstack-ansible-rabbitmq_server master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-rabbitmq_server/+/77205518:11
openstackgerritMerged openstack/openstack-ansible-rsyslog_client stable/victoria: Remove examples for variables which no longer exist  https://review.opendev.org/c/openstack/openstack-ansible-rsyslog_client/+/77222318:12
openstackgerritMerged openstack/openstack-ansible-os_placement master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_placement/+/77204418:14
openstackgerritMerged openstack/openstack-ansible-galera_server master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/77201218:19
openstackgerritMerged openstack/openstack-ansible-os_masakari master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_masakari/+/77203718:19
openstackgerritMerged openstack/openstack-ansible-os_rally master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_rally/+/77204518:20
openstackgerritMerged openstack/openstack-ansible-repo_server master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-repo_server/+/77205618:20
openstackgerritMerged openstack/openstack-ansible-rsyslog_client master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-rsyslog_client/+/77205718:20
spateljrosser: i didn't use -e venv_rebuild=true?  so question is why Build wheels for the packages to be installed into the venv getting skipped18:23
openstackgerritMerged openstack/ansible-role-systemd_networkd master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/ansible-role-systemd_networkd/+/77200518:26
openstackgerritMerged openstack/openstack-ansible-os_blazar master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_blazar/+/77202118:26
openstackgerritMerged openstack/openstack-ansible-os_glance master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_glance/+/77202818:26
openstackgerritMerged openstack/openstack-ansible-os_mistral master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_mistral/+/77203818:26
openstackgerritMerged openstack/openstack-ansible-os_swift master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_swift/+/77204918:26
openstackgerritMerged openstack/openstack-ansible-os_magnum master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_magnum/+/77203518:32
openstackgerritMerged openstack/openstack-ansible-os_manila master: Move manila pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_manila/+/77228718:32
openstackgerritMerged openstack/openstack-ansible-os_sahara master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_sahara/+/77204818:32
openstackgerritMerged openstack/openstack-ansible-os_barbican master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_barbican/+/77202218:39
openstackgerritMerged openstack/openstack-ansible-os_ceilometer master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_ceilometer/+/77202418:39
openstackgerritMerged openstack/openstack-ansible-os_gnocchi master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/77203018:39
openstackgerritMerged openstack/openstack-ansible-os_tacker master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_tacker/+/77205018:39
openstackgerritMerged openstack/openstack-ansible-os_horizon master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_horizon/+/77203218:46
openstackgerritMerged openstack/openstack-ansible-os_mistral master: Move mistral pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_mistral/+/77229118:51
openstackgerritMerged openstack/openstack-ansible-os_nova master: [reno] Stop publishing release notes  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/77204118:52
*** miloa has quit IRC18:53
prometheanfirewhy are release notes stopping?18:55
prometheanfireah, integrated reno18:56
prometheanfireI looked at backlog and didn't see U->V upgrade conversations, was it the rabbit cluster join issue?18:56
openstackgerritMerged openstack/openstack-ansible-os_neutron master: L3 agent cleanup_on_shutdown  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/77182619:13
*** andrewbonney has quit IRC19:16
openstackgerritMerged openstack/openstack-ansible-os_barbican master: Move barbican pip packages from constraints to requirements  https://review.opendev.org/c/openstack/openstack-ansible-os_barbican/+/77226019:22
spatelnoonedeadpunk: quick question i have this "neutron-17.1.0.dev119-py3-none-any.whl" in /var/www/repo/os-releases/22.0.0.0rc1/ubuntu-20.04-x86_6419:25
*** cloudnull has quit IRC19:26
spateldoes os-neutron-install.yml remove this file before running?19:26
*** cloudnull has joined #openstack-ansible19:26
noonedeadpunkno, unless you run with venv_rebuild19:37
spatelhmm!19:41
spatellet me run some experiment to understand why that file disappearing19:41
spatelis it possible lsyncd is removing ?19:42
noonedeadpunkyes, totally19:48
noonedeadpunkthat file should be on the container where lsyncd is running. if it on some other host, the lsyncd will drop it19:49
*** maharg101 has joined #openstack-ansible19:51
spatelhmm19:53
spatellet me check19:54
spatelnoonedeadpunk: i do have neutron-17.1.0.dev119-py3-none-any.whl in all repo19:56
spatelthat is how lsyncd works right19:56
*** maharg101 has quit IRC19:56
spatelit copy files from repo-1 to 2 and 319:56
noonedeadpunkyep19:56
noonedeadpunkbut it should not be deleted19:56
spatelhmm! let me keep digging... i think i am slowly understanding how these nuts and bolts working here :)19:57
noonedeadpunkhaha19:58
spatelnoonedeadpunk: are we going to release rc3 or stable ?20:02
spatelhow much work is pending for victoria?20:02
spatelI was thinking to do minor upgrade from rc1 to rc2 or wait for stable one20:02
*** simondodsley has quit IRC20:14
*** hindret has quit IRC20:14
*** hindret has joined #openstack-ansible20:16
jrosserspatel: 22.0.0 is released, i mentioned this the other day20:16
*** simondodsley has joined #openstack-ansible20:16
jrosserrc2 is the same SHA sa 22.0.020:16
jrosser*as20:16
spatelso its pretty much stable release right?20:17
jrosser?20:17
jrosser22.0.0 is the first stable release tag on the stable/victoria branch, after the previous rc120:18
spatelyour answered my question20:18
jrosserthe release process requires the final non-rc release to be at the same point as the last rc20:18
jrosserand as we had merged fixes on top of rc1, we had to make rc2, then put 22.0.0 tag at the same place as rc220:18
spatelgot it so 22.0.0 = rc220:20
jrossercorrect20:20
jrosserit's just a git tag, no magic20:20
spatel:)20:21
spatelI am seeing very strange thing with my issue.. let me explain what is going on20:21
spatelwhen i use -e venv_rebuild=yes its destroying whole folder in repo /var/www/repo/os-releases/22.0.0.0rc1/ubuntu-20.04-x86_6420:22
spateland re-creating again20:22
spatelis that normal behavior?20:22
spatelI think that is what my problem is20:23
spatelSo lets if i run os-nova-install.yml -e venv_rebuild=yes it destroying repo folder and recreating but neutron-* whl package is missing, so next time when i run os-neutron-install.yml -e venv_rebuild=yes it destroying repo folder and re-creating without nova-* whl package20:25
spatelthat is why i am seeing this issue everytime i add new compute node20:26
spateljrosser: ^^20:26
spatelThis is like chicken egg issue20:26
jrosserhttps://github.com/openstack/ansible-role-python_venv_build/blob/master/tasks/python_venv_wheel_build.yml#L43-L5420:32
spatelit does delete folder but is creating issue for me.20:34
spatelnova deleting neutron and neutron deleting nova whl files20:34
jrosserright, and when you run the role next time it should rebuild the wheels anyway20:34
jrosserbecause of this https://github.com/openstack/ansible-role-python_venv_build/blob/master/tasks/python_venv_wheel_build.yml#L12820:35
jrosserif for some reason it does not (like in a paste from earlier) then that condition is not being made true20:35
jrosserso the first link i gave you is where all the files get deleted20:36
jrosserwhen you re-run it, the files are re-created and the condition for the second link (do the wheel build) will be true20:36
jrosserthat should build the wheels and put them in the var20:37
jrosserwww/repo directory20:37
spatelyes20:37
jrosserbut you see on later runs that the wheels are not present20:38
jrosserso either they are not being built on the repo server that you think they are20:38
jrosseror there is some lsyncd/rsync thing where they are getting deleted because the ordering of the repo servers is somehow messed up20:39
jrosseryou will be able to see that perhaps in the lsync log20:39
spatelsee - http://paste.openstack.org/show/802020/20:39
spatelI can see wheels getting builds on repo-1 and then lsyncd copying them to repo-2 and repo-320:40
spatelI can see all contents are synced20:40
spatelcan i stop lsyncd on repo-1 and try ?20:41
spatelquestion when i run -e venv_rebuild=yes in that case does it going to rebuild all whl including nova and neutron also right?20:44
spatelI am seeing with os-nova-install.yml its just build nova whl not neutron and that is why i am in that race condition20:45
jrosserso20:45
spatelseems like bug20:46
jrosserlook at the condition here https://github.com/openstack/ansible-role-python_venv_build/blob/stable/victoria/tasks/python_venv_wheel_build.yml#L12820:46
jrosserwhen is that true?20:46
spatelone of this file - /var/www/repo/os-releases/22.0.0.0rc1/nova-22.0.0.0rc1-requirements.txt20:47
spatelright20:48
jrosseras i read the code now, running venv_rebuild=true will delete the wheels in ubuntu-18.04-x86_64 or whatever the subdirectory is20:49
spatelyes i can see its destroying whole folder20:49
jrosserbut it will leave behind everything else except the service you are deploying, because of https://github.com/openstack/ansible-role-python_venv_build/blob/stable/victoria/tasks/python_venv_wheel_build.yml#L49-L5220:50
jrosserso it seems that if you venv_rebuild=true for nova, it will leave the neutron constraints/requirements/global-constraints files present20:50
spatelthey are present there20:51
jrosserwhich means that next time you come to run neutron install, the wheels are gone but the condition that says to build them is false20:51
spatelwhen you saying file is changed does that means timestamp or content of files?20:52
jrosserfor example _requirement_file is changed20:52
jrosserthat is the registered result from this _requirement_file is changed20:53
jrosseroops20:53
jrosserhttps://github.com/openstack/ansible-role-python_venv_build/blob/stable/victoria/tasks/python_venv_wheel_build.yml#L79-L8620:53
jrosserif the ansible task is 'changed', i.e the content in the file is different20:53
jrossernow i am really surprised that we don't see this before, because i do venv_rebuild=true all the time in AIO for developing stuff20:54
jrosserbut that is generally for just one service20:54
spateli haven't seen this issue also but suddenly seeing with new deployment20:55
jrosseri would be really interested to know what noonedeadpunk thinks of this20:55
spatelsomething definitely changed20:55
*** dtantsur is now known as dtantsur|afk20:55
spateljrosser: i have noticed whenever i run -e venv_rebuild its changing timestamp of nova-22.0.0.0rc1-requirements.txt file21:00
spatelfeels like new file21:00
jrosserfor which playbook?21:00
spatelnova or neutron any playbook21:00
spatelcurrently i can see nova-22.0.0.0rc1-requirements.txt file has current timestamp21:01
jrosserthats what i'm trying to explain i think21:01
jrosserfor the playbook you run (lets say nova), it will delete those files21:01
jrosserhttps://github.com/openstack/ansible-role-python_venv_build/blob/stable/victoria/tasks/python_venv_wheel_build.yml#L49-L5221:01
jrosserbut not the ones for any other service21:01
jrosserthis is why it seems to work for you when you try nova with venv_rebuild=true21:02
jrosserbut then the next service will fail21:02
spatelyes21:02
jrosserthe ones that get deleted trigger the wheel build for that service (only) when they are re-created21:02
spateldo you think this is still related to lsync issue?21:03
jrossernothing to do with lsync21:03
spatelok21:03
jrosseras far as i can see, it looks really like a logic error in python_venv_build21:04
spatelgood if we know what is going on so i stop poking around :)21:04
jrosserso - delete all those .txt files in /var/www/repo/.... and you should be good21:04
spatellets see if noonedeadpunk has anything to offer us tomorrow21:04
jrosserthat will force the wheel build21:04
jrosserwithout deleting any existing wheels21:05
spatelyou want me to delete nova/neutron *.txt file from /var/www/repo ?21:05
spateland re-run build?21:05
jrosserare you still stuck?21:06
spateli am not stuck (workaround is -e venv_rebuild=yes)21:06
jrosserok then leave it21:06
spateli can add computes but trying to fix this issue so my other team member don't ask why its failing :)21:06
spatelI need to educate them for this issue.21:07
jrosserwell i think i would advise not using venv_rebuild=true, or telling people to use it to 'fix' things21:07
jrosseras that will create mythology :)21:08
spatelYes that is what i m not telling them21:08
spateli want to fix this issue not following bad practice21:08
jrosserif you want to clean your current deployment remove all those txt files and run without venv_rebuild=true21:08
jrosseri think it will then re-create everything you need21:08
spatelall *txt or just for nova/neutron?21:08
jrosserwell, if all the wheels have been deleted then all those files are unhelpful now21:09
spatellet me try with just nova/neutron first and if won't help then delete all21:10
spatelin worst case i will upgrade to 22.0.021:10
jrosseryeah, then your compute add should just work21:10
spatelLets talk to noonedeadpunk tomorrow with fresh mind :)21:10
spateldo you want me to open bug to keep trace of it or we can poke him here ?21:10
jrosserhere is good for now21:11
spatelcool!21:11
jrosserthanks for taking the time to look into it21:11
*** gshippey has quit IRC21:12
spatelno worry! i want to make sure we are good before anyone else see this issue :)21:12
spatelthank to you for sticking around with this issue21:12
jrosserno problem :)21:12
spateljrosser: here is the result of experiment, i deleted nova*.txt and neutron*.txt from repo and re-run without venv_rebuild21:20
spatelsuccessfully able to run both playbook and i can see both nova*.whl and neutron*.whl files21:20
spatelthis is good state21:21
jrosserfabulous21:21
jrosserthat kind of confirms what is happening21:21
spatelnow if i add next compute it won't error out because they can see both whl files21:21
spatellet me prepare next compute node21:21
spatelnow if i run venv_rebuild=yes then i will go back to same chicken egg issue21:22
jrosseri think so yes21:22
spateljust confirming21:22
spatelso tomorrow we need to find out how to fix that21:23
*** SecOpsNinja has left #openstack-ansible21:44
*** jbadiapa has quit IRC22:06
*** spatel has quit IRC22:50
*** tosky has quit IRC23:18
*** admin0 has quit IRC23:39
*** cshen has quit IRC23:48

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!