Tuesday, 2024-02-27

opendevreviewfrankming proposed openstack/ironic master: Modify ESP configuring script of redfish document  https://review.opendev.org/c/openstack/ironic/+/90995303:38
opendevreviewfrankming proposed openstack/ironic master: Modify ESP configuring script of redfish document  https://review.opendev.org/c/openstack/ironic/+/90995303:40
opendevreviewfrankming proposed openstack/ironic master: Modify ESP configuring script of redfish document  https://review.opendev.org/c/openstack/ironic/+/90995303:43
opendevreviewfrankming proposed openstack/ironic master: Fix iscsi url generate method for ipxe  https://review.opendev.org/c/openstack/ironic/+/91030007:00
dtantsurarne_wiebalck: hey, good morning. Do you folks rely on python-hardware and extra data inspection?08:37
opendevreviewfrankming proposed openstack/ironic master: Fix iscsi url generate method for ipxe  https://review.opendev.org/c/openstack/ironic/+/91030008:55
opendevreviewDmitry Tantsur proposed openstack/ironic master: [WIP] Add inspection PXE filter service  https://review.opendev.org/c/openstack/ironic/+/90799109:03
opendevreviewRiccardo Pittau proposed openstack/ironic-python-agent-builder master: Update link to ipmitool repository  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/91021609:32
opendevreviewMerged openstack/ironic master: neutron: do not error if no cleaning/provisioning on launch  https://review.opendev.org/c/openstack/ironic/+/90993712:31
TheJuliagood morning14:08
opendevreviewRiccardo Pittau proposed openstack/ironic master: [WIP] move back to plain pyasn1 in 2024.2  https://review.opendev.org/c/openstack/ironic/+/91034214:12
opendevreviewJulia Kreger proposed openstack/ironic stable/2023.2: neutron: do not error if no cleaning/provisioning on launch  https://review.opendev.org/c/openstack/ironic/+/91031014:13
opendevreviewJulia Kreger proposed openstack/ironic stable/2023.1: neutron: do not error if no cleaning/provisioning on launch  https://review.opendev.org/c/openstack/ironic/+/91031114:13
opendevreviewRiccardo Pittau proposed openstack/ironic-python-agent-builder master: Update ipmitool version to 1.8.19  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/91034414:31
TheJuliadtantsur: it occurs to me that some operators do for the extensive data, but may be the path is to make it "more optional" or a dib element plugin?15:24
dtantsurTheJulia: it is a plugin.  The problem is: the library itself is maintained by rpittau alone, and we don't have a use case for it.15:25
JayFlet me have a look15:29
JayFmy downstream might use it15:29
JayFThis is a major downside of putting our stuff in libraries15:29
JayFif this was a feature fully implemented in IPA, I suspect we don't have this conversation15:29
dtantsurYeah, although in this case it was the other way around: we just reused a library15:30
JayFah, and rpittau took it over 15:30
JayFI saw redhat-cip in the org and assumed we/you all built it15:30
dtantsurYep. Red Hat "inherited" it from the eNovance acquisition.15:30
dtantsurAnd the OSP HardProv team, later OCP Metal team ended up responsible for it.15:31
dtantsur(could be a good PTG topic btw)15:33
TheJulia++++15:43
*** nfedorov_ is now known as jingvar15:56
opendevreviewDmitry Tantsur proposed openstack/ironic master: [WIP] Add inspection PXE filter service  https://review.opendev.org/c/openstack/ironic/+/90799116:19
opendevreviewDmitry Tantsur proposed openstack/ironic master: [WIP] Add RPC for the PXE filter service  https://review.opendev.org/c/openstack/ironic/+/91036516:21
opendevreviewDmitry Tantsur proposed openstack/ironic master: [WIP] Add inspection PXE filter service  https://review.opendev.org/c/openstack/ironic/+/90799116:23
dtantsurI wonder if something is still wrong with eventlet. Using Event.wait locks the complete process so much it only responds to SIGKILL.16:29
JayFWhat eventlet version, what python version?16:31
JayFEasy way to tell if it's eventlet usually is going from 3.11+ to <3.1016:31
JayFthere's very little that changed around how that works in older pythons16:31
JayFbut if we can get a reproducing failure I can point someone at it16:31
dtantsurPython 3.9.16, eventlet==0.35.116:31
JayFoh, that's a bad version isn't it?16:31
dtantsurhuh, then why is it in u-c?16:31
JayFah, I was thinking 0.34.1 I think16:31
dtantsurI'm testing https://review.opendev.org/c/openstack/ironic/+/907991. Basically just starting the new service.16:32
JayFyeah 34.1 was the bad one, confirmed16:32
dtantsurIt starts then locks up until the event timeout passes. Then it handles Ctrl+C.16:32
* dtantsur is pondering a busy loop instead of an event..16:37
JayFdtantsur: you didn't monkey_patch16:40
JayFhttps://review.opendev.org/c/openstack/ironic/+/907991/8#message-6a4cad63ec16df488b72eba54288e81ced3a19a816:40
dtantsurJayF: it's in cmd.__init__ already16:43
JayF__init__ doesn't run when you have a main16:43
JayF__init__ runs on import16:43
JayFnot on calls to main16:43
JayFaiui16:43
dtantsurJayF: then nothing in ironic monkey patches :)16:43
dtantsurbut no, you cannot import a submodule without running __init__16:44
JayFhmmmm16:44
dtantsur(we rely on it for all our commands)16:44
JayFyep, you're right16:44
JayFso it prints no messages?16:45
JayFnothing angry from eventlet16:45
JayFso that means you must be patched in16:45
dtantsurI don't see anything.. I also use eventlet.queue and spawn explicitly.16:47
JayFCan you lay out the exact case you're seeing, and perhaps a simple reproduction that is doable by someone less-openstacky?16:49
JayFI'm basically trying to see if I can get this to a point where I can ask someone more eventlet-y to look at it16:49
JayFbut that's only going to happen if we can repro in a venv/outside devstack16:49
JayFdtantsur: have we used https://docs.python.org/3/library/fcntl.html with eventlet before? 16:50
JayFIt probably doesn't even get that far, does it?16:51
dtantsurI'll see if I can come up with an easier reproducer16:54
dtantsur(I don't quite understand the context of the question around fcntl)16:54
rpittaugood night! o/16:58
opendevreviewVerification of a change to openstack/ironic-python-agent-builder master failed: Update tinyipa to tinycore 15.x  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/91016917:10
opendevreviewVerification of a change to openstack/ironic-python-agent-builder master failed: Update link to ipmitool repository  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/91021617:10
opendevreviewVerification of a change to openstack/ironic-python-agent-builder master failed: Update ipmitool version to 1.8.19  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/91034417:10
JayFdtantsur: I'm mainly just suspect it could be responsible for a deadlock; that's all17:42
TheJuliaHey folks, https://review.opendev.org/c/openstack/ironic-tempest-plugin/+/902171 would love a review. It would enable us to run multiple distinct boot_interface scenario jobs on the same overall job. Specifically so we cover more boot interfaces without adding more scenario jobs.19:00
JayFTheJulia: +2a19:05
JayFhttps://review.opendev.org/c/openstack/releases/+/910398 I'll note I just +1'd this patch to move V, W, X to unmaintained/ namespace19:29
JayFPlease let me know if at some point you all stop caring about V so we can EOL it :)19:29
JayFAt this point, my personal investment stops around W19:29
JayF(but you know me, I'll backport things anyway)19:29
TheJuliaI only care about 2023.1 and W19:40
opendevreviewJulia Kreger proposed openstack/ironic-tempest-plugin master: Invoke tests with fake interfaces  https://review.opendev.org/c/openstack/ironic-tempest-plugin/+/90993919:40
TheJuliawow that took a long time to post into gerrit19:40
JayFwe can't keep "W" without a chain of upgrades to master19:42
JayFso that means you care about everything W+19:42
*** elodilles is now known as elodilles_pto19:46
TheJuliacare is on different levels, truthfully19:56
JayFif it was maintained19:56
JayFwe wouldn't prefix with "un-" :D 19:56
opendevreviewJulia Kreger proposed openstack/ironic master: ci: fix dnsmasq downgrade package location  https://review.opendev.org/c/openstack/ironic/+/91043619:57
TheJuliaflagged https://review.opendev.org/c/openstack/ironic/+/910436 as ironic-week-prio as it is causing numerous job failures19:58
TheJuliawell, it seeks to fix what causes many failures19:58
TheJuliahttps://zuul.opendev.org/t/openstack/status#ironic seems a bit like https://media1.tenor.com/m/6031wf2pdKwAAAAd/you-got-red-on-you-you-for-a-stain.gif19:59
TheJuliaWell, looks like we may just want to disable grenade to merge it and then backport the fix to branches which are also impacted.20:13
JayFif we need to do that, I'm onboard20:16
JayFwill likely not have free time to look at this until tomorrow or later20:16
TheJuliaI'm going to let the change on the master branch run for now20:17
TheJuliajust to make sure it is good/happy20:17
opendevreviewVerification of a change to openstack/ironic-tempest-plugin master failed: Test multiple boot interfaces as part of one CI job  https://review.opendev.org/c/openstack/ironic-tempest-plugin/+/90217120:21
JayFTheJulia: So looking at https://review.opendev.org/c/openstack/ironic/+/894460/18/zuul.d/ironic-jobs.yaml#769 -- the multitenant job is failing; kinda expected given how it's setup; but I wanted to ensure that changing the regexp is the right path to get that job to not run20:30
JayF(all the other jobs pass)20:30
JayFand yes, I'll DRY up the multinode stuff in a follow-up (I'm trying right now to get a job working to make nova happy)_20:31
JayFignore sean's comment, he enabled the job overnight in nova so we have job outputs20:33
JayFTheJulia: actually... I can't make sense as to why the multitenant job fails there20:35
JayFooooh, yeah, I know why20:35
JayFno, I don't20:35
JayFwtf20:35
JayFThe only difference is setting IRONIC_SHARDS and a shard name20:37
JayFthey are hooked up to the same flavor20:37
JayFI cannot explain why one works and one does not20:37
JayFand it's concerning to me as it could be a real bug(?)20:37
JayFhmm, I see a comment that might be a pointer to this being broken in our devstack plugin20:39
JayFhttps://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/log/controller/logs/screen-ir-api.txt#2378-2389 maybe related20:44
JayFfor now I'm skipping that job20:44
JayFand just doing BaremetalBasicOps20:44
JayFand going and and DRY'ing it20:44
opendevreviewJay Faulkner proposed openstack/ironic master: [CI] Support for running with shards  https://review.opendev.org/c/openstack/ironic/+/89446020:52
opendevreviewJay Faulkner proposed openstack/ironic master: [CI] Support for adding dummy shards  https://review.opendev.org/c/openstack/ironic/+/91044020:52
JayF^ that change swaps it to only do BaremetalBasicOps, and I went ahead and DRY'd it up20:53
JayFsean-k-mooney: ^ that should actually run itself this time, as a bonus lol20:53
sean-k-mooneyack20:54
sean-k-mooneyi may not get time to redo my patch tonight but if can do it quickly ill let it run and check it in the morning20:54
sean-k-mooneyso with the split i shoudl jsut depnd on https://review.opendev.org/c/openstack/ironic/+/894460 for now right20:54
JayFyes20:55
JayF910440 is separate, just for dummy shards20:55
JayFwhat does your patch need?20:55
JayFit should be trivial, yeah?20:55
sean-k-mooneyso ironic-tempest-ipa-wholedisk-bios-agent_ipmitool-tinyipa20:56
JayFno20:56
sean-k-mooneyis the sanity check job we run on every patch20:56
JayFoh, yeha20:56
JayFThat's the one you want me to hook up to dummy when it's' built?20:56
JayFI can swing that20:56
sean-k-mooneywell we could or we could change what job we run20:56
JayFlet me restack it real quick to be that way, shouldn't be hard20:56
JayFyou are running the most passe' job possible20:57
sean-k-mooneyyour updating ironic-tempest-uefi-redfish-vmedia20:57
JayFBIOS! IPMI!20:57
JayFbah! 20:57
JayFlol20:57
JayFyeah, just because redfish is newer and shinier but it doesn't matter20:57
sean-k-mooneybasically im asking20:57
JayFI'll flip it around20:57
sean-k-mooneyif you could have just one ironic job on every nova patch20:57
JayFI want, for purposes of this conversation, the minimal change possible in nova20:57
sean-k-mooneywhat woudl you like it to be20:57
JayFand do not want you to address my answer to that question :)20:57
JayFlol20:57
JayFprobably something uefi+redfish reflects the majority of how things are run20:58
sean-k-mooneyok so i think we can just swap to ironic-tempest-uefi-redfish-vmedia20:58
sean-k-mooneywhen that is working with the dumy nodes20:58
JayFeasy enough20:58
JayFjust know my choice was probably decided as much by where my cursor was in the file20:58
JayFand I just lucked out to hit a good choice for that use lol20:59
JayFactually, I wonder if nova should care about vmedia vs pxe job20:59
TheJuliaunlikely, tbh20:59
JayFany ideas on that multitenant weirdness?20:59
JayFit makes me nervous but I don't know anything about that job20:59
TheJuliaI haven't been able to look, I am in a meeting at the moment20:59
JayFand sean-k-mooney is fairly certain it can't be because of sharding20:59
JayFack20:59
JayFmainly looking for you to pat me on the head and say it's OK LOL21:00
sean-k-mooneynova largely should not care about the particalar of the ironic driver21:00
JayFonly things I can really think of is BFV might be nice21:01
JayFbecause that's in our nova driver too21:01
JayFbut it probably trades coverage in one area for coverage in another21:01
JayFand I know volume stuff is a pain point for nova jobs so...21:01
TheJuliaJayF: all the changes on the job config you posted made me raise my eyebrow immediately, but I need to look at the actual results when I have time to do so21:01
sean-k-mooneyya so we would be happy to test bfv with ironic21:01
sean-k-mooneywell actully that because of a libvirt regression21:02
JayFTheJulia: the new version is DRY'd, the only changes are to tempest_test_regexp and adding IRONIC_SHARD* vars21:02
sean-k-mooneywell qemu 21:02
JayFso how about this21:02
JayFwe focus right now on "good enough for sharding to merge"21:02
sean-k-mooneybasically we have some josb fail because qemu does nto give up the volume when we ask it too21:02
JayFbut I promise you more time after to "improve nova<>ironic ci generally"21:02
sean-k-mooneyand sometime we hit slow node issues with lvm runing on a loopback device21:02
JayFwhich will include making this dummy shard stuff work, hooking up a post-run thing to ensure n-cpu didn't cross shards, and picking a better job for nova to run21:02
sean-k-mooneyright BFV is not needed for sharding21:03
sean-k-mooneybut we can iterage on this after FF21:03
JayFwell ideally, we get dummy shards + the post-run stuff workign well enough21:03
JayFthat we can drop the shard multinode21:03
JayFbecause we haven't even evaluated the fact that I gotta get that merged here21:03
TheJuliaJayF:  https://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/log/controller/logs/screen-ir-api.txt#2378-2389 is unrelated21:03
JayFthat's an NGS vif detach bug, right?21:03
TheJuliaJayF: where is the job saying that your seeing fail due to https://review.opendev.org/c/openstack/ironic/+/894460/18/zuul.d/ironic-jobs.yaml#741 with the multitenant job?21:04
TheJuliasince the test is not run21:04
TheJuliano, it is a trust/who is authoritative issue21:04
JayFsean made it run in nova, lemme find it21:04
JayFhttps://review.opendev.org/c/openstack/nova/+/91033321:05
TheJuliaand reality of "we need to do the thing to move it forward"21:05
TheJuliaokay21:07
TheJuliaso the issue is, not all scenario jobs are appropriate for all situations21:07
TheJuliahttps://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_2a3/910333/4/check/ironic-tempest-ipa-wholedisk-direct-tinyipa-multinode-shard/2a3f86e/testr_results.html21:07
TheJuliabasic ops and single tenant pass21:08
JayFTheJulia: my concern is simple: I'll note that https://review.opendev.org/c/openstack/ironic/+/894460/19/zuul.d/ironic-jobs.yaml represents the diff between this job and the one that passes21:08
TheJuliamulti-tenant is a specialized case test which requires additional configuration AFAIK21:08
JayFTheJulia: I can't answer *why it doesn't pass on shard job*21:08
TheJuliaso you *cannot* just say run everything, *espesically* in scenario21:08
JayFbecause it's identical config, just with sharding on top21:08
JayF(ignore the tempest_test_regex change in that patchset; it was modified to avoid this issue and get a clean pass, but I still wanna know *why*)21:09
JayFbecause it seems to me if it were possible for sharding to cause that job to fail, and *only sharding* that may be a canary of a potential issue21:09
JayFbut I can't connect dots as to where/how it could be, so I am a little stuck21:09
TheJuliaThe way I see it right now, they are unrelated, but maybe I'm not seeing the same concern21:10
sean-k-mooneyi was plannign to start with this  tempest_test_regex: (^tempest\..*compute\..*(reboot|rebuild).*)21:10
JayFTheJulia: If all I do is add sharding, and setup the two n-cpus on separate shards, why should that make multitenant tests fail?21:11
JayFThat's the exact question I'm trying to answer21:11
JayFand "we expect it to fail because X" is a good answer21:11
JayFbut AIUI there should be *zero* about having different shards setup that causes this to fail21:11
TheJuliamy gut feeling is you've resurrected job config we just don't expect to be run anymore21:14
TheJuliabut I need to be able to pivot to focus to know for sure21:14
JayFYeah, that assumption is invalid. That job is passing alongside the failing job21:15
TheJuliawhere?21:15
JayF> ironic-tempest-ipa-wholedisk-direct-tinyipa-multinode https://zuul.opendev.org/t/openstack/build/8f54392ac8ec4739ad1d782b245e7cfa : SUCCESS in 1h 30m 41s (non-voting)21:15
JayFhttps://review.opendev.org/c/openstack/ironic/+/894460/19#message-114d559904d38dc5d0c0a9e0081294cbaed5f4cf21:15
JayFthat link is good to use, to results21:15
TheJulia\o/21:15
JayFthat is the only reason I WTF; that I see this job passing21:15
JayFthe -shard variant is not passing the same test21:16
JayFwith shards being the only difference21:16
JayFwhich is hrm-inducing21:16
TheJuliadid you look at the actual failure?21:16
TheJuliatesttools.matchers._impl.MismatchError: 'PING 10.0.100.223 (10.0.100.223) 56(84) bytes of data.\n64 bytes from 10.0.100.223: icmp_seq=1 ttl=64 time=6.22 ms\n64 bytes from 10.0.100.223: icmp_seq=2 ttl=64 time=0.458 ms\n64 bytes from 10.0.100.223: icmp_seq=3 ttl=64 time=0.434 ms\n64 bytes from 10.0.100.223: icmp_seq=4 ttl=64 time=0.475 ms\n\n--- 10.0.100.223 ping statistics ---\n4 packets transmitted, 4 received, 0% packet 21:16
TheJulialoss, time 3008ms\nrtt min/avg/max/mdev = 0.434/1.895/6.216/2.494 ms\n' matches Contains(' bytes from 10.0.100.223') <-- deja vu21:16
JayFYeah, so in the sharding case the end nodes can communicate21:17
JayFwhereas in the non-sharding case they can't21:17
JayFor is there something else I'm missing?21:17
TheJuliathe match is failing 21:19
TheJuliathe text it is trying to compare suggests success21:19
TheJuliabut hey, don't know why it doesn't want to match it right now21:19
JayFit's the opposite, it's a negative match21:19
JayFand it found it21:19
JayFthat ping /is not supposed to succeed/ AIUI21:19
TheJuliaoh shoot21:20
sean-k-mooney    pre-run: playbooks/ci-workarounds/pre.yaml21:20
sean-k-mooneyis there a reason you added that21:20
JayFhttps://github.com/openstack/ironic/blob/master/playbooks/ci-workarounds/pre.yaml21:20
JayFlooks like it does multi-node-bridge?21:21
sean-k-mooneythat should be done in the parent job right21:21
JayFoooh good point21:21
JayFI see21:21
sean-k-mooneyas in the exiting multi node josb pre playbooks will run before this jobs21:21
sean-k-mooneyin an onion style 21:21
sean-k-mooneyso that proably not needed21:22
opendevreviewJay Faulkner proposed openstack/ironic master: [CI] Support for running with shards  https://review.opendev.org/c/openstack/ironic/+/89446021:22
opendevreviewJay Faulkner proposed openstack/ironic master: [CI] Support for adding dummy shards  https://review.opendev.org/c/openstack/ironic/+/91044021:22
JayFthat's the updated one21:22
sean-k-mooneyack21:22
sean-k-mooneyso BaremetalBasicOps is the new tempest regex21:23
JayFyep21:24
TheJuliaJayF: I think I would need to dig at the job log itself and compare to the test for sure, because out of the box, yeah, that should be failing. I *will* note 21:34
TheJuliaerr21:34
TheJuliasorry, still on a short call that is dragging on21:34
TheJuliaso, the *test* was originally intended if I'm remembering correctly to be a vm in one tenant and a baremetal in another21:35
JayFoooh21:35
JayFI'm going to step away and walk the dog, I'd like to figure out why that's failing but we've hopped over that issue for now21:36
JayFand sean-k-mooney is fairly confident that even if the fail is real (doubtful, still) that it's not likely caused on nova side21:37
JayFso I think the job gives us confidence in any event w/r/t what Nova devs are lookin'21:37
sean-k-mooneyi also dont think this is nova related21:37
sean-k-mooneyhttps://review.opendev.org/c/openstack/nova/+/91033321:37
sean-k-mooneyi have slit out the job change21:37
JayFsean-k-mooney: I'll note we're going from 4->2 nodes booted with the regexp change :)21:38
JayFsean-k-mooney: I think we really need the post-run bit to validate to have confidence the CI job will, unattended, report if things are broken (even if logs tell us it's working)21:39
sean-k-mooneyso what ill like do is create a nova-ironic-shard job that inherits form either the multi node or dumy node job21:39
sean-k-mooneyand just tweaks the regex to enable more nova related tests21:39
JayFbooting more instances to validate in Ironic is ... not my favorite approach, as each machine we boot makes the job take longer and exposes us to more false alarms21:39
JayFso just have some awareness of how chonky our fake machines are21:39
JayFnested virt is terrible, only thing more terrible is NOT nested virt ;) 21:40
sean-k-mooneywell our nova libvirt jobs are also booting vms in the vms21:40
sean-k-mooneyand thatn mainly what they are for21:40
JayFI guess you have to own the libvirt, don't you21:41
sean-k-mooneywhat would be nice form an nova perspective (you dont have to do this) is for use in our gate to add more usage of the nova api21:41
JayFso no magical "rehome my top level hypervisors"21:41
sean-k-mooneywel we have libvirt installed in teh zuul vms just like ye do21:41
JayFsean-k-mooney: lets have this conversation during some of the downtime between cycles, I'm on board to improve coverage ... on the flip side, we're getting close to "feature complete" in ironic driver21:42
sean-k-mooneyits just nova talks to it directly instead of via ironic and vbmc21:42
JayFwe have nothing slated for it, moving forward, outside of the metadata stuff we're also working on together21:42
* TheJulia collapses after meeting21:42
sean-k-mooneycool but we dont have a lot of coverage or the integration in general21:42
JayFI am going to actually go walk the dog now :) my brain keeps looking at the clock expecting it to be late bceause I'm exhausted21:42
JayFsean-k-mooney: yeah, I realized that goes both ways21:43
JayFsean-k-mooney: if it's more stable, even more reason to ensure it doesn't break21:43
sean-k-mooneyso testing BFV and other things would likely be good if we dont already have that in the existing job21:43
JayFmy feelings on our nova driver are way, way, up since we've got the client/sdk split through21:43
sean-k-mooneyi am not really sure how much ironic-tempest-ipa-wholedisk-bios-agent_ipmitool-tinyipa tests21:43
JayFthe funniest thing about that job name21:43
JayFwe haven't had "agent_ipmitool" style driver names in AGES21:44
JayFI suspect we kept it to not break nova21:44
JayFso getting onto a better job is not a bad idea21:44
TheJuliaI tried to rename it ages+1 or ages+2 cycles ago21:44
sean-k-mooney...21:44
JayFand on the ironic side maybe even indicate it's used for nova, too21:44
sean-k-mooneyhttps://d393ab92b65d6ff2eea5-fc707f543607a38fac44776c15f601da.ssl.cf5.rackcdn.com/906992/5/check/ironic-tempest-ipa-wholedisk-bios-agent_ipmitool-tinyipa/f001dcc/testr_results.html21:44
TheJuliabut devstack's gate was so faily that I never succeeded21:44
sean-k-mooneyit runs one test21:44
JayFwow, rescue, and bfv21:44
JayFthat's honestly not bad21:44
JayFin terms of horizontal api coverage21:45
sean-k-mooneyam no21:45
sean-k-mooneywell its only runnign est_baremetal_server_ops_wholedisk_image21:45
sean-k-mooneyits skiped BFV21:45
JayFoh, yeah, I see that now heh21:45
JayFyeah, we'll improve this too21:45
JayFthis stuff /does/ get tested in Ironic gate though21:46
sean-k-mooneyso the new job your adding has more coverage then the the existing job...21:46
JayFso putting in nova gate just is an earlier warning system21:46
JayFwhich I think is why our focus has been on "make that job basic and small and super reliable" over coverage21:46
JayFthat is a decision we can revisit, but I'll put it this way: I never want the "Gate Status" portion of the TC meeting to contain "Ironic jobs are keeping Nova stuff from landing"21:46
JayFlol21:47
sean-k-mooneysure but its a very expeinsive way to boot one vm21:47
sean-k-mooneyanyway you should go walk your dog21:47
JayFyes, I should21:47
JayF:D21:47
JayFhttps://usercontent.irccloud-cdn.com/file/2ip0E65I/FB_IMG_1633191782370.jpg (dog tax paid)21:48
JayFbrb :)21:48
TheJuliaoh noes21:48
TheJuliathe dnsmasq version 2.80 is quietly exiting again21:49
opendevreviewyatin proposed openstack/ironic master: Source install dnsmasq-2.87  https://review.opendev.org/c/openstack/ironic/+/88812121:51
TheJuliaI bet that will be happy21:52
TheJuliasame plan, we could just disable grenade and backport to fix it21:52
* TheJulia begins wishing we never merged the mutlitenancy test22:40
JayFTheJulia: I am back, I have other things I *can* do, but I'm willing to hate that test with you if you think it's helpful :)22:47
opendevreviewJulia Kreger proposed openstack/ironic master: ci: Source install dnsmasq-2.87  https://review.opendev.org/c/openstack/ironic/+/88812122:50
JayFsean-k-mooney: noting that ironic ci is busted now, all those jobs are going to fail fast22:50
TheJuliaThe weird thing is they are on the same logical subnet, so of course ping will work22:51
JayFyeah I'm not sure I grok how the actual test is passing22:51
JayFthat's mainly where my wtf is coming from22:51
JayFTheJulia: +2 888121, mainly because I trust the gate to complain if you did it wrong :)22:52
JayFTheJulia: that way also you can land it later if it passes 22:53
TheJuliaJayF: yakin did it and it passed the gate ages ago, so it should still work22:53
JayFack; wfm22:53
JayFthere's probably value is getting a ppa up22:54
TheJuliastevebaker[m]: fyi https://review.opendev.org/c/openstack/ironic/+/888121 which we'll need to backport to stable/2023.2 (and likely further, thanks ubuntu!)22:54
TheJuliaI guess it is just a little frustrating since it is a known issue which kind of got shrugged at if memory serves22:54
TheJuliabut the older package route was only going to work for so long22:55
JayFI read the LP bug more as "hell if I know" and they punted :( 22:55
JayFTheJulia: you have that LP bug # at hand, perhaps?22:56
TheJulianot handy, I just remember we noted the link where it was known in dnsmasq22:56
JayFhttps://bugs.launchpad.net/ubuntu/+source/dnsmasq/+bug/202675722:56
stevebaker[m]the source install looks cleaner anyway, heh22:58
JayFthat is ... ignoring the real issue in a sense22:58
JayFI have a downstream that may be deploying Bobcat Ironic on Ubuntu.22:58
JayFWhat do I tell them?22:58
JayFIt doesn't work? Build your own dnsmasq? 22:59
JayFSo I think I'll try to find time to make a better answer; even if we have to doc "use our PPA because ubuntu won't fix 2026757"22:59
TheJuliaubuntu won't pull in the updated version of dnsmasq because that is a change for a known issue on config updates in the version they are shipping23:00
JayF*blink*23:01
JayFdo you mean because they document it as a known issue, they won't fix it?23:01
JayFOr did I misread that?23:01
TheJuliait is known to dnsmasq folk23:01
JayFoooh so the idea is, we don't have a moving-forward fix23:01
JayFonly a moving-backwards unbreak23:01
JayFugh23:01
TheJuliaand 2.87 is known good as well, but that would mean incrementing their package version23:02
JayFthey aren't willing to backport a fix?23:02
TheJulia... ubuntu?23:02
JayFyes23:02
TheJuliaI mean, they are wanting to pin it down exactly, I think thekelly's post mentioning it pins it down to like 2 different changes I think23:03
TheJuliaits been a while unfortunately23:04
TheJuliaso pivoting subjects to the multitenancy job, even the exemplar job logs have them on the same network( 99253a6c-b699-4d2f-8c79-8d1ec7d9a8f8 10.0.100.16 1640f2be-f306-4a99-97a7-d9bd3f38f64e 10.0.100.249) https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_8f5/894460/18/check/ironic-tempest-ipa-wholedisk-direct-tinyipa-multinode/8f54392/controller/logs/tempest_log.txt23:06
JayFso we've reoriented from "why does the sharding job not pass" to "why does the actual job pass"23:10
JayFthat's comforting, weirdly23:10
TheJuliabasically yeah23:13
TheJuliaso, the way the job is designed, is it uses the *same* IP range for two different networks23:13
TheJuliaand being in tenants, they should have entirely separate config23:14
TheJuliaand then we end up wiring the nodes to different vlans23:14
TheJuliaand it confirms if things are happy23:14
TheJulia... or not23:14
JayFso really is just testing NGS23:16
JayFand/or Ironic/Neutron/NGS interations23:17
TheJuliabasically yeah23:18
TheJuliaand if the underlying contract is solid there, or not23:18
TheJuliaand in this case, I can *see* the tags getting set23:18
TheJuliaI just don't know why they don't take effect or hold23:18
JayFso this is just a mechanism of the test thing23:18
JayFnot an actual problem, most likely23:18
JayFif tags are getting passed through, NGS is doing the thing23:18
TheJuliayeah, what is happening is almost like, the tags are getting lost23:19
TheJulia... in fact, we should have some of it preserved23:19
JayFwhere are the tags set?23:19
JayFyou wanna VC while looking at it?23:19
JayFnot sure how much brain I have left, but I have time left in the day :)23:19
TheJuliahttps://1f56561c0b806c2ef991-4e4c7f9e4e66be3b004244bfc6fd74b6.ssl.cf5.rackcdn.com/910436/1/check/ironic-standalone-redfish/676e82b/post-job-network-ovs.txt23:19
JayFTheJulia: I should be seeing blank, yes?23:20
TheJuliaI don't think so23:20
TheJuliaI'd have to go look at the post job steps23:20
JayFI'm saying more like23:20
JayFit appears blank23:20
TheJuliayes, it is23:20
JayFis that a local web browser issue23:20
JayFor an empty file23:20
JayFlol okay23:20
JayFTheJulia: ... all those logs load in as empty to me23:21
JayFokay, there we go23:22
JayFWTF23:22
TheJuliahttps://github.com/openstack/ironic/blob/master/playbooks/ci-workarounds/get_extra_logging.yaml#L2623:22
JayFseems like this is all "ovs is in a strange/bad state"23:23
JayFbut again: how is that possible with the shard name set differently?!23:23
JayFwait23:23
JayFis there a possibility we're digging a random/weird failure23:23
JayFwe've not seen this happen 2x yet23:23
TheJuliaI'm fairly sure it is a weird/random failure which is unrelated23:23
JayFack23:23
JayFsgtm23:23
TheJuliashard is all about the input into nova-compute23:23
TheJuliaand a filtered view23:23
TheJuliathat wouldn't change the behavior or outcome like 7 steps removed23:24
TheJuliaI think that sort of sums it up, there is some sort of ovs problem on the job/hosts/config where somehow everything ends up wired together23:26
TheJuliaand it shouldn't, and that test *really* is built around finding a bad/misbehaving ml2 plugin23:27
opendevreviewJulia Kreger proposed openstack/ironic stable/2023.2: ci: Source install dnsmasq-2.87  https://review.opendev.org/c/openstack/ironic/+/91044423:33
TheJuliait looks like we don't need to take that further23:36
opendevreviewJay Faulkner proposed openstack/ironic master: Remove downgrade_dnsmasq; 1.90 is upstream now  https://review.opendev.org/c/openstack/ironic/+/91044523:47
JayFfor those not seeing the chat in #openstack-tc: fungi found and pointed out that 1.90 was pushed to Jammy repos a couple weeks back23:48
JayFso while I don't wanna stop Julia's from hitting the gate, I stacked a full removal of the downgrade on top23:48
JayFif that passes we can backport that down and be fixed, too23:48
JayFbut let Julia's fix land since our CI is busted already and it's close23:48
* JayF &23:49
fungito be clear, i just stumbled across it when looking to see if it would make sense to ask for the version from noble to be added to jammy-backports23:49
TheJulia2.90 :)23:51
opendevreviewJulia Kreger proposed openstack/ironic stable/2023.2: Special case lenovo UEFI boot setup  https://review.opendev.org/c/openstack/ironic/+/91044623:55

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!