Thursday, 2024-04-04

jandersTheJulia yes02:39
janders(sorry for slow response)02:40
TheJuliajanders: just wondering if can sync up/get an update on virtual media. Looks like hpe may have put in a url length limit03:59
TheJuliaat least, one that is pretty aggressively short03:59
jandersoh dear that doesn't sound like good news04:00
jandersI'm totally bogged down with an urgent lifecycle controller related issue, hopefully will get on top of that in a day or two04:01
jandersunfortunately I don't have much news on push-based vMedia - I haven't been able to find time to push it further, transition back from parental leave hasn't been exactly easy04:02
janderswill try to plan better for this one this quarter04:03
TheJuliajanders: ack, okay. No worries. I suspect we need to likely point out to vendors, at least with the url that there are both standards where there is a reasonable maximum url, and reason for longer urls in anything dealing with automation.04:06
TheJuliaAnyway, I’m going to likely try and get some sleep.04:06
jandersgood night, sleep well04:06
janderswe'll chat more soon04:06
dtantsurTheJulia: mmm, I see. but we still need to make it random, so that it cannot be guessed? (although, it can be opt-in for now to avoid backport issues)07:20
rpittaugood morning ironic! o/07:26
rpittauTheJulia: for the secure boot topic I've added it to Thursday since we have still some time 07:31
drannouHello Ironic ! I'm playing with soft raid in Ironic, and I'm suprrised about something: during a recycling process, IPA is doing a shred, which cost a lot. In my usecase, I'm recycling a host that have 2 NVME and 1 SSD. They all managed instant erase (so should take a couple of seconds), and in my case it took 1 hour for shred to be executed. Would it be better to just09:14
drannou"delete the raid", "clean all devices" (with instant erase if device support), and "re-create the raid"09:14
iurygregorygood morning Ironic10:48
*** sfinucan is now known as stephenfin12:25
dtantsurdrannou: yeah, it's a known omission..12:34
dtantsurdrannou: you can enabling delete_configuration in automated cleaning, then build configuration at deploy time (via deploy templates / steps). I haven't tried it.12:35
dtantsurCERN folks have experience with sw RAID, but they're hard to catch nowadays.12:35
drannoudtantsur: Yeah, as an operator, it's pretty weired to have to first make a manual clean to activate MDADM, and then call provide for an automatic clean to have a host in available state12:36
drannouAs an operator, calling a provide to a host should put the host in the required configurationb12:37
dtantsurClean steps were invented long before deploy steps, so we still tend to lean towards the former where we should recommend the latter12:37
drannouRight now I have to : activate the raid (node set --target-raid-config), manual clean to activate the raid, AND ask for a global clean (via "provide"), that's very weird. And more over if I don't make the manual clean, the provide will go on without error12:40
drannouit also means that if my final customer completely destroy the RAID, when he delete the instance, will it recreate the RAID ?12:41
TheJuliadtantsur: the value in the publisher be randomized and matched to the kernel command line, totally13:06
dtantsurdrannou: if you go down the deploy steps path, you can make sure the RAID is rebuilt every time. Otherwise, only if you do the manual cleaning dance each time.13:09
dtantsurTheJulia: nice ++13:09
TheJuliaI really wish it was just the uuid field, but that seems to be hard mapepd out to the creation time13:10
TheJuliamapped13:10
* TheJulia sips coffee13:10
drannoudtantsur: ok, can you confirm me that the mindset is : clean is there to REMOVE things, and deploy will configure (so create things that are not there) ? That's an importnat point for my SED ongoing development13:11
dtantsurdrannou: "cleaning" is a bad name in retrospective (better than the initially proposed "zapping" but still). It's more of what we used to call "ready state" at some point at RH.13:12
dtantsurDeploy steps were meant to allow instance-specific tailoring. But the actual difference is vague.13:12
drannoudtantsur: yeah, even more  with raid, why raid1, and not raid 10 ? is it up to the admin to define that ? 13:15
drannoubut if it's the instance customer, how should he configure that with nova ?13:15
drannouno easy answer 13:15
dtantsurdrannou: linking traits with deploy templates13:15
dtantsurmay or may not be easy depending on how you look13:16
drannoufor stock management, that's pretty hard13:16
TheJuliadrannou: could you elaborate on what you mean by "that" ? I ask because we're discussing a whole next evolution on deploy templates13:17
dtantsurI don't disagree with that, but I believe we've exhausted all ohter options with Nova13:17
* dtantsur lets the actually knowledgeable person aka TheJulia drive the conversation further :)13:17
TheJuliaoh, yeah, *that* is.... difficult since they want everything directly value discoverable13:17
* TheJulia will fondly remember the "we will never agree on any naming" forum session13:19
drannoudtantsur: that's hard : if, as a public cloud provider, you want to propose "some" raid management, if you have to "hard fix" the type of each host, how do you know that your customers will need 10 raid1, 2 raid 0, 5 raid-10 ?13:20
TheJuliaoh, it doesn't have to be fixed hard per host13:20
TheJuliaBut you have to establish some meaning behind flavors13:21
drannouif you let your customer be able to say that during the spawn, that's easier :)13:21
TheJuliayou publishize/detail those out, and then you map those flavors to have a trait which matches a deployment template pre-configured in ironic.13:21
drannoubut well we are too far, for the moment I need to be sure that if I tag a host as raid1, the host is in raid1 during the deploy :)13:22
TheJuliayes, except scheduling needs to be acutely aware of the fine details behind the ask, and what that translates out to to see if it is possible.13:22
TheJuliaAnd you *can* kind of do that, with traits today but you have to articulate the meaning if that makes sense13:23
TheJuliaThere is a whole aspect of quantitive versus qualitative traits to the backstory though. Hardware folks are generally seeking quantitative matching because we're generally in worlds of absolutes, which doesn't entirely align with a virtual machine in a hypervisor because while you are functionally restricted by the limits of the hardware, that is your upper bound and your not given the whole machine outright.13:47
opendevreviewMohammed Boukhalfa proposed openstack/sushy-tools master: Add fake_ipa inspection, lookup and heartbeater to fake system  https://review.opendev.org/c/openstack/sushy-tools/+/87536614:02
Continuitydrannou: we decided we would force our customers to have a RAID1 on the first 2 disks of a certain size, for that very reason. It was hard to allow the customer to choose, or to provide loads of flavours with it pre loaded.14:27
TheJuliaContinuity: I've heard similar from folks14:27
ContinuityTo be honest with NVMe and the longevity of the drives, im starting to wonder if SW Raid is even *needed*. Not having that resilience makes the operational part of my mind wince....14:28
TheJuliaheh14:29
ContinuityBut its a toss up between ease of management and protection of the customers service.14:29
TheJuliaYeah14:29
TheJuliaDisks failing, ages ago, was largely the spinny parts14:29
ContinuityAs long as we don't shred the drive every time it cleans :D14:29
TheJulia"whiirrrrrrrrrrl.... knock knock knock..... whiirrrrrrrrrrlllll"14:29
ContinuityLove that noise...14:30
TheJuliaYeah14:30
TheJuliaContinuity: I ran one of the largest image farms at it's time of *just* scanned documents in litigation back in the 2000s, and sooo many platters that came out of our data center stripped of the magnetic coating from head crashes and got hung on the walls14:32
ContinuityComputing has lost something since the removal of spinning parts....14:33
ContinuityIts not as visceral 14:33
ContinuityNow its just noisy fans14:33
TheJulialargely, yes14:33
iurygregoryis just me or what we recommend in https://opendev.org/openstack/bifrost/src/branch/master/doc/source/user/troubleshooting.rst#obtaining-ipa-logs-via-the-console is not supported?  I've found https://opendev.org/openstack/ironic/src/branch/master/releasenotes/notes/remove-DEPRECATED-options-from-%5Bagent%5D-7b6cce21b5f52022.yaml 14:33
TheJuliaoh, the docs need to be fixed, kernel_append_params ?14:34
iurygregoryI think it would be the case 14:47
iurygregorythe person who found will submit a patch =)14:47
dtantsurTheJulia: you may want to jump on https://review.opendev.org/c/openstack/ironic-specs/+/912050 real quick if you want to prevent `[driver]verify_ca` in favour of `[conductor]bmc_verify_ca`14:50
dtantsurI'm fine either way tbh14:50
opendevreviewMohammed Boukhalfa proposed openstack/sushy-tools master: Add fake_ipa inspection, lookup and heartbeater to fake system  https://review.opendev.org/c/openstack/sushy-tools/+/87536615:17
dtantsurmeanwhile, looking for a 2nd review https://review.opendev.org/c/openstack/ironic/+/91497215:20
opendevreviewRiccardo Pittau proposed openstack/ironic master: Fix redifsh detach generic vmedia device method  https://review.opendev.org/c/openstack/ironic/+/91497815:25
opendevreviewRiccardo Pittau proposed openstack/ironic master: Fix redifsh detach generic vmedia device method  https://review.opendev.org/c/openstack/ironic/+/91497815:26
TheJuliadtantsur: the proposal seems to work for me15:36
dtantsurack, thanks for checking15:38
rpittaugood night! o/16:06
opendevreviewJulia Kreger proposed openstack/ironic-tempest-plugin master: Unprovision iso ramdisk boot from test  https://review.opendev.org/c/openstack/ironic-tempest-plugin/+/91498018:16
opendevreviewSteve Baker proposed openstack/ironic-python-agent master: Step to clean UEFI NVRAM entries  https://review.opendev.org/c/openstack/ironic-python-agent/+/91456321:24
opendevreviewJulia Kreger proposed openstack/ironic master: Inject a randomized publisher id  https://review.opendev.org/c/openstack/ironic/+/91502221:31
TheJuliadtantsur: Take a look at ^ and let me know. It excludes the irmc driver's own direct calls (why?!) to the cd image generation which I'm sort of split on21:34
opendevreviewJacob Anders proposed openstack/sushy-oem-idrac master: [DNM] Wait for BIOS configuration job to complete  https://review.opendev.org/c/openstack/sushy-oem-idrac/+/91509222:31
janders( /me hacking on sushy-oem trying to get around "one LC job at the time" issue breaking adjusting BIOS settings)22:32
opendevreviewJacob Anders proposed openstack/sushy-oem-idrac master: [DNM] Wait for BIOS configuration job to complete  https://review.opendev.org/c/openstack/sushy-oem-idrac/+/91509222:33

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!