Wednesday, 2019-08-14

*** rcernin has joined #openstack-ironic00:08
*** whoami-rajat has joined #openstack-ironic00:09
*** igordc has joined #openstack-ironic00:09
*** trident has quit IRC00:38
*** gyee has quit IRC00:48
*** rcernin has quit IRC00:58
*** rcernin has joined #openstack-ironic01:14
*** hamzy has joined #openstack-ironic01:17
*** rh-jelabarre has joined #openstack-ironic01:58
*** rloo has quit IRC01:58
*** ash2307 has joined #openstack-ironic02:13
*** ash2307 has left #openstack-ironic02:16
*** whoami-rajat has quit IRC02:28
openstackgerritMerged openstack/networking-generic-switch master: Add Mellanox MLNX-OS Switch support  https://review.opendev.org/64256502:32
*** strigazi has quit IRC03:09
*** strigazi has joined #openstack-ironic03:10
*** whoami-rajat has joined #openstack-ironic03:20
*** ash2307 has joined #openstack-ironic03:34
openstackgerritKaifeng Wang proposed openstack/ironic-inspector master: [TEST] Update non-standalone job to use uwsgi  https://review.opendev.org/67572403:39
*** gkadam has joined #openstack-ironic03:41
*** gkadam has quit IRC03:41
*** gkadam has joined #openstack-ironic03:43
*** gkadam has quit IRC04:00
*** absubram has quit IRC04:11
*** mkrai has joined #openstack-ironic04:17
*** stendulker has joined #openstack-ironic04:18
*** dsneddon has quit IRC04:24
*** absubram has joined #openstack-ironic04:30
*** rh-jelabarre has quit IRC04:42
*** igordc has quit IRC05:02
*** dsneddon has joined #openstack-ironic05:11
openstackgerritDigambar proposed openstack/ironic stable/rocky: DRAC: Fix OOB introspection to use pxe_enabled flag in idrac driver  https://review.opendev.org/64836005:15
*** dsneddon has quit IRC05:17
*** dsneddon has joined #openstack-ironic05:21
*** dsneddon has quit IRC05:26
*** ash2307 has left #openstack-ironic05:31
*** mkrai has quit IRC05:33
*** mkrai has joined #openstack-ironic05:36
*** adrianc has quit IRC05:37
openstackgerritKaifeng Wang proposed openstack/ironic-inspector master: [TEST] Update non-standalone job to use uwsgi  https://review.opendev.org/67572405:46
*** absubram has quit IRC05:48
*** dsneddon has joined #openstack-ironic05:58
*** dsneddon has quit IRC06:03
*** rcernin has quit IRC06:16
*** jtomasek has joined #openstack-ironic06:20
*** kaifeng has joined #openstack-ironic06:21
*** rcernin has joined #openstack-ironic06:31
openstackgerritMerged openstack/sushy-tools master: Follow-up fixes  https://review.opendev.org/67462706:41
arne_wiebalckGood morning, ironic!06:44
kaifengmorning arne o/06:44
*** dsneddon has joined #openstack-ironic06:46
arne_wiebalckHey kaifeng o/06:47
arne_wiebalckIn case people are interested, we have summarized some of our findings when scaling ironic with nova here: https://techblog.web.cern.ch/techblog/post/nova-ironic-at-scale/06:49
*** dsneddon has quit IRC06:51
*** dsneddon has joined #openstack-ironic07:01
*** trident has joined #openstack-ironic07:01
*** tesseract has joined #openstack-ironic07:10
*** e0ne has joined #openstack-ironic07:22
kaifengarne_wiebalck: thanks for the article, it says "if an Ironic instance is already provisioned it can???t be moved to a different compute-node and in case of a nova-compute failure the user can't perform any API operation for his instance.", I am wondering why07:24
kaifengif a compute node is offline, the instance won't be shifted to another compute node?07:25
*** mkrai has quit IRC07:29
*** mkrai has joined #openstack-ironic07:31
*** mkrai has quit IRC07:38
*** rcernin has quit IRC07:56
arne_wiebalckkaifeng: no, the code does not do this (there is now a proposal to change this, though)08:00
*** dougsz has joined #openstack-ironic08:00
*** lucasagomes has joined #openstack-ironic08:02
kaifengarne_wiebalck: thanks, do you have the link to the proposal?08:11
*** rpittau|afk is now known as rpittau08:18
rpittaugood morning ironic! o/08:18
*** tssurya has joined #openstack-ironic08:22
*** rcernin has joined #openstack-ironic08:23
*** dougsz has quit IRC08:27
*** derekh has joined #openstack-ironic08:29
*** rcernin has quit IRC08:29
*** dsneddon has quit IRC08:32
kaifengmorning rpittau08:32
rpittauhey kaifeng :)08:33
*** rcernin has joined #openstack-ironic08:38
*** dougsz has joined #openstack-ironic08:40
*** stendulker has quit IRC08:41
openstackgerritMerged openstack/ironic master: Add deploy steps for Redfish BIOS interface  https://review.opendev.org/64206008:42
arne_wiebalckkaifeng: This one looks like it should help: https://review.opendev.org/#/c/67153409:00
patchbotpatch 671534 - nova - ironic: take over instances from down compute serv... - 1 patch set09:00
* arne_wiebalck is upgrading our ironic deployment atm09:00
*** dsneddon has joined #openstack-ironic09:02
*** rcernin has quit IRC09:04
rpittauarne_wiebalck: good luck :)09:08
arne_wiebalckrpittau: thx :)09:11
kaifengarne_wiebalck: thanks!09:12
openstackgerritMerged openstack/ironic-python-agent master: Fixes get_holder disks with nvme drives  https://review.opendev.org/67562009:14
*** dsneddon has quit IRC09:17
*** dsneddon has joined #openstack-ironic09:28
*** dsneddon has quit IRC09:33
*** kaifeng has quit IRC09:43
*** ociuhandu has joined #openstack-ironic09:49
openstackgerritraphael.glon proposed openstack/ironic-python-agent master: Softraid: partitioning fixes  https://review.opendev.org/67481909:50
openstackgerritraphael.glon proposed openstack/ironic-python-agent master: image extension, install_bootloader improvements  https://review.opendev.org/67487909:53
openstackgerritShivanand Tendulker proposed openstack/ironic master: Add new method 'apply_configuration' to RAIDInterface  https://review.opendev.org/67426910:07
*** adrianc has joined #openstack-ironic10:07
openstackgerritShivanand Tendulker proposed openstack/ironic master: Add iLO RAID deploy steps  https://review.opendev.org/67427110:07
openstackgerritShivanand Tendulker proposed openstack/ironic master: WIP: Add iDRAC RAID deploy steps  https://review.opendev.org/64173110:08
*** ociuhandu has quit IRC10:14
*** ociuhandu has joined #openstack-ironic10:15
*** alexmcleod has joined #openstack-ironic10:16
* arne_wiebalck finished the upgrade to Stein (+ s/w RAID) \o/10:22
*** verma-varsha has joined #openstack-ironic10:22
*** verma-varsha1 has joined #openstack-ironic10:27
*** dougsz has quit IRC10:28
*** verma-varsha has quit IRC10:29
*** verma-varsha1 has quit IRC10:29
*** bnemec has quit IRC10:34
*** bnemec has joined #openstack-ironic10:37
*** verma-varsha has joined #openstack-ironic10:39
*** dougsz has joined #openstack-ironic10:44
*** bnemec has quit IRC10:45
*** bnemec has joined #openstack-ironic10:49
*** priteau has joined #openstack-ironic10:50
*** adrianc has quit IRC11:00
*** adrianc has joined #openstack-ironic11:03
*** bnemec has quit IRC11:04
openstackgerritPradip Kadam proposed openstack/ironic master: DRAC : clear_job_queue clean step to fix pending bios config jobs  https://review.opendev.org/67402111:08
*** bnemec has joined #openstack-ironic11:09
*** bnemec has quit IRC11:13
TheJuliaarne_wiebalck: \o/11:14
TheJuliaGood morning everyone!11:14
TheJuliarpioso: spec merged correct? if so needs-spec is not really needed. I don't see an issue applying the tag though11:19
TheJuliaarne_wiebalck: very interesting read yesterday. I was very shocked about how long it takes for a single nova-compute to update the list of resources in nova. Like... alarmingly shocked.11:21
arne_wiebalckTheJulia: User visible shocking, actually :)11:23
arne_wiebalckTheJulia: FWIU, this is only/mostly needed to clean up inconsistencies, though.11:24
TheJuliaarne_wiebalck: Yeah, I'm suspecting all of those checks are soft of like death by a thousand cuts outside of the deepcopy11:25
arne_wiebalckTheJulia: exactly11:25
arne_wiebalckTheJulia: perfectly fine for hosts with virtual instances11:26
*** bnemec has joined #openstack-ironic11:26
TheJuliaSounds like nova/ironic ought to try and discuss this and see if any of us can somehow reduce the thousand cuts and the deepcopy11:27
TheJuliaefried: ^^^11:27
arne_wiebalckBelmiro (who has done most of the analysis in our team) is in touch with efried.11:27
*** dsneddon has joined #openstack-ironic11:29
TheJuliaokay, awesome11:29
arne_wiebalckTheJulia: We have also looked into sharding, but will not do this before nova is on Stein as well.11:29
arne_wiebalckTheJulia: Should happen in the coming weeks.11:30
openstackgerritIlya Etingof proposed openstack/ironic master: Add Redfish Virtual Media Boot support  https://review.opendev.org/63845311:30
arne_wiebalckTheJulia: This may be a candidate for backporting: https://review.opendev.org/#/c/650942/11:30
patchbotpatch 650942 - ironic - Do not tear down node upon cleaning failure - 6 patch sets11:30
*** bnemec has quit IRC11:31
openstackgerritIlya Etingof proposed openstack/ironic master: Add set_boot_device hook in `redfish` boot interface  https://review.opendev.org/67212311:31
openstackgerritIlya Etingof proposed openstack/ironic master: Add `filename` parameter to Redfish virtual media boot URL  https://review.opendev.org/67105411:31
TheJuliaarne_wiebalck: Okay, I know yahoo folks have made mention of insanely long update loops to reconcile, but their fleet is insanely large, comparable times for you guys with a much smaller enrolled fleet kind of caused me to drop my jaw11:32
arne_wiebalckTheJulia: Not sure the resource tracker already exists in the release they run.11:34
arne_wiebalckTheJulia: They had issues with power synchronization I think.11:34
*** dsneddon has quit IRC11:34
TheJulianot anywhere near current day form11:34
TheJuliayeah11:34
openstackgerritIlya Etingof proposed openstack/ironic master: [WIP] Add iDRAC boot interface  https://review.opendev.org/67249811:35
TheJuliaI wonder if nova tried to perform updates in smaller sweeps with ironic. i.e. if we made node searchable by time since last update... if that would at least reduce the set of machines downward...11:36
*** bnemec has joined #openstack-ironic11:38
* arne_wiebalck is somewhat out of his depth on this one11:39
arne_wiebalckI think there are some fundamental (algorithmic) things to improve first.11:39
TheJuliaIndeed, but if we could change the fundimental access pattern to be more helpful as to what to process through resource tracker... The downside is things eventually disappearing :\11:40
arne_wiebalckAgreed. How about we check with the nova folks to see if/how ironic can help with improving things in this area? Eric and Belmiro have been working on this during the past weeks, so I guess they would have an answer quite quickly.11:43
TheJulia++11:44
jrollmorning everyone11:44
TheJuliaGood morning jroll11:44
rpittauhey jroll :)11:44
* jroll loves this conversation11:44
*** bnemec has quit IRC11:45
arne_wiebalckjroll: o/11:45
* arne_wiebalck will check with Belmiro11:45
TheJuliaarne_wiebalck: thanks11:46
arne_wiebalckjroll: do you still work on this one: https://review.opendev.org/#/c/671534?11:46
patchbotpatch 671534 - nova - ironic: take over instances from down compute serv... - 1 patch set11:46
arne_wiebalckjroll: sorry, I guess so (somehow thought the year was 2018)11:47
jrollarne_wiebalck: I don't have time to work on it - I put up a similar POC downstream for people to take over and thought I'd push it upstream in case someone found it useful :)11:47
arne_wiebalckjroll: ok, I see11:48
jrollI'm not sure if our downstream folks even took it yet11:48
*** bnemec has joined #openstack-ironic11:48
arne_wiebalckjroll: we will need sth like this once we shard11:48
jrollthey just finished up our ocata upgrade (finally!) and this is a bug in the backlog11:48
TheJuliajroll: mind if I comment on that change that your okay with someone taking it over?11:49
arne_wiebalckjroll: you have sharded nova-compute?11:49
jrollarne_wiebalck: can you define shard? :)11:49
jrollTheJulia: I'll do it11:49
TheJuliajroll: awesome, thanks~!11:49
arne_wiebalckjroll: multiple nova-computes11:49
arne_wiebalckjroll: we currently have one with 3k nodes11:49
jrollarne_wiebalck: ah ok, we're still playing with it, but are running multiple nova-computes11:50
jrollso a couple things here11:50
jroll1) agree you need something to take over instances, but the existing code without that isn't any worse than what you have today (if a nova-compute service goes down, those instances can't be managed), so you could still go ahead and try it11:51
arne_wiebalckjroll: yes11:51
TheJuliahmm... failure to find disks has killed the standalone job twice in the last 24 hours11:52
jroll2) we did a hack where one nova-compute process spawns multiple nova-compute services, I highly recommend not doing this, it's been painful :) start multiple instances of the services instead11:52
arne_wiebalckjroll: ok, good to know!11:52
TheJulia2 sounds very ouchy11:53
jroll3) restart nova-compute services one at a time, or else the hash ring churn gets insane. any computes starting at the same time see the others as down, and so they take all the nodes11:53
jrollI think that's it11:53
jroll2 was brutal, but was a good experiment11:53
arne_wiebalckjroll: ai! 3) sounds nasty11:53
arne_wiebalckjroll: what's the main reason you run with multiple nova-computes?11:53
jrollarne_wiebalck: HA and reducing resource tracker runtimes11:54
arne_wiebalckjroll: w/o take over there is no HA, is there? ;)11:55
arne_wiebalckjroll: how many nodes do you have per nova-compute?11:55
*** bnemec has quit IRC11:55
jrollooo, that deepcopy is rough, I don't think that's in ocata11:55
arne_wiebalckjroll: I think that came later, yes11:55
jrollarne_wiebalck: heh, there's HA for new instances at least :P11:56
arne_wiebalckjroll: true!11:56
jrollI'm not sure how many offhand, I'd have to check11:56
arne_wiebalckstill 1000s I guess11:56
jrollfor sure11:56
arne_wiebalckjroll: thanks a lot!11:57
*** belmoreira has joined #openstack-ironic11:58
*** dsneddon has joined #openstack-ironic11:59
jrollpagination always confuses me, I knew there had to be more than 1k servers in this deployment :P11:59
jrollarne_wiebalck: you're welcome :)11:59
*** bnemec has joined #openstack-ironic11:59
TheJuliajroll: eh... I think there is hard coded only return 1k nodes unless told otherwise.12:00
TheJuliawhich... likely needs to change12:00
jrollyeah12:00
jrollwell, it's configurable12:00
* TheJulia doesn't remember it being configurable but then \o/12:01
jrollmax_limit I think12:01
jroll--limit 0 will make the client grab all, and also make me wait >:(12:01
jrollarne_wiebalck: the first cluster we started having problems with was around 3400 nodes per nova-compute instance12:02
arne_wiebalckjroll: ai ... we're at 3k now, but next deliveries arrive in a couple of weeks I believe12:02
TheJuliajroll: yeah, maybe a flag to say "forget pagination, just stream the results back to me" ?12:02
jrollTheJulia: interesting, could be fun :)12:03
TheJuliahow much time is that wait presently?12:03
TheJuliawith pagination trying to do it's thing12:03
TheJuliaroughly12:03
jrollheh12:03
*** rh-jelabarre has joined #openstack-ironic12:03
jrollgetting about 10k nodes per minute12:04
jrolla little more12:04
TheJuliagiven how the object model works, that is not bad12:04
jrollso... 6s per api call12:04
jrollyeah12:04
*** dsneddon has quit IRC12:04
*** bnemec has quit IRC12:04
jrollnot sure how much of that is on each end, of course, the client takes a bit to format too12:04
TheJuliaThat is true, but I seem to remember the bulk of the enforcement is in the api because of the line return limit for raw clients12:05
jrollyeah, pagination happens in the API, I just mean the client takes time to format results12:05
TheJuliaAnyway, allowing a processor to begin working on a list while being returned might be useful, although I suspect the client handling that may not actually return until the transfer completes.12:06
TheJuliayeah12:06
*** bnemec has joined #openstack-ironic12:10
*** dsneddon has joined #openstack-ironic12:10
*** belmoreira has quit IRC12:13
*** ociuhandu has quit IRC12:14
openstackgerritJulia Kreger proposed openstack/ironic master: Follow-up to power sync reno  https://review.opendev.org/67640012:14
*** dsneddon has quit IRC12:16
TheJuliarpittau: When you have a couple minutes free, please take a look at https://review.opendev.org/#/c/67426912:17
patchbotpatch 674269 - ironic - Add new method 'apply_configuration' to RAIDInterface - 6 patch sets12:17
rpittauTheJulia: ack12:17
openstackgerritIlya Etingof proposed openstack/ironic master: Add Redfish Virtual Media Boot support  https://review.opendev.org/63845312:19
openstackgerritIlya Etingof proposed openstack/ironic master: Add set_boot_device hook in `redfish` boot interface  https://review.opendev.org/67212312:20
openstackgerritIlya Etingof proposed openstack/ironic master: Add `filename` parameter to Redfish virtual media boot URL  https://review.opendev.org/67105412:20
openstackgerritIlya Etingof proposed openstack/ironic master: [WIP] Add iDRAC boot interface  https://review.opendev.org/67249812:21
TheJuliaeek... rebased them all12:22
TheJuliaoh, whew... gerrit did the right thing12:23
* rpittau getting lost in ephemeral disks12:23
TheJuliarutro!12:27
TheJuliahjensas: Any any ideas on the networking-baremetal traffic passing failure?12:27
openstackgerritVarsha Verma proposed openstack/sushy-tools master: Add Storage and Storage Controllers resource support  https://review.opendev.org/67433912:27
*** bnemec has quit IRC12:29
hjensasTheJulia: I'm still looking at it, no clues yet. I'm setting up a local devstack to reproduce the job atm.12:30
TheJuliahjensas: ack, I was suspecting that would have to be the next step :(12:30
TheJulialet me know if there is anything I can do to help, otherwise I'm going to try and stay out of the way and work on reviewing patches.12:30
hjensasTheJulia: ok.12:31
*** bnemec has joined #openstack-ironic12:33
*** rcernin has joined #openstack-ironic12:34
openstackgerritJulia Kreger proposed openstack/ironic stable/stein: Check for deploy.deploy deploy step in heartbeat  https://review.opendev.org/67615112:38
*** rcernin has quit IRC12:40
*** bnemec has quit IRC12:41
*** belmoreira has joined #openstack-ironic12:42
*** belmoreira has quit IRC12:43
*** bnemec has joined #openstack-ironic12:44
*** priteau has quit IRC12:48
*** bnemec has quit IRC12:49
*** dsneddon has joined #openstack-ironic12:49
*** belmoreira has joined #openstack-ironic12:51
openstackgerritMerged openstack/sushy master: Implements adapter checking  https://review.opendev.org/66996312:53
*** bnemec has joined #openstack-ironic12:54
*** rloo has joined #openstack-ironic12:55
*** dsneddon has quit IRC12:56
*** dsneddon has joined #openstack-ironic13:02
*** jcoufal has joined #openstack-ironic13:04
*** bnemec has quit IRC13:07
*** dsneddon has quit IRC13:08
*** bnemec has joined #openstack-ironic13:10
*** bnemec has quit IRC13:16
*** beekneemech has joined #openstack-ironic13:16
*** belmoreira has quit IRC13:28
*** beekneemech has quit IRC13:33
*** belmoreira has joined #openstack-ironic13:34
*** dsneddon has joined #openstack-ironic13:36
TheJuliaarne_wiebalck: would be super appreciative of any IPA reviews if you have time, mainly because there are a couple software raid related changes/fixes.13:37
*** bnemec has joined #openstack-ironic13:39
openstackgerritMerged openstack/ironic master: Ansible: fix partition_configdrive for logical root_devices  https://review.opendev.org/67464313:44
efriedo/ arne_wiebalck TheJulia -- looks like I missed a discussion about perf?13:44
*** cdent has joined #openstack-ironic13:45
*** belmoreira has quit IRC13:46
*** bnemec has quit IRC13:47
efriedIt occurs to me that belmoreira might benefit from this (credit cdent) to narrow down places where we can get the most bang https://docs.openstack.org/nova/latest/contributor/testing/eventlet-profiling.html13:47
cdent(that needs a caveat/warning about how it fails to work well when rpc is involved, and workarounds for dealing with that, but for the most part is goodness)13:48
*** bnemec has joined #openstack-ironic13:51
*** dsneddon has quit IRC13:52
*** bnemec has quit IRC13:55
TheJuliaNeat, it does sound like they went fairly down the rabbit hole and found various issues/pain points. At least that was my perception from reading what they published.13:57
*** belmoreira has joined #openstack-ironic13:58
*** bnemec has joined #openstack-ironic13:58
arne_wiebalckThe main question from the ironic side is probably if ironic can do sth to help or if the improvements are mostly needed in the general nova code (i.e. even outside the ironic driver, let alone how ironic provides data to nova).14:03
*** sthussey has joined #openstack-ironic14:04
jrollarne_wiebalck: oh, I forgot to mention one other way to shard is with the conductor locality feature14:04
arne_wiebalckMy understanding from a discussion with belmoreira earlier was that the main improvements are expected to come from the shared code in nova (shared code meaning used for virtual and physical instances).14:04
jrollin ironic, it shards conductors, but you can also tie nova-compute instances to the group14:05
arne_wiebalckjroll: conductor locality?14:05
cdentarne_wiebalck: yes, that's my understanding too14:05
jrollthe main benefit in nova being that it pulls a subset of nodes from ironic to iterate over in the resource tracker, instead of all14:05
* jroll finds docs14:05
jrollarne_wiebalck: https://docs.openstack.org/ironic/latest/admin/conductor-groups.html14:05
cdentnova likes to lock over loops that it was designed to think are small but with ironic are big14:05
arne_wiebalckcdent: right14:06
cdent(they are also big in clustered hypervisors)14:06
jrollcdent: ++14:06
*** belmoreira has quit IRC14:06
arne_wiebalckcdent: we are considering to have multiple nova-computes to mitigate, but from jroll's (and our) experience this has limits14:07
* cdent nods14:07
jrollwelllll, it has caveats. if you spun an n-cpu per ironic node, you'd be golden :P14:07
cdentthousands of tiny n-cpu containers?14:08
jroll"tiny" is relative, but yeah14:08
cdent(is not actually the worst idea every)14:08
cdentever14:08
jrollthat would be hard to manage, but not impossible14:08
cdentjust tells k8s to do it, doesn't it fix everything ;)14:08
*** dsneddon has joined #openstack-ironic14:09
jrollsure, run my k8s for me and that sounds great :D14:09
jrollre: what can ironic do: the one big change that ironic could do is to take over management of placement for ironic nodes. if we were putting the right data in placement as it changes, we could just noop the resource tracker in nova. but this comes with all sorts of problems in the developer world14:09
cdentI was gonna say that too, but it feels churlish for me to mention it first, so I'm glad you did14:10
arne_wiebalckhmm, isn't the rt a little less important than its name implies?14:11
* arne_wiebalck maybe does not fully understand its role14:11
jrollin short, the resource tracker is what takes the ironic node data and puts it in placement for the scheduler to use14:11
arne_wiebalckbut only for new resources, no?14:11
jrollexisting as well14:12
arne_wiebalckI don't think so.14:12
jrolle.g. if you put an available node in maintenance mode, it will notice, and drop the available resources14:12
arne_wiebalckI thought this was based on explicit events.14:12
jrollif you update the resource class on a node, it updates placement accordingly, etc14:12
jrollit is not :(14:12
jrollit should be, but is not14:12
cdentIs it already a given in these discussion that nova needs to be involved at all (presumably for the sake of the api?)? If that's fungible, the options open up broadly.14:12
jroll(this model would also work for new resources)14:13
jrollcdent: it is not always a given, but a single API for all compute resources is a major benefit that some orgs would like to continue to take advantage of14:13
*** dsneddon has quit IRC14:13
*** absubram has joined #openstack-ironic14:15
rpittauTheJulia: I noticed that in https://review.opendev.org/666591 we're mixing legacy neutron libs, I don't think that is advisable, and I'm actually struggling understanding where that come from at the moment :/14:17
patchbotpatch 666591 - ironic-python-agent-builder - [WIP] Update tinycore from 8.x to 10.x - 12 patch sets14:17
*** absubram has quit IRC14:19
*** ociuhandu has joined #openstack-ironic14:21
*** absubram has joined #openstack-ironic14:25
*** ociuhandu has quit IRC14:26
arne_wiebalckcdent: jroll: Any suggestions how to best follow up on this discussion and come up with a plan on how to move forward?14:26
cdentarne_wiebalck: I missed the start of the conversation so I'm not clear on the full context14:26
*** belmoreira has joined #openstack-ironic14:27
arne_wiebalckcdent: I think the trigger was the performance issues in the resource tracker for larger ironic deployments. efried and belmoreira are working on this.14:27
arne_wiebalckcdent: TheJulia offered help from the ironic side (if needed).14:28
*** jcoufal_ has joined #openstack-ironic14:29
cdentOne thing I think is probably worth exploring is using a deque as the source of things do during the locked loop and just do N things per go. Each time you process something, put it back on the end of the deque14:29
cdenthowever14:29
cdentin some cases that's going to upset the view of reality14:29
cdentand managing that becomes the tricky bit14:29
cdentI would think that if belmoreira and efried are on the case in some fashion, that's pretty good odds that things will be improved14:30
arne_wiebalckagreed14:30
cdentas belmoreira pointed out in his blog post there are some definite problem in the ProviderTree data structure that are probably amenable to improvement14:30
arne_wiebalckright14:31
*** jcoufal has quit IRC14:31
rpiosoTheJulia: I tidied up https://storyboard.openstack.org/#!/story/2004592 yesterday by adding a couple of the tags we discussed, changing the project of the task associated with the spec to openstack/ironic-specs, and adding a new task for the implementation. I left the description there, even though it's been superseded by the spec.14:32
arne_wiebalckI was just thinking if we take a step back, there may be even more efficient options/ideas (like "does nova need to be involved?").14:32
arne_wiebalckTotally fine with letting efried and belmoreira finish their analysis :)14:32
TheJuliarpittau: got a link to a log?14:33
*** ijw has joined #openstack-ironic14:33
arne_wiebalckTheJulia: I'll try to have a look at the IPA patches.14:33
*** dsneddon has joined #openstack-ironic14:36
jrollarne_wiebalck: I do think we should have a bigger discussion about "can we eliminate the RT for the ironic driver", which might should start in the ML or a spec14:36
rpiosoTheJulia: On a different topic, I would like us to discuss prospective homes for sushy OEM extension Git repos during our next weekly meeting. Spoiler alert! I'm thinking of the opendev.org/x namespace. May I add it to the Discussion section of the agenda?14:38
arne_wiebalckjroll: That should better happen before too much work goes into optimizing it.14:39
jrollarne_wiebalck: good point, I somewhat agree, though it's going to be a long road and doesn't hurt to improve it in the meantime14:41
rpiosojroll: You may be interested in ^^^14:41
arne_wiebalckjroll: also true14:41
jrollrpioso: the 'x' namespace is for unofficial projects, so I would just go ahead and put them there now. if ironic decides to take them in we can move them :)14:42
arne_wiebalckjroll: as major changes will much longer, I'd think that the PTG might be a good place to start the discussion, but it seems many will skip the Shanghai one14:42
jrollarne_wiebalck: agree and agree :(14:42
*** dsneddon has quit IRC14:47
jrollarne_wiebalck: now that I think about it, I don't think cdent will be there either? so best to start on the ML14:47
cdenti will not, no14:48
cdentML helps to filter out a lot of the shared language building prior to in person chat, too14:49
jroll++14:50
*** priteau has joined #openstack-ironic14:50
rpittauTheJulia: sorry, got mixed up with other things and inverted the logs, that is actually not using legacy, while other patches are14:50
arne_wiebalckcdent: jroll: ok, I'll check with belmoreira (to also check with efried), I think it'd be best if they drive this as they probably have the best overview atm14:51
efriedarne_wiebalck: I'm not at the PTG either14:51
efriedbut yeah, we can do things in IRC/ML14:52
jroll++14:52
arne_wiebalckefried: ack, thx!14:52
*** dsneddon has joined #openstack-ironic14:55
*** ociuhandu has joined #openstack-ironic14:57
*** dsneddon has quit IRC15:01
*** absubram has quit IRC15:09
*** alexmcleod is now known as alexmcleod|bbl15:11
*** ociuhandu has quit IRC15:11
*** ociuhandu has joined #openstack-ironic15:12
*** jcoufal_ has quit IRC15:12
*** ijw_ has joined #openstack-ironic15:12
*** jcoufal has joined #openstack-ironic15:13
*** ijw has quit IRC15:14
TheJuliarpioso: Is there really a need. It is not like the ironic community has to approve x/ namespace usage and there has been a past desire to not pull specific vendor items into the project's scope. Perhaps one day there might be the case for a separate namespace though....15:16
TheJuliaarne_wiebalck: I really think the ML is the best place because the PTG is likely to be a specific slice of contributorship and I suspect most PTG time will actually be more of the "bridge building" and listening to issues/needs sort of time. Anything discussed there would need to go back to the ML anyway.15:18
TheJuliarpittau: I'm afraid you have me confused, are you referring to networking-baremetal?15:19
rpittauTheJulia: sorry, I'm referring to the migration to tinycore 10.x in ironic-python-agent-builder15:20
*** ianychoi has joined #openstack-ironic15:20
TheJuliaso, we did merge a change the default template to not use legacy this week, I guess we might need to do it there depending on job config...15:21
rpittauTheJulia: no, my bad, I confused two different logs, that is actually using the updated neutron, not legacy15:22
TheJuliait happens, that was earlier in the week too so older job logs may be slightly confusing15:23
arne_wiebalckTheJulia: Checking with belmoreira, the plan was to upgrade to Stein first, then re-assess (as it brings some patches in this area).15:25
arne_wiebalckTheJulia: The (nova) upgrade should happen during the next weeks.15:26
arne_wiebalckTheJulia: OK with you if we wait and pick this up then?15:27
TheJuliaarne_wiebalck: absolutely. One thing though, I'm going to be absent for a good chunk of September.15:28
* TheJulia needs to send that email15:28
rpiosoTheJulia: I thought we could share thoughts like those with the ironic community. Either way, I'm good with proceeding.15:28
rpiosojroll, TheJulia: Thank you!15:29
arne_wiebalckTheJulia: ok15:29
TheJuliarpioso: I think if the project as a whole wants to reconsider, then that might be a good topic. I'm just not sure the outcome would really go anywhere unless we had quorum of others who may be impacted/affected or who may directly benefit.15:30
TheJuliarpioso: in otherwords, it might also be a good mailing list topic :)15:30
*** dsneddon has joined #openstack-ironic15:33
rpiosoTheJulia: Since that one day may be quite some time in the future, we'll proceed to create the repo in 'x'. No need to hold it up while discussing prospective futures. After it exists, an ML topic may be worthwhile to inform the community of its existence and possible paths forward should critical mass be achieved.15:38
*** ijw has joined #openstack-ironic15:39
TheJuliarpioso: I was thinking the exact same thing!15:39
*** dsneddon has quit IRC15:39
rpiosoTheJulia: \o/15:40
rpiosoTheJulia: Is #openstack-infra the place to go to ask about creating it or is TC approval needed first?15:42
*** ijw_ has quit IRC15:42
jrollthere's a guide for this :)15:43
jrollrpioso: https://docs.openstack.org/infra/manual/creators.html15:43
*** ijw_ has joined #openstack-ironic15:46
*** absubram has joined #openstack-ironic15:48
*** jcoufal has quit IRC15:48
*** ijw has quit IRC15:49
rpiosojroll: Now that's a guide! Thanks, again.15:50
mbuilhey guys, when using virtualbmc version 1.5, if I don't specify a log file, at some point the vbmc server crashes. Have you ever noticed this? This was not happening in 1.3, so I suspect this was introduced by the new architecture but no idea why :(15:50
jrollrpioso: no problem :)15:52
rpittaubye all, see you on monday, long weekend here :)15:55
*** rpittau is now known as rpittau|afk15:55
*** tesseract has quit IRC15:56
*** tssurya has quit IRC16:05
*** tssurya has joined #openstack-ironic16:05
*** belmoreira has quit IRC16:07
*** cdent has quit IRC16:12
*** lucasagomes has quit IRC16:15
etingofmbuil, o/ I've not encountered that, however if you do, your best bet would be to open up an issue on storyboard enclosing your config and traceback16:18
TheJuliambuil:  Interesting... I wonder upon detach, since stdout is not a valid option, if python blows upon trying to write to it for logging....16:21
TheJuliaetingof: ^^^ just a crazy idea16:22
*** dsneddon has joined #openstack-ironic16:27
mbuiletingof: ok, I'll do that :). Another thing that I found is that when doing vbmc delete server_A, the config gets deleted but the process that was created when adding it stays there but now as "[vbmc] <defunct>". Looks like a zombie process16:28
mbuil5     0 131170 131097  20   0      0     0 exit   Z    ?          0:02 [vbmc] <defunct>16:28
mbuilI'll add that to the storyboard too16:29
*** ricolin has quit IRC16:29
*** dsneddon has quit IRC16:33
*** ijw_ has quit IRC16:34
*** alexmcleod|bbl has quit IRC16:37
*** tssurya has quit IRC16:39
*** belmoreira has joined #openstack-ironic16:39
*** ociuhandu_ has joined #openstack-ironic16:41
*** belmoreira has quit IRC16:41
*** fungi has quit IRC16:42
*** fungi has joined #openstack-ironic16:43
*** ociuhandu has quit IRC16:44
*** ociuhandu_ has quit IRC16:45
*** dsneddon has joined #openstack-ironic16:47
*** adrianc has quit IRC16:50
*** e0ne has quit IRC16:51
*** adrianc has joined #openstack-ironic16:52
*** dsneddon has quit IRC16:53
etingofmbuil, this all sounds very weird... can it be related to python version you are on, I wonder...16:55
*** ociuhandu has joined #openstack-ironic16:56
*** derekh has quit IRC17:00
*** ociuhandu has quit IRC17:00
openstackgerritMerged openstack/ironic stable/stein: Fixes power-on failure for 'ilo' hardware type  https://review.opendev.org/67445817:00
etingofmbuil, also, if you could set logging level to debug and see what happens in the log, that can possibly reveal something... or not17:02
etingofTheJulia's crazy idea is too reasonable17:04
*** ijw has joined #openstack-ironic17:05
*** jcoufal has joined #openstack-ironic17:12
openstackgerritMerged openstack/ironic master: Enable testing software RAID in the standalone job  https://review.opendev.org/67510217:14
*** dsneddon has joined #openstack-ironic17:25
TheJuliaetingof: sorry :(17:29
TheJuliaI'll strive harder for crazy ideas!17:29
TheJuliaWhat if it was from an exploding fuel tank flying out of a nearby computer running kerbal space program?!? ;)17:30
*** dsneddon has quit IRC17:30
etingofdo you mean - PSU, not rocket booster perhaps?17:31
TheJuliaWell, I recently started packing the science packages you could place on far off places... and they do require solar panel packages.......17:32
etingofironic goes beyond the clouds, it seems!17:34
*** priteau has quit IRC17:37
etingofit seems there is a design problem in sushy OEM class hierarchy, but fixing it clearly might introduce a backward hiccup17:38
*** dougsz has quit IRC17:41
TheJuliaetingof: well, was it not already moderately not working until we just merged a patch recently?17:46
TheJuliaI guess regardless, it would be a major version bump17:46
* etingof is taking it as a suggestion not to bother with backward compatibility for OEM! \o/17:47
etingofotherwise it's quite messy17:47
*** verma-varsha has joined #openstack-ironic18:05
*** gyee has joined #openstack-ironic18:06
*** e0ne has joined #openstack-ironic18:08
*** dsneddon has joined #openstack-ironic18:09
*** e0ne has quit IRC18:10
*** ociuhandu has joined #openstack-ironic18:11
*** verma-varsha has quit IRC18:12
*** dsneddon has quit IRC18:14
*** dsneddon has joined #openstack-ironic18:28
*** verma-varsha has joined #openstack-ironic18:29
*** ociuhandu has quit IRC18:34
*** verma-varsha has quit IRC18:46
TheJuliaI might be awful for thinking that, and others may disagree, but no real reason to maintain it if it hasn't been used and we just fixed it. Just need to do the major version bump19:43
etingofthat's my thinking as well19:45
rpiosoetingof: Are you preparing to cut a new sushy release?19:46
etingofrpioso, not yet, we seem to require OEM resource model change in sushy19:47
rpiosoetingof: Has that change already been proposed or merged?19:48
etingofno, I am still experimenting with that19:48
rpiosoetingof: Gotcha. I'll be on the lookout for it. Thanks!19:50
*** e0ne has joined #openstack-ironic20:14
*** ijw has quit IRC20:17
*** ijw has joined #openstack-ironic20:24
*** ijw has quit IRC20:27
openstackgerritJulia Kreger proposed openstack/python-ironicclient master: WIP: Remove the ironic command  https://review.opendev.org/67651520:28
*** ijw has joined #openstack-ironic20:28
*** ijw has quit IRC20:30
*** ijw has joined #openstack-ironic20:30
openstackgerritMatt Riedemann proposed openstack/ironic stable/rocky: CI: remove quotation marks from TEMPEST_PLUGINS variable  https://review.opendev.org/67651720:37
*** mriedem has joined #openstack-ironic20:38
*** ijw has quit IRC20:38
*** ijw has joined #openstack-ironic20:38
*** efried has quit IRC20:45
*** efried has joined #openstack-ironic20:46
*** mriedem has left #openstack-ironic20:56
*** jtomasek has quit IRC21:01
*** jcoufal has quit IRC21:10
*** e0ne has quit IRC21:11
*** e0ne has joined #openstack-ironic21:13
*** e0ne has quit IRC21:13
*** dsneddon has quit IRC21:56
openstackgerritSteve Baker proposed openstack/metalsmith master: Allow reserve_node to backfill from existing node  https://review.opendev.org/67652522:32
openstackgerritSteve Baker proposed openstack/metalsmith master: Allow reserve_node to backfill from existing node  https://review.opendev.org/67652522:36
*** absubram has quit IRC23:29
*** ijw has quit IRC23:29
*** sthussey has quit IRC23:53

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!